Bayesian Computation Statistics
Rashin Nimaei; Farzad Eskandari
Abstract
The recent advancements in technology have faced an increase in the growth rate of data.According to the amount of data generated, ensuring effective analysis using traditional approaches becomes very complicated.One of the methods of managing and analyzing big data ...
Read More
The recent advancements in technology have faced an increase in the growth rate of data.According to the amount of data generated, ensuring effective analysis using traditional approaches becomes very complicated.One of the methods of managing and analyzing big data is classification.%One of the data mining methods used commonly and effectively to classify big data is the MapReduceIn this paper, the feature weighting technique to improve Bayesian classification algorithms for big data is developed based on Correlative Naive Bayes classifier and MapReduce Model.%Classification models include Naive Bayes classifier, correlated Naive Bayes and correlated Naive Bayes with feature weighting.Correlated Naive Bayes classification is a generalization of the Naive Bayes classification model by considering the dependence between features.%This paper uses the feature weighting technique and Laplace calibration to improve the correlated Naive Bayes classification.The performance of all described methods are evaluated by considering accuracy, sensitivity and specificity, accuracy, sensitivity and specificity metrics.
Bayesian Computation Statistics
Ehsan Ormoz; Farzad Eskandari
Abstract
This paper introduces a novel semiparametric Bayesian approach for bivariate meta-regression. The method extends traditional binomial models to trinomial distributions, accounting for positive, neutral, and negative treatment effects. Using a conditional Dirichlet process, we develop a model to compare ...
Read More
This paper introduces a novel semiparametric Bayesian approach for bivariate meta-regression. The method extends traditional binomial models to trinomial distributions, accounting for positive, neutral, and negative treatment effects. Using a conditional Dirichlet process, we develop a model to compare treatment and control groups across multiple clinical centers. This approach addresses the challenges posed by confounding factors in such studies. The primary objective is to assess treatment efficacy by modeling response outcomes as trinomial distributions. We employ Gibbs sampling and the Metropolis-Hastings algorithm for posterior computation. These methods generate estimates of treatment effects while incorporating auxiliary variables that may influence outcomes. Simulations across various scenarios demonstrate the model’s effectiveness. We also establish credible intervals to evaluate hypotheses related to treatment effects. Furthermore, we apply the methodology to real-world data on economic activity in Iran from 2009 to 2021. This application highlights the practical utility of our approach in meta-analytic contexts. Our research contributes to the growing body of literature on Bayesian methods in meta-analysis. It provides valuable insights for improving clinical study evaluations.
Machine Learning
Sahar Abbasi; Radmin Sadeghian; Maryam Hamedi
Abstract
Multi-label classification assigns multiple labels to each instance, crucial for tasks like cancer detection in images and text categorization. However, machine learning methods often struggle with the complexity of real-life datasets. To improve efficiency, researchers have developed feature selection ...
Read More
Multi-label classification assigns multiple labels to each instance, crucial for tasks like cancer detection in images and text categorization. However, machine learning methods often struggle with the complexity of real-life datasets. To improve efficiency, researchers have developed feature selection methods to identify the most relevant features. Traditional methods, requiring all features upfront, fail in dynamic environments like media platforms with continuous data streams. To address this, novel online methods have been created, yet they often neglect optimizing conflicting objectives. This study introduces an objective search approach using mutual information, feature interaction, and the NSGA-II algorithm to select relevant features from streaming data. The strategy aims to minimize feature overlap, maximize relevance to labels, and optimize online feature interaction analysis. By applying a modified NSGA-II algorithm, a set of non-dominantsolutions is identified. Experiments on eleven datasets show that the proposed approach outperforms advanced online feature selection techniques in predictive accuracy, statistical analysis, and stability assessment.
Machine Learning
Mohammad Zahaby; Iman Makhdoom
Abstract
Breast cancer (BC) is one of the leading causes of death in women worldwide. Early diagnosis of this disease can save many women’s lives. The Breast Imaging Reporting and Data System (BIRADS) is a standard method developed by the American College of Radiology (ACR). However, physicians have had ...
Read More
Breast cancer (BC) is one of the leading causes of death in women worldwide. Early diagnosis of this disease can save many women’s lives. The Breast Imaging Reporting and Data System (BIRADS) is a standard method developed by the American College of Radiology (ACR). However, physicians have had a lot of contradictions in determining the value of BIRADS, and all aspects of patients have not been considered in diagnosing this disease using the methods that have been used so far. In this article, a novel decision support system (DSS) has been presented. In the proposed DSS, firstly, c-mean clustering was used to determine the molecular subtype for patients who did not have this value by combining the mammography reports processing along with hospital information systems (HIS) obtained from their electronic files. Then several classifiers such as convolutional neural networks (CNN), decision tree (DT), multi-level fuzzy min-max neural network (MLF), multi-class support vector machine (SVM), and XGboost were trained to determine the BIRADS. Finally, the values obtained by these classifiers were combined using weighted ensemble learning with the majority voting algorithm to obtain the appropriate value of BIRADS. This helps physicians in the early diagnosis of BC. Finally, the results were evaluated in terms of accuracy, specificity, sensitivity, positive predicted value (PPV), negative predicted value (NPV), and f1-measure by the confusion matrix. The obtained values were, 97.94%, 98.79%, 92.08%, 92.34%, 98.80%, and 92.19% respectively.
Bayesian Computation Statistics
Mahdieh Bayati
Abstract
This study generalizes the joint empirical likelihood (JEL) which is named the joint penalized empirical likelihood(JPEL) and presents a comparative analysis of two innovative empirical likelihood methods: the restricted penalized empirical likelihood (RPEL) and the joint penalized empirical likelihood. ...
Read More
This study generalizes the joint empirical likelihood (JEL) which is named the joint penalized empirical likelihood(JPEL) and presents a comparative analysis of two innovative empirical likelihood methods: the restricted penalized empirical likelihood (RPEL) and the joint penalized empirical likelihood. These methods extend traditional empirical likelihood approaches by integrating criteria based on the minimum variance and unbiasedness of the estimator equations. In RPEL, estimators are obtained under these two criteria, while JPEL facilitates the joint application of the estimator equations used in RPEL, allowing for broader applicability.\\We evaluate the effectiveness of RPEL and RJEL in regression models through simulation studies, and evaluate the performance of RPEL and JPEL, focusing on parameter accuracy, model selection (as measured by the Empirical Bayesian Information Criterion), predictive accuracy (Mean Square Error), and robustness to outliers. Results indicate that RPEL consistently outperforms JPEL across all criteria, with RPEL yielding simpler models and more reliable estimates, particularly as sample sizes increase. These findings suggest that RPEL provides greater stability and interpretability for regression models, making it a superior choice over JPEL for the scenarios tested in this study.