In biomedicine, binary classification problems are involved in diagnostic but also, for instance, in personalized medicine. The objective is to use information for correctly allocating subjects in groups. Frequently, ...
详细信息
In biomedicine, binary classification problems are involved in diagnostic but also, for instance, in personalized medicine. The objective is to use information for correctly allocating subjects in groups. Frequently, this information implies high-dimensional data. An adequate classification rule is a trade-off between the sensitivity and the specificity. The ROC curve helps to understand, evaluate and compare the accuracy of classification processes. We propose a procedure for estimating the optimal classification rules based on a penalized estimator of the underlying probability distribution functions. We study its asymptotic properties. Through Monte Carlo simulations, we compare our proposal with a support vector machine-based ROC curve. We illustrate its practical use in a real-world problem. Results suggest that, despite some techniques promise to improve the results provided by traditional methods, in the binary classification problem, the limit is the actual relationship among the density functions.
This paper proposes a cellular automata-based solution of a binary classification problem. The proposed method is based on a two-dimensional, three-state cellular automaton (CA) with the von Neumann neighborhood. Sinc...
详细信息
This paper proposes a cellular automata-based solution of a binary classification problem. The proposed method is based on a two-dimensional, three-state cellular automaton (CA) with the von Neumann neighborhood. Since the number of possible CA rules (potential CA-based classifiers) is huge, searching efficient rules is conducted with use of a genetic algorithm (GA). Experiments show an excellent performance of discovered rules in solving the classificationproblem. The best found rules perform better than the heuristic CA rule designed by a human and also better than one of the most widely used statistical method: the k-nearest neighbors algorithm (k-NN). Experiments show that CAs rules can be successfully reused in the process of searching new rules.
The generalized receiver-operating characteristic, gROC, curve considers the classification ability of diagnostic tests when both larger and lower values of the marker are associated with higher probabilities of being...
详细信息
The generalized receiver-operating characteristic, gROC, curve considers the classification ability of diagnostic tests when both larger and lower values of the marker are associated with higher probabilities of being positive. Its empirical estimation implies to select the best classification subsets among those satisfying particular condition. Both strong and weak consistency have already been proved. However, using the same data for both to select the classification subsets and to calculate its gROC curve leads to an over-optimistic estimate of the real performance of the diagnostic criteria on future samples. In this work, the bias of the empirical gROC curve estimator is explored through Monte Carlo simulations. Besides, two cross validation based algorithms are proposed for reducing the overfitting. The practical application of the proposed algorithms is illustrated through the analysis of a real world dataset. Simulation results suggest that the empirical gROC curve estimator returns optimistic approximations, especially, in situations in which the diagnostic capacity of the marker is poor and the sample size is small. The new proposed algorithms improve the estimation of the actual diagnostic test accuracy, and get almost unbiased gAUCs in most of the considered scenarios. However, the cross-validation based algorithms reported larger L-1-errors than the standard empirical estimators, and increment the computational cost of the procedures. As online supplementary material, this manuscript includes an R function which wraps up the implemented routines.
The overlap coefficient (OVL) measures the common area between two or more density functions. It has been used for measuring the similarity between distributions in different research fields including astronomy, econo...
详细信息
The overlap coefficient (OVL) measures the common area between two or more density functions. It has been used for measuring the similarity between distributions in different research fields including astronomy, economy or sociology, among others. Recently, different authors have studied the use of the OVL coefficient in the binary classification problem. They argue that, in particular settings, it could provide better accuracy measure than other stablished indices. We prove here that the OVL coefficient does not provide additional information to the Youden index and that, the potential advantages previously reported are based on the assumption that the classification rules underlying any classification process always assign more probability of being positive to the larger values of the marker. Particularly, we prove that, for a fixed continuous marker, the OVL coefficient is equivalent to the Youden index associated with the optimal classification rules based on this marker. We illustrate the problem studying the capacity of the white blood cells count to identify the type of disease in patients having either acute viral meningitis or acute bacterial meningitis.
A good diagnostic test should show different behavior on both the positive and the negative populations. However, this is not enough for having a good classification system. The binary classification problem is a comp...
详细信息
A good diagnostic test should show different behavior on both the positive and the negative populations. However, this is not enough for having a good classification system. The binary classification problem is a complex task, which implies to define decision criteria. The knowledge of the level of dissimilarity between the two involved distributions is not enough. We also have to know how to define those decision criteria. The length of the receiver-operating characteristic curve has been proposed as an index of the optimal discriminatory capacity of a biomarker. It is related not with the actual but with the optimal classification capacity of the considered diagnostic test. One particularity of this index is that its estimation should be based on parametric or smoothed models. We explore here the behavior of a kernel density estimator-based approximation for estimating the length of the receiver-operating characteristic curve. We prove the asymptotic distribution of the resulting statistic, propose a parametric bootstrap algorithm for confidence intervals construction, discuss the role that the bandwidth parameter plays in the quality of the provided estimations and, via Monte Carlo simulations, study its finite-sample behavior considering four different criteria for the bandwidth selection. The practical use of the length of the receiver-operating characteristic curve is illustrated through two real-world examples.
Crack detection of the concrete bridge is an essential index for the safety assessment of bridge structure. It is more important to check the whole structure than to check the accuracy in the damage assessment. Howeve...
详细信息
Crack detection of the concrete bridge is an essential index for the safety assessment of bridge structure. It is more important to check the whole structure than to check the accuracy in the damage assessment. However, the traditional deep learning model method cannot completely detect the crack structure, which challenges image-based crack detection. For this reason, we propose deep bridge crack classification (DBCC)-Net as a classification-based deep learning network. By pruning the Yolox, the regression problem of the target detection is converted to the binary classification problem to avoid the network performance degradation caused by the translation invariance of the convolutional neural network (CNN). In addition, the network post-processing and a two-stage crack detection strategy are proposed to enable the network to detect cracks and extract crack morphology in high-resolution images quickly. In the first stage, DBCC-Net realizes the coarse extraction of crack position based on image slice classification. In the second stage, the complete crack morphology is extracted from the location suggested by the semantic segmentation network. Experimental results show that the proposed two-stage method has 19 frames per second (FPS) and 0.79 Miou (mean intersection over union) at the actual bridge images with 2560x2560 pixels. Although FPS is reduced, the Miou value is 7.8% higher than other methods, proving this paper's practical value.
Flood maps based on Earth Observation (EO) data inform critical decision-making in almost every stage of the disaster management cycle, directly impacting the ability of affected individuals and governments to receive...
详细信息
Flood maps based on Earth Observation (EO) data inform critical decision-making in almost every stage of the disaster management cycle, directly impacting the ability of affected individuals and governments to receive aid as well as informing policies on future adaptation. However, flood map validation also presents a challenge in the form of class imbalance between flood and non-flood classes, which has rarely been investigated. There are currently no established best practices for addressing this issue, and the accuracy of these maps is often viewed as a mere formality, which leads to a lack of user trust in flood map products and a limitation in their operational use and uptake. This paper provides the first comprehensive assessment of the impact of current EO-based flood map validation practices. Using flood inundation maps derived from Sentinel-1 synthetic aperture radar data with synthetically generated controlled errors and Copernicus Emergency Management Service flood maps as the ground truth, binary metrics were statistically evaluated for the quantification of flood detection accuracy for events under varying flood conditions. Especially, class specific metrics were found to be sensitive to the class imbalance, i.e. larger flood magnitudes result in higher metric scores, thus being naturally biased towards overpredicting classifiers. Metric stability across error percentiles and flood magnitudes was assessed through standard deviation calculated by bootstrapping to quantify the impact of sample selection subjectivity, where stratified sampling schemes exhibited the lowest standard deviation consistently. Thoughtful sample and response design were critical, with probability-based random sampling and proportional or equal class allocation vital to producing robust accuracy estimates comparable across study sites, error classes, and flood magnitudes. Results suggest that popular evaluation metrics such as the F1-Score are in fact unsuitable for accurate chara
The distribution of humanitarian aid is a vital issue for humanity's future. In recent years, the management of humanitarian crises has become more crucial than it was a decade ago. Due to the volatility and urgen...
详细信息
The distribution of humanitarian aid is a vital issue for humanity's future. In recent years, the management of humanitarian crises has become more crucial than it was a decade ago. Due to the volatility and urgency that characterize such situations, one of the most important challenges globally is the optimization of decisions regarding the timely distribution of aid during humanitarian operations. Our main goal is to develop an innovative decision-making tool, essential for non-profit organizations and governments that aims at the prompt selection of the location of the distribution center of humanitarian aid, in cases of natural or human-made di-sasters. The proposed tool is based on network science principles and can be used for selecting a suitable node for the installation of a distribution center during the beginning of a humanitarian crisis, considering that networks have a volatile nature and require quick decisions. For the configuration of the proposed tool we use a combi-nation of a classical heuristic algorithm and predictive models based on a binary classification problem with the support of a supervised deep neural network. It is developed using the R programming language with the contribution of the "Shiny" package (web application framework for R) along with other packages for network analysis, data manipulation and visualization.
As it is well known, decision tree is a kind of data-driven classification model, and its primary core is the split criterion. Although a great deal of split criteria have been proposed so far, almost all of them focu...
详细信息
As it is well known, decision tree is a kind of data-driven classification model, and its primary core is the split criterion. Although a great deal of split criteria have been proposed so far, almost all of them focus on the global class distribution of the training data. However, they ignored the local class imbalance problem that commonly appears during the decision tree induction over balanced or roughly balanced binary class data sets. In the present study, this problem is investigated in detail and an adaptive approach based on multiple existing split criteria is proposed. In the proposed scheme, the local class imbalanced ratio is considered as the weight factor to weigh the importance between these split criteria so as to determine the optimal splitting point at each internal node. In order to evaluate the effectiveness of the proposed method, it is applied on twenty roughly balanced real-world binary class data sets. Experimental results show that the proposed method not only outperforms all other methods, but also improves the prediction accuracy of each class.
Facial Kinship Verification involves determining whether two face images belong to relatives, a task that is particularly challenging due to subtle differences in facial features and large intra-class variations. In r...
详细信息
Facial Kinship Verification involves determining whether two face images belong to relatives, a task that is particularly challenging due to subtle differences in facial features and large intra-class variations. In recent years, deep learning models have shown great promise in addressing this problem. In this work, we propose a Vision Transformer (ViT) model for facial Kinship Verification, leveraging the proven effectiveness of Transformer architectures in Natural Language Processing. The Vision Transformer is trained end-to-end on two benchmark datasets: the large-scale Families in the Wild (FIW) dataset, consisting of thousands of face images with corresponding kinship labels, and the smaller KinFaceW-II dataset. Our model employs multiple attention mechanisms to capture complex relationships between facial features and produce a final kinship prediction. Experimental results demonstrate that our approach outperforms state-of-the-art methods, achieving an average accuracy of 92% on the FIW dataset and an F1 score of 0.85. The Euclidean distance metric further enhances the classification of kin and non-kin pairs. These findings confirm the effectiveness of Vision Transformer models for facial Kinship Verification and underscore their potential for future research in this domain.
暂无评论