Diabetes is a disorder that develops in the human body when blood glucose or sugar levels are extremely high. machinelearning (ML) is subfield of Artificial Intelligence (AI) that is built on the idea that systems an...
详细信息
Tracking and recording human activities have been a major interest in the iSpace, for this purpose different recognition and clustering techniques have been used, like using a learning Classifier System and data Minin...
详细信息
ISBN:
(纸本)9781424475629
Tracking and recording human activities have been a major interest in the iSpace, for this purpose different recognition and clustering techniques have been used, like using a learning Classifier System and datamining Techniques. These techniques share the common factor of database dependence and there was actually little effort into making the system to understand the way human were behaving in a given time in the space. Using Artificial Intelligence techniques, we present a work that reads and classifies user object activity
This paper presents a study on using a concept feature to detect web phishing problem. Following the features introduced in Carnegie Mellon Anti-phishing and Network Analysis Tool (CANTINA), we applied additional doma...
详细信息
ISBN:
(纸本)9780769539232
This paper presents a study on using a concept feature to detect web phishing problem. Following the features introduced in Carnegie Mellon Anti-phishing and Network Analysis Tool (CANTINA), we applied additional domain top-page similarity feature to a machinelearning based phishing detection system. We preliminarily experimented with a small set of 200 web data, consisting of 100 phishing webs and another 100 non-phishing webs. The evaluation result in terms of f-measure was up to 0.9250, with 7.50% of error rate.
Liver diseases have produced a big data such as metabolomics analyses, electronic health records, and report including patient medical information, and disorders. However, these data must be analyzed and integrated if...
详细信息
ISBN:
(纸本)9783319746906;9783319746890
Liver diseases have produced a big data such as metabolomics analyses, electronic health records, and report including patient medical information, and disorders. However, these data must be analyzed and integrated if they are to produce models about physiological mechanisms of pathogenesis. We use machinelearning based on classifier for big datasets in the fields of liver to Predict and therapeutic discovery. A dataset was developed with twenty three attributes that include the records of 7000 patients in which 5295 patients were male and rests were female. Support Vector machine (SVM), Boosted C5.0, and Naive Bayes (NB), datamining techniques are used with the proposed model for the prediction of liver diseases. The performance of these classifier techniques are evaluated with accuracy, sensitivity, specificity.
In mining massive datasets, often two of the most important and immediate problems are sampling and feature selection. Proper sampling and feature selection contributes to reducing the size of the dataset while obtain...
详细信息
ISBN:
(纸本)9781479941735
In mining massive datasets, often two of the most important and immediate problems are sampling and feature selection. Proper sampling and feature selection contributes to reducing the size of the dataset while obtaining satisfactory results in model building. Theoretically, therefore, it is interesting to investigate whether a given dataset possesses a critical feature dimension, or the minimum number of features that is required for a given learningmachine to achieve "satisfactory" performance. (Likewise, the critical sampling size problem concerns whether, for a given dataset, there is a minimum number of data points that must be included in any sample for a learningmachine to achieve satisfactory performance.) Here the specific meaning of "satisfactory" performance is to be defined by the user. This paper addresses the complexity of both problems in one general theoretical setting and shows that they have the same complexity and are highly intractable. Next, an empirical method is applied in an attempt to find the approximate critical feature dimension of datasets. It is demonstrated that, under generally reasonable assumptions pertaining to feature ranking algorithms, the critical feature dimension are successfully discovered by the empirical method for a number of datasets of various sizes. The results are encouraging in achieving significant feature size reduction and point to a promising way in dealing with big data. The significance of the existence of crucial dimension in datasets is also explained.
The evolution of modern approach in knowledge systems, decision support systems and clinical constraints estimation algorithms that formulate machinelearning, soft computing and datamining in presenting a new outloo...
详细信息
ISBN:
(纸本)9789380544199
The evolution of modern approach in knowledge systems, decision support systems and clinical constraints estimation algorithms that formulate machinelearning, soft computing and datamining in presenting a new outlook for health informatics domain. Health is then clearly understood as the essential part while describing a person's sense of well-being. The delivery of health care services therefore considers as higher proportion, and played an efficient role in information and communication technologies for its effective distribution mechanism. datamining in health informatics are developing into optimistic area for producing vision from diverse data set. datamining techniques are proved to he as a valuable resource for health care informatics. The main scope of writing this paper is to analyse the effectiveness of datamining techniques in health informatics and compare various techniques, approaches or methods and different tools used and its effect on the healthcare industry. The main motive of using datamining application in healthcare systems is to exploit a machine driven tool for identifying and circulating useful and relevant healthcare information.
Increasing trade volume adds up various challenges and risks for customs to maintain balance between trade facilitation and strong border control. With limited resources and manpower, it39;s quite difficult to have ...
详细信息
ISBN:
(纸本)9781538662274
Increasing trade volume adds up various challenges and risks for customs to maintain balance between trade facilitation and strong border control. With limited resources and manpower, it's quite difficult to have exhaustive physical examination of all import and export consignments. To balance control and facilitation Revised Kyoto Convention (RKC) and World Trade Organization (WTO) Trade Facilitation Agreement (TFA) have clearly stated about implementation of effective risk management system. In this paper, deep learning model was trained and tested to segregate high risk and low risk consignment on randomly selected 200,000 data from Nepal Customs of the year 2017. Model was tested using supervised learning utilizing inspection result provided by Nepal Customs. Deep learning has improved accuracy and seizure rate than that of decision Tree (DT) and Support Vector machine (SVM). All three methods have achieved a better result than current rule based risk management system. ANN had achieved better result than DT and SVM, by achieving 81% of seizure rate under 9% inspection.
This paper analyses the main research methods of power text mining technology in detail, and discusses the research hotspots of power text named entity recognition and named entity relationship extraction based on mac...
详细信息
Student success is important to institutions of learning. learning institutions offer academic support to their students to ensure success. Knowledge Tracing (KT), the task of modelling student knowledge based on thei...
详细信息
In learning search, it is assumed that a search algorithm has adequate opportunities to traverse the problem state space, and to sense and learn the topological (or probability) knowledge about the state space. In thi...
详细信息
ISBN:
(纸本)9781538643013
In learning search, it is assumed that a search algorithm has adequate opportunities to traverse the problem state space, and to sense and learn the topological (or probability) knowledge about the state space. In this paper, a new search algorithm based on Bayesian learning (algorithm BLS) is proposed, which finds the solution path for any given state pair (s, t) in the state space efficiently. The algorithm depicts the search as a process of pattern recognition for the current path. During its depth-first search, the BLS classifies the current path, depending on the probability distribution functions acquired in previous Bayesian learning, and prunes it if necessary. It is proved that the mean complexity of BLS grows linearly with the length of the solution path after sufficient training.
暂无评论