We propose an extension with deterministically timed multiactions of discrete time stochastic and immediate Petri box calculus (dtsiPBC), previously presented by I.V. Tarasyuk, H. Macià and V. Valero. In dtsdPBC,...
详细信息
This study compares Chinese news classification results of machine learning (ML) and deep learning (DL). In processing ML, we chose Support Vector Machine (SVM) and Naive Bayes (NB) to form three models: Word2Vec-SVM,...
详细信息
ISBN:
(纸本)9781728128177
This study compares Chinese news classification results of machine learning (ML) and deep learning (DL). In processing ML, we chose Support Vector Machine (SVM) and Naive Bayes (NB) to form three models: Word2Vec-SVM, TFIDF-SVM, and TFIDF-NB. Since NB assumes that the words are independent, this is different from the concept of related word distribution in Word2Vec, so the combination with NB is excluded. In processing DL, we adopted Bidirectional Long Short-Term Memory (Bi-LSTM), Long Short-Term Memory (LSTM) and Convolutional Neural Network (CNN), and used Word2Vec for word embedding. Experimental results showed that with proper word preprocessing, the difference of classification accuracy of ML and DL models is actually very small. Although the results show that Bi-LSTM performs the most accurate and has the lowest Loss compared to other DL techniques, its implementation process is the most time consuming. This study affirms the excellent results of CNN, while its Loss is the highest of the DL models. We also found that Word2Vec-SVM was superior to TFIDF-SVM in terms of efficiency, but its accuracy is not as good as expected. To summarize the classification accuracy in Bi-LSTM, LSTM, CNN, Word2vec-SVM, TFIDF-SVM, and NB are 89.3%, 88%, and 87.54%, 85.32%, 87.35%, 86.56%, respectively.
Data plays the most significant role to attain efficiency in performing a task using Machine Learning (ML) techniques. Metadata (MD) represents data of data. MD extraction and data attribute selection play a vital rol...
详细信息
ISBN:
(纸本)9781728148274
Data plays the most significant role to attain efficiency in performing a task using Machine Learning (ML) techniques. Metadata (MD) represents data of data. MD extraction and data attribute selection play a vital role in defining the performance of ML models. The study in this paper focuses on the role of MD, data attributes and data models that define the learning capability of ML to evolve with human-like capability to learn and draw inferences. To evolve with such artificially intelligent autonomous systems, the study in this paper is a preliminary step towards applying ML techniques on textual data for performing syntactic analysis, further to evolve with semantic and behavioral analysis. Based on the rigorous survey study and observations, this paper concludes with the description of the parameters to quantify the performance of ML model which are essential to define the performance characteristics of ML. The increased deployment of ML is observed in the recent Artificial Intelligence arena, and hence the study contributes towards evolving performance parameters in applications that employ ML techniquestextbf.
Every programming language has its own attributes, advantages and its own syntax. The logical reasoning applied by the programmer also requires awareness of the syntax specific to that language. Writing correct code d...
详细信息
ISBN:
(纸本)9781728136950
Every programming language has its own attributes, advantages and its own syntax. The logical reasoning applied by the programmer also requires awareness of the syntax specific to that language. Writing correct code depends heavily on syntax proficiency. Perhaps this can be asserted as an impediment of general programminglanguages. The new learners of a particular programming language find it difficult to cope up with the syntax requirement of that particular language. This not only increases the time required to learn a language but also shifts the focus of the user from logical reasoning. In order to shun the tedious approach of learning the syntax of a language, the approach of converting the user's logic drafted in natural language directly into the appropriate programming syntax can be used. This approach will not only grant the user the ability to use natural language but will eliminate the syntax dependency as well. Since logic construction for a solution to a problem is constrained by the syntax of a programming language, we propose a system that allows the user to provide a simple English statement as input to the system, which will then be translated into syntactically correct *** show that the system works efficiently with more than 80% accuracy. With each iteration, the dataset gets trained and updated, further increasing the precision and recall of the system. We also convey that, with the help of this system the syntax dependency can be eliminated, thereby increasing the user's efficiency.
programminglanguages are emerging as a challenging and interesting domain for machine learning. A core task, which has received significant attention in recent years, is building generative models of source code. How...
详细信息
The most prominent machine learning (ML) methods in use today are supervised, meaning they require ground-truth labeling of the data on which they are trained. Annotating data is arduous and expensive. Additionally, d...
详细信息
ISBN:
(数字)9781728125060
ISBN:
(纸本)9781728125077
The most prominent machine learning (ML) methods in use today are supervised, meaning they require ground-truth labeling of the data on which they are trained. Annotating data is arduous and expensive. Additionally, data sets for image object detection may be annotated by drawing polygons, drawing bounding boxes, or providing single points on targets. Selection of annotation technique is a tradeoff between time to annotate and accuracy of the annotation. When annotating a dataset for machine object recognition algorithms, researchers may not know the most advantageous method of annotation for their experiments. This paper evaluates the performance tradeoffs of three alternative methods of annotating imagery for use in ML. A neural network was trained using the different types of annotations and compares the detection accuracy of and differences between the resultant models. In addition to the accuracy, cost is analyzed for each of the models and respective datasets.
In case of no fixed infrastructure (military applications and emergency rescue operations) and we need to build a network with low cost, Wireless sensor networks (WSNs) are useful. We have no fixed routing protocol, o...
详细信息
ISBN:
(纸本)9781728139593
In case of no fixed infrastructure (military applications and emergency rescue operations) and we need to build a network with low cost, Wireless sensor networks (WSNs) are useful. We have no fixed routing protocol, or intrusion detection technique available for them because WSNs are dynamic in nature and individual nodes of the network are required for this to be done. Nodes are mobile in most of the applications of WSNs, so they depend on battery power and availability of limited resources which shows that power consumption is an effective research area for performing a set of tasks in WSNs. To deal with such an issue, machine learning (ML) techniques (self-learning algorithms, working without programming or human intervention) can be applied effectively according to the application requirement. In this paper, we have done comparative about several ML-based techniques for WSNs. In addition, we also analyzed ML techniques for clustering and energy harvesting. At the end, we present a summary of ML techniques for both clustering and energy harvesting with some open issues.
Code summarization, aiming to generate succinct natural language description of source code, is extremely useful for code search and code comprehension. It has played an important role in software maintenance and evol...
详细信息
ISBN:
(纸本)9780999241127
Code summarization, aiming to generate succinct natural language description of source code, is extremely useful for code search and code comprehension. It has played an important role in software maintenance and evolution. Previous approaches generate summaries by retrieving summaries from similar code snippets. However, these approaches heavily rely on whether similar code snippets can be retrieved, how similar the snippets are, and fail to capture the API knowledge in the source code, which carries vital information about the functionality of the source code. In this paper, we propose a novel approach, named TL-CodeSum, which successfully uses API knowledge learned in a different but related task to code summarization. Experiments on large-scale real-world industry Java projects indicate that our approach is effective and outperforms the state-of-the-art in code summarization.
The practical success of widely used machine learning (ML) and deep learning (DL) algorithms in Artificial Intelligence (AI) community owes to availability of large datasets for training and huge computational resourc...
详细信息
ISBN:
(纸本)9781728136950
The practical success of widely used machine learning (ML) and deep learning (DL) algorithms in Artificial Intelligence (AI) community owes to availability of large datasets for training and huge computational resources. Despite the enormous practical success of AI, these algorithms are only loosely inspired from the biological brain and do not mimic any of the fundamental properties of neurons in the brain, one such property being the chaotic firing of biological neurons. This motivates us to develop a novel neuronal architecture where the individual neurons are intrinsically chaotic in nature. By making use of the topological transitivity property of chaos, our neuronal network is able to perform classification tasks with very less number of training samples. For the MNIST dataset, with as low as 0.1% of the total training data, our method outperforms ML and matches DL in classification accuracy for up to 7 training samples/class. For the Iris dataset, our accuracy is comparable with ML algorithms, and even with just two training samples/class, we report an accuracy as high as 95.8%. This work highlights the effectiveness of chaos and its properties for learning and paves the way for chaos-inspired neuronal architectures by closely mimicking the chaotic nature of neurons in the brain.
Feature extraction has a vibrant part in Machine learning (ML) to identify the data patterns with optimum accuracy. We proposed some significant features to predict the student's institution or university based on...
详细信息
ISBN:
(数字)9781728132457
ISBN:
(纸本)9781728132464
Feature extraction has a vibrant part in Machine learning (ML) to identify the data patterns with optimum accuracy. We proposed some significant features to predict the student's institution or university based on their answers in the technological survey. Four experiments were performed in IBM SPSS Modeler version 18.2 using 4 ML to resolve the binary classification problem. In the university prediction problem., the uppermost accuracy of 94.26% is provided by eXtreme Gradient Boosting Tree (XGBT) and suggested 18 significant features out of a total of 37. Further., the Artificial Neural Network (ANN) with boosting scored second maximum accuracy of 93.96% and recommended 10 significant features; Support Vector Machine (SVM) provided third-highest accuracy of 92.45% with the recommendation of 12 features; and Random Tree (RT) attained the least accuracy 92.15% with recommendation of 10 important features. The findings of the paper conclude that the XGBT classifier outperformed others in prediction. Also., a noteworthy dissimilarity was found between XGBT's accuracy and SVM's accuracy., RT's accuracy.
暂无评论