Cloud storage is quickly becoming the norm in today's online infrastructure because of the low administration costs and convenient accessibility. The cloud storage system offers crucial advantages by letting cloud...
详细信息
Cloud storage is quickly becoming the norm in today's online infrastructure because of the low administration costs and convenient accessibility. The cloud storage system offers crucial advantages by letting cloud users keep their massive amounts of data in the cloud. Better performance during data retrieval necessitates attending to the privacy and security of data stored in the cloud. The necessary information can be obtained from the cloud using the current improved semantic feature abstraction method E-TFIDF (Enhanced-Term Frequency-Inverse Document Frequency). There was still no solution for the problem of secure cloud-based data access. Current method SEDFS (Searchable Encrypted data File Sharing) system applied along with limited and precise keyword quest; indexing procedure necessary for improvement. The research focuses on SHE (Secured Hybrid Search), a method for conducting encrypted multi-keyword searches on cloud platforms. Using a category-based method, an index list is built for each document depending on its classification. Clearly, when compared to both the current system and the adaptively secure SSE technique, it has been found that the proposed approach significantly improves encryption and decryption throughput in the shortest possible time. In addition, the suggested algorithm shortens the amount of time required to index and conduct a search on a variety of documents. There was also a decrease in the amount of time spent searching and indexing, and an increase in the efficiency with which keywords could be extracted.
The advent of COVID-19 highlights the need for big data-driven medical applications, the Internet of Medical Things, and smart healthcare. The biological information collected is strictly private. This enormous quanti...
The advent of COVID-19 highlights the need for big data-driven medical applications, the Internet of Medical Things, and smart healthcare. The biological information collected is strictly private. This enormous quantity of biological information is, however, beyond the capacity of current health care systems. Therefore, cloud computing has become the norm for archiving and sharing data. The combined information is then put to use in a variety of ways, including research and the identification of previously unknown facts. Textual forms (such as test results, prescriptions, and diagnoses) are the norm for biological data. Unfortunately, there are a number of security dangers and assaults that may be made against such data, including infringements on privacy and confidentiality. The security of biological data has come a long way, yet the majority of current methods still cause considerable delays and cannot support real-time replies. To improve the healthcare system, this study suggests a unique fog-enabled blockchain based privacy-preserving approach that makes use of machine learning. The suggested model efficiently carries out Medical Entity Recognition as it is built on four different machine learning models. In terms of detection rate and accuracy, the suggested model is 97% and 98%, respectively, better than the state-of-the-art replicas, as shown by the experiments. The sanitization approach outdoes the state-of-the-art by 8 percentage points when it comes to preserving utilities.
The advancement of research in the field of natural language processing has made peoples’ daily lives much easier, with numerous applications at the disposal. The traditional methods like the Hidden Markov Model, the...
详细信息
The advancement of research in the field of natural language processing has made peoples’ daily lives much easier, with numerous applications at the disposal. The traditional methods like the Hidden Markov Model, the CRF classifier, the Naive Bayes classifier, and others are being replaced by neural networks in recent times. However, most of these methods work considerably well only with huge amounts of training data and, hence are not suitable for languages that are poor in terms of trainable resources. The challenge is to make the system work considerably well with minimal training. This paper presents research work to understand the effect of training size for part of speech tagging, which is one of the preliminary tasks for any NLP application. Experiments are conducted to understand the training size required for standard techniques to perform with high accuracy. The results of the experiments conducted for English and Assamese are presented in this paper.
In recent years, with the rapid development of computer technology, people have put forward higher requirements for data mining tools and methods. This paper mainly studied multi-objective optimization problems in the...
In recent years, with the rapid development of computer technology, people have put forward higher requirements for data mining tools and methods. This paper mainly studied multi-objective optimization problems in the field of construction engineering. By establishing models, information such as building structures and buildings was modeled and processed, and influencing factors were analyzed and predicted. After being quantified, the examples were used to illustrate the application effect, and the model was optimized. The test results showed that the dataprocessing efficiency of the multi-objective optimization model for construction engineering based on the data mining association rule algorithm was around 74% to 77% before optimization, and the dataprocessing efficiency after optimization was between 80% to 90%. At the same time, appropriate index parameter values were selected based on the actual situation to maximize the synergy between engineering projects and optimize economic benefits, thereby achieving the goal of improving construction efficiency and quality.
Anomaly detection is the identification of items, events, or observed values that do not meet the expected situation or other conditions in the data set. In this paper, an anomaly detection algorithm of IoT (Internet ...
Anomaly detection is the identification of items, events, or observed values that do not meet the expected situation or other conditions in the data set. In this paper, an anomaly detection algorithm of IoT (Internet of Things) intelligent data acquisition terminal based on deep learning is proposed. In the case design, the Text-CNN (Text-convolutional Neural Network) is adopted as the algorithm of supervised detection scheme. Split the collected IoT data set to obtain network traffic data of a plurality of different devices, and label these data in categories; Input the data into the established Text-CNN algorithm model through NLP; In the sample category output module, train the model, save the trained model, test it, and output the sample category. The test results show that the algorithm proposed in this chapter is superior to other literature algorithms in anomaly detection, and it is suitable for anomaly detection of complex data, especially high-dimensional data.
Efficient and accurate recording of electronic prescription data is of paramount importance in the ever-evolving field of healthcare. As this investigation delves deeper, a groundbreaking solution emerges, adept at br...
Efficient and accurate recording of electronic prescription data is of paramount importance in the ever-evolving field of healthcare. As this investigation delves deeper, a groundbreaking solution emerges, adept at bridging the speed gap between traditional handwritten methods and the digital landscape. This innovative technique is founded on a profound understanding of the cognitive processes underpinning manual prescription creation, resulting in a significant overhaul of the digital interface. This approach revolves around the integration of predictive algorithms and tailored templates into the electronic prescription system, with the goal of matching the speed of traditional handwritten methods. This substantial enhancement not only streamlines workflow but also enhances patient care, optimizes pharmacy operations, and provides valuable data for research, promising a transformative impact on the healthcare industry.
The exponential growth of healthcare data poses significant challenges for clinical researchers who strive to identify meaningful patterns and correlations. The complexity of this data arises from its high dimensional...
详细信息
As the most important parts, bearings are widely used in many types of mechanical and electronic equipment. Bearing failure often leads to serious consequences. Therefore, the fault detection of bearings is the key to...
详细信息
In the last decade, Twitter data has become one of the most valuable research sources for many areas including health, marketing, security, and politics. Researchers prefer Twitter data since it is completely public a...
In the last decade, Twitter data has become one of the most valuable research sources for many areas including health, marketing, security, and politics. Researchers prefer Twitter data since it is completely public and can be easily downloaded using Twitter APIs. The recent intensive use of Twitter data makes it difficult for researchers to follow or analyze its research. In this paper, we summarize most of the predictable patterns, aspects, and attitudes from Twitter data and analyze the performance and feasibility of the algorithms used. Moreover, we describe the current popular Twitter datasets used in various domains and applications. Current challenges and research gaps are discussed, and some recommendations are given for future works from different perspectives.
A literature review is an essential part of research. Beginning researchers who would like to conduct research in any field commonly review previous papers to identify trends and gaps in research. However, conducting ...
A literature review is an essential part of research. Beginning researchers who would like to conduct research in any field commonly review previous papers to identify trends and gaps in research. However, conducting a literature review is challenging, particularly for novice researchers. First, synthesising previous related works can be time-consuming. Second, accessing expensive databases such as Scopus and Web of Science might be inaccessible for novice researchers in developing countries. Thus, the purpose of this paper is to present a review performed by data crawling ArXiv and using latent Dirichlet allocation (LDA), a topic modelling algorithm. In this paper, we provide an overview of the topic models of natural disasters and artificial intelligence. Papers from 2017–2022 were crawled using Python with specific keywords, and only papers written in English were eligible for further analysis. Then, the papers were pre-processed by lowering the text, removing punctuation and stop words, tokenising, and lemmatising with POS tagging techniques. Pre-processed papers were then trained using the LDA algorithm. For evaluation, coherence metrics were used to determine the optimal number of topics. There were five topics resulting from the topic modelling algorithm. The clusters were qualitatively analysed and labelled as image classification, algorithms, social media analysis, natural language processing, and machine learning. Further findings are also discussed.
暂无评论