Fractal represents a special feature of nature and functional objects. However, fractal based computing can be applied to many research domains because of its fixed property resisted deformation, variable parameters a...
详细信息
As the number of patent applications increases yearly, the negation relation between patents has become intertwined, which makes it difficult for constructing negation relation in patent examination manually. Therefor...
详细信息
Generated method of transcriptional regulatory networks remains an important research in biology. Many approaches have been proposed to construct transcriptional regulatory networks. However, with the increase of ChIP...
详细信息
With the intense competition of global intellectual property, the increasing patents promote the potentiality of patent transactions. Patent valuation is the premise of the patent transaction. Automatic patent valuati...
详细信息
ISBN:
(数字)9781728169262
ISBN:
(纸本)9781728169279
With the intense competition of global intellectual property, the increasing patents promote the potentiality of patent transactions. Patent valuation is the premise of the patent transaction. Automatic patent valuation faces some challenging issues from valuation feature to valuation model. To solve the above issues, we propose a Bayesian graph convolutional neural network based patent valuation model. In the model, the valuation objects are defined, from which to some valuation features are extracted. Valuation scenario is the constructed, on which Bayesian graph convolutional neural network is used to generate patent value. We evaluate our model by comparing the state-of-the-art model on patent data sets. The results show that our model outperforms other models in the evaluation measurements.
作者:
Hanlin LiuJiantao ZhouHua LiCollege of Computer Science
Inner Mongolia University Hohhot China Inner Mongolia Key Laboratory of Discipline Inspection and Supervision Big Data
Inner Mongolia Engineering Laboratory for Big Data Analysis Technology Ministry of Education Inner Mongolia Engineering Laboratory for Cloud Computing and Service Software Inner Mongolia Key Laboratory of Social Computing and Data Processing National & Local Joint Engineering Research Center of Intelligent Information Processing Technology for Mongolian Engineering Research Center of Ecological Big Data Hohhot China
Anomaly detection refers to the identification of data objects that deviate from the general data distribution. One of the important challenges in anomaly detection is handling high-dimensional data, especially when i...
Anomaly detection refers to the identification of data objects that deviate from the general data distribution. One of the important challenges in anomaly detection is handling high-dimensional data, especially when it contains a large number of redundant attributes. Although traditional Isolation Forest (iForest) and Extended Isolation Forest (EIF) have performed well in anomaly detection, the detection accuracy is still extremely influenced by redundant attributes. In this paper, we propose a two-tier architecture named RS-EIF, to alleviate the negative impact of redundant attributes. Specifically, the RS-EIF combines the rough set theory and EIF algorithm and provides an effective basis for the EIF-based anomaly detection method by removing redundant attributes with rough set theory. Experimental results show that the RS-EIF algorithm performs better than the iForest and EIF algorithms in terms of anomaly detection rate, and the results of the combination of rough set theory and EIF are more prominent compared to the other combination methods.
Knowledge Graph (KG) is an essential research direction that involves storing and managing knowledge data, but its incompleteness and sparsity hinder its development in various applications. Knowledge Graph Reasoning ...
Knowledge Graph (KG) is an essential research direction that involves storing and managing knowledge data, but its incompleteness and sparsity hinder its development in various applications. Knowledge Graph Reasoning (KGR) is an effective method to solve this limitation via reasoning missing knowledge based on existing knowledge. The graph Convolution Network (GCN) based method is one of the state-of-the-art approaches to this work. However, there are still some problems, such as the insufficient ability to perceive graph structure and the poor effect of learning data features which may limit the reasoning accuracy. This paper proposes a KGR architecture based on a graph sequence generator and multi-head self-attention mechanism, named GaM-KGR, to improve the above problems and enhance prediction accuracy. Specifically, the GaM-KGR first introduces the graph generation model into the field of KGR for graph representation learning to obtain the hidden features in the data so that enhancing the reasoning effect and then embeds the generated graph sequence into the multi-head self-attention mechanism for subsequent processing to improve the graph structure perception ability of the proposed architecture, so that it can process the graph structure data more appropriately. Extensive experimental results show that the GaM-KGR architecture can achieve the state-of-the-art prediction results of current GCN-based models.
Knowledge Graph is an important research field that involves the storage and management of knowledge, but the incompleteness and sparsity of Knowledge Graphs hinder their application in many fields. Knowledge Graph Re...
详细信息
Knowledge Graph is an important research field that involves the storage and management of knowledge, but the incompleteness and sparsity of Knowledge Graphs hinder their application in many fields. Knowledge Graph Reasoning aims to alleviate this problem by completing missing paths or identifying wrong paths between entities. Graph Convolution Network (GCN) based methods are one of the state-of-the-art approaches to this work. However, it is difficult to directly generalize to unknown nodes and utilizes valid information from the local neighborhood which results in poor flexibility and extensibility and will loss of important information. This paper presents EG-KGR, a plug-and-play knowledge reasoning model based on enhanced graph sampling and aggregate inductive learning algorithm to relieve the above problems and enhance existing GCN-based methods. Specifically, EG-KGR supports incremental characteristics, uses inductive learning to replace transductive learning, and designs random sampling and local information sampling optimization methods to improve the model's generalization ability, prediction accuracy, and running speed. Extensive experimental results show that our EG-KGR can achieve optimal results.
作者:
Feiyu WangJian-tao ZhouCollege of Computer Science
Inner Mongolia University Hohhot Inner Mongolia China Inner Mongolia Engineering Laboratory for Cloud Computing and Service Software
Inner Mongolia Key Laboratory of Social Computing and Data Processing Inner Mongolia Engineering Laboratory for Big Data Analysis Technology Engineering Research Center of Ecological Big Data Ministry of Education National & Local Joint Engineering Research Center of Intelligent Information Processing Technology for Mongolian China
Cloud storage services have been used by most businesses and individual users. However, data loss, service interruptions and cyber attacks often lead to cloud storage services not being provided properly, and these in...
Cloud storage services have been used by most businesses and individual users. However, data loss, service interruptions and cyber attacks often lead to cloud storage services not being provided properly, and these incidents have caused financial losses to users. Second, traditional and single-cloud model disaster recovery services are no longer suitable for the current complex cloud storage systems. Therefore, a scheme to provide disaster recovery for cloud storage services in a multi-cloud storage environment is needed in real production. In this paper, we propose a disaster recovery scheme based on blockchain technology. The proposed scheme outlined in this study aims to address the issue of data availability within the cloud storage landscape. The proposed scheme achieves this goal by dividing data into hot and cold categories, verifying the integrity of copy data via blockchain technology, and utilizing blockchain networks to manage multi-cloud storage systems. Experimental findings demonstrate that the proposed scheme yields superior results in terms of computation and time overheads.
In a multi-cloud storage system, provenance data records all operations and ownership during its lifecycle, which is critical for data security and audibility. However, recording provenance data also poses some challe...
In a multi-cloud storage system, provenance data records all operations and ownership during its lifecycle, which is critical for data security and audibility. However, recording provenance data also poses some challenging security and storage issues. In this paper, we present a secure and efficient multi-cloud storage data source scheme, BMDP. We use blockchain technology to ensure the secure storage of provenance data and design a smart contract to utilize the provenance data to ensure the proper operation of the multi-cloud storage system. Finally, we analyze the scheme’s safety and do simulation experiments to show that the scheme has practicality.
作者:
Junchao ChenJian-tao ZhouXinyu HaoInner Mongolia Engineering Laboratory for Cloud Computing and Service Software
Inner Mongolia Key Laboratory of Social Computing and Data Processing Inner Mongolia Engineering Laboratory for Big Data Analysis Technology College of Computer Science Inner Mongolia University Engineering Research Center of Ecological Big Data Ministry of Education National & Local Joint Engineering Research Center of Intelligent Information Processing Technology for Mongolian Hohhot China
The accuracy of data analysis depends on data quality, and addressing data consistency issues is a key challenge to improve it. Constant Conditional Functional Dependency (CCFD) is an effective approach that ensures d...
详细信息
ISBN:
(数字)9798350376968
ISBN:
(纸本)9798350376975
The accuracy of data analysis depends on data quality, and addressing data consistency issues is a key challenge to improve it. Constant Conditional Functional Dependency (CCFD) is an effective approach that ensures data consistency by enforcing bindings of semantically related values, thus providing quality assurance for data analysis and decision-making processes. However, with the growth of data scale, especially the increasing number of data tuples and attributes, existing single-machine CCFD discovery algorithms face issues of low computational efficiency and lengthy computation time. This paper proposes a time-efficient distributed CCFD discovery algorithm (DCCFD). Through the optimization of data preprocessing and index mapping, the data organization structure is enhanced, laying the foundation for the discovery of CCFDs under distributed conditions. The Spark parallel computing framework is used to partition the dataset, which accelerates the parallel loading and processing of data. Additionally, this algorithm ensures accuracy and processing speed when discovering dependencies by efficiently generating frequent itemsets and verifying CCFDs in parallel. Experiments on multiple real datasets show that, especially with the complex Airline dataset, the DCCFD algorithm not only accurately discovers CCFDs, but also reduces the average running time by 75.64% compared with the preCFDMiner algorithm.
暂无评论