Fully capturing contextual information and analyzing the association between entity semantics and type is helpful for joint extraction task: 1) The context can reflect the part of speech and semantics of entity. 2) Th...
详细信息
ISBN:
(纸本)9781450385053
Fully capturing contextual information and analyzing the association between entity semantics and type is helpful for joint extraction task: 1) The context can reflect the part of speech and semantics of entity. 2) The entity type is closely related to the relation between entities. Previous research used to simply embed the contextual information into shallow layer of the model, ignoring the association between entity semantics and type. In this paper, we propose a graph network with full-information modeling to explicitly model different-level information in the text. The contextual information of entity is dynamically embedded in each span representation to improve the reasoning ability. To capture the fine-grained association between the semantics and type of entity, the graph network uses the feature of entity types to generate edge information between different nodes. Experimental results show that our model outperforms previous models on the CoNLL04 dataset and obtains competitive results on the SciERC dataset in both entity recognition and relation extraction. Extensive additional experiments further verify the effectiveness of the model.
Payload anomaly detection can discover malicious behaviors hidden in network packets. It is hard to handle payload due to its various possible characters and complex semantic context, and thus identifying abnormal pay...
详细信息
ISBN:
(纸本)9781665421263
Payload anomaly detection can discover malicious behaviors hidden in network packets. It is hard to handle payload due to its various possible characters and complex semantic context, and thus identifying abnormal payload is also a non-trivial task. Prior art only uses the n-gram language model to extract features, which directly leads to ultra-high-dimensional feature space and also fails to capture the context semantics fully. Accordingly, this paper proposes a word embedding-based context-sensitive network flow payload anomaly detection method (termed WECAD). First, WECAD obtains the initial feature representation of the payload through the word embedding-based method. Then, we propose a corpus pruning algorithm, which applies the cosine similarity clustering and frequency distribution to prune inconsequential characters. We only keep the essential characters to reduce the calculation space. Subsequently, we propose a context learning algorithm. It employs the co-occurrence matrix transformation technology and introduces the backward step size to consider the order relationship of essential characters. Comprehensive experiments on real-world intrusion detection datasets validate the effectiveness of our method.
Network anomaly detection is important for detecting and reacting to the presence of network attacks. In this paper, we propose a novel method to effectively leverage the features in detecting network anomalies, named...
详细信息
Network anomaly detection is important for detecting and reacting to the presence of network attacks. In this paper, we propose a novel method to effectively leverage the features in detecting network anomalies, named FDEn, consisting of flow-based Feature Derivation (FD) and prior knowledge incorporated Ensemble models (En pk). To mine the effective information in features, 149 features are derived to enrich the feature set of the original data with covering more characteristics of network traffic. To leverage these features effectively, an ensemble model En pk, including CatBoost and XGBoost, based on the bagging strategy is proposed to first detect anomalies by combining numerical features and categorical features. And then, En pk adjusts the predicted label of specific data by incorporating the prior knowledge of network security. We conduct empirically experiments on the data set provided by the Network Anomaly Detection Challenge (NADC), in which we obtain average improvement up to 61.6%, 31.7%, 50.2%, and 45.0%, in terms of the cost score, precision, recall and F1-score, respectively.
Isolation forest (iForest) has been emerging as arguably the most popular anomaly detector in recent years due to its general effectiveness across different benchmarks and strong scalability. Nevertheless, its linear ...
详细信息
With the increasing adoption of graph neural networks (GNNs) in the graph-based deep learning community, various graph programming frameworks and models have been developed to improve the productivity of GNNs. The cur...
详细信息
ISBN:
(纸本)9781665443326
With the increasing adoption of graph neural networks (GNNs) in the graph-based deep learning community, various graph programming frameworks and models have been developed to improve the productivity of GNNs. The current GNN frameworks choose GPU as an essential tool to accelerate GNN training. However, it is still challenging to train GNNs on large graphs with limited GPU memory. Unlike traditional neural networks, generating mini-batch data by sampling in GNNs requires some complicated tasks such as traversing the graph to select neighboring nodes and gathering their features. This process takes up most of the training and we find the main bottleneck comes from transferring nodes features from CPU to GPU through limited bandwidth. In this paper, We propose a method Reusing Batch Data for the problem of data transmission. This method utilizes the similarity between adjacent mini-batches to reduce repeated data transmission from CPU to GPU. Furthermore, to reduce the overhead introduced by this method, we design a fast algorithm based on GPU to detect repeated nodes’ data and achieve shorter additional computation time. Evaluations on three representative GNN models show that our method can reduce transmission time by up to 60% and speed the end-to-end GNN training by up to 1.79× over the state-ofthe-art baselines. Besides, Reusing Batch Data can effectively save GPU memory footprint by about 19% to 40% while still reducing the training time compared to the static cache strategy.
SNNs have achieved great attention in recent years as they contain neurons more like those in the brain and use spikes to encode and transmit information efficiently among neurons with lower energy consumption. A Reti...
详细信息
ISBN:
(纸本)9781450385893
SNNs have achieved great attention in recent years as they contain neurons more like those in the brain and use spikes to encode and transmit information efficiently among neurons with lower energy consumption. A Retina-LGN-V1 structure-like spiking neuron network (RLVSL-SNN) is proposed in this paper. It is inspired by the structure of mammalian primary visual pathway, and simulates different biological structures of Retina, LGN and V1. Noise reduction circuit and light adaptation circuit are also simulated for enhancing the robustness of its extracted features. RLVSL-SNN is a bio-plausible neuron network as it has firing rates of neurons in each layer that are similar to those of biological experiments. Besides, a full-connected SNN (FC SNN) is implemented following RLVSL-SNN for classification to evaluate the extracted features. The additive spiking timing dependent plasticity (STDP) learning rules and the ANN-to-SNN conversion method are utilized to train RLVSL-SNN and FC SNN, respectively. The experiments on MNIST dataset have verified that RLVSL-SNN is comparable to AlexNet for classification by features from spikes.
The implicitly coupled pressure-based algorithm is widely acknowledged for its superior convergence and robustness in solving incompressible flow problems. However, the increased expansion scale of equations and diffi...
详细信息
In JointCloud Computing, multi-party participation introduces complexity and uncertainty. For all participants in JointCloud Computing, both continuous supervision and necessary privacy protection are required. Tradit...
详细信息
In JointCloud Computing, multi-party participation introduces complexity and uncertainty. For all participants in JointCloud Computing, both continuous supervision and necessary privacy protection are required. Traditional supervision methods usually adopt the centralized information interaction mode, which has such defects as collusion of interests, single point of failure, privacy disclosure, etc. Building a decentralized supervision mechanism has become a new research direction. In this paper, we propose PPSS, a privacy-preserving supervision scheme based on blockchain, which decentralizes the supervision of the participants in JointCloud Computing, and combines the “double encryptions” and “threshold encryption” technologies to provide privacy protection. While making full use of the decentralization of the blockchain, a committee is established to carry out the analysis and decision-making tasks in terms of supervision and privacy protection. Experimental results indicate that PPSS can balance performance and security by reasonably configuring the committee.
In this paper, we present OpenMedIA, an open-source toolbox library containing a rich set of deep learning methods for medical image analysis under heterogeneous Artificial Intelligence (AI) computing platforms. Vario...
详细信息
DRAM failure prediction is a vital task in AIOps, which is crucial to maintain the reliability and sustainable service of large-scale data centers. However, limited work has been done on DRAM failure prediction mainly...
详细信息
暂无评论