Extracting relevant information from legal documents is a challenging task due to the technical complexity and volume of their content. These factors also increase the costs of annotating large datasets, which are req...
详细信息
ISBN:
(纸本)9798400704314
Extracting relevant information from legal documents is a challenging task due to the technical complexity and volume of their content. These factors also increase the costs of annotating large datasets, which are required to train state-of-the-art summarization systems. To address these challenges, we introduce CivilSum, a collection of 23,350 legal case decisions from the Supreme Court of India and other Indian High Courts paired with human-written summaries. Compared to previous datasets such as IN-Abs, CivilSum not only has more legal decisions but also provides shorter and more abstractive summaries, thus offering a challenging benchmark for legal summarization. Unlike other domains such as news articles, our analysis shows the most important content tends to appear at the end of the documents. We measure the effect of this tail bias on summarization performance using strong architectures for long-document abstractivesummarization, and the results highlight the importance of long sequence modeling for the proposed task. CivilSum and related code are publicly available to the research community to advance textsummarization in the legal domain.
With the increase of knowledge, there is a need for summarization systems that will direct the person to the area they are interested in without any waste of time. In this work, Turkish news headlines have been predic...
详细信息
ISBN:
(纸本)9781728119045
With the increase of knowledge, there is a need for summarization systems that will direct the person to the area they are interested in without any waste of time. In this work, Turkish news headlines have been predicted by using encoder-decoder model from deep learning methods. Abstraction based textsummarization method has been used during the generation of headlines. The system has been trained with recurrent neural networks by developing encoder-decoder model. The word embeddings of the words in news texts have been generated by using Fasttext that is very commonly used model in the literature recently. The system has been tested separately by training the first sentence, first two sentences and full-text of each news. The success of the system is measured by ROUGE score and semantic similarity score. According to the experimental results, it has been observed the model trained with full-text of news outperforms among the other models.
abstractive text summarization helps people quickly obtain the key information of an article, and existing models generate fluent summaries but often suffer from factual consistency problems, a key issue that current ...
详细信息
.abstractive text summarization plays an important role in the field of natural language processing. However, the abstractivetext summary adopts deep learning research method to predict words often appears semantic i...
详细信息
.abstractive text summarization plays an important role in the field of natural language processing. However, the abstractivetext summary adopts deep learning research method to predict words often appears semantic inaccuracy and repetition and so on. at the present stage, in order to solve the problem that semantic inaccuracy, we propose an MS-Pointer Network that based on the multi-head self-attention mechanism, which a multi-head self-attention mechanism is introduced in the basic encoder-decoder model. Since multi-head self-attention can combine input words into the encoder-decoder arbitrarily, and given a higher weight of these words that combination of the semantics, thereby achieving the purpose of enhancing the semantic features of the text, so that the abstractivetext summary is more semantically structured, And the multi-head self-attention mechanism add the position information of the input text, which can enhance the semantic representation of the text. At the same time, in order to solve the problem of out of vocabulary, a pointer network is introduced on the seqtoseq with a multi-head attention mechanism. The model is referred to as MS-Pointer Network. We used CNN/Daily Mail and Gigaword datasets to validate our model, and uses the ROUGE metric to measure model. Experiments have shown that abstractivetext summaries generated using the multi-head self-attention mechanism outperforming current open state-of-the-art two points averagely.
A novel textsummarization framework referred to as Skip-Though Vector and Bi-encoder Based Automatic textsummarization (STV-BEATS) is proposed in this paper. STV-BEATS utilizes - (a) skip-though vector to generate s...
详细信息
A novel textsummarization framework referred to as Skip-Though Vector and Bi-encoder Based Automatic textsummarization (STV-BEATS) is proposed in this paper. STV-BEATS utilizes - (a) skip-though vector to generate sentence-based embedding;and (b) Long Short-Term Memory (LSTM) based deep autoencoder to reduce dimensions of skip thought vectors. STV-BEATS works in the conjunction of extractive and abstractivesummarization models to enhance the overall quality of the results. For each sentence, relevance and novelty metrics are calculated on the intermediate representation of the deep autoencoder to generate the final sentence score. The highly scored sentences are selected to generate an extractive summary. On the other hand, the abstractive part is composed of two encoders and a decoder which works as - (a) the first GRU-based bi-directional encoder and decoder work as basic sequence-to-sequence model on the extractive summary;and (b) the second GRU-based unidirectional encoder is used for fine encoding. Extensive computer experiments are conducted to determine the effectiveness of the STV-BEATS. Three standard benchmark datasets, namely, CNN/Daily Mail, DUC-2004, and DUC-2002 are used during experiments. Further, recall-oriented understudy for gisting evaluation (ROUGE) is used for validation of the STV-BEATS. Result reveals that the proposed STV-BEATS is capable of effective textsummarization and achieves substantially better results over the state-of-the-art models. (C)& nbsp;2022 Elsevier B.V. All rights reserved.
The application of neural networks in natural language processing, including abstractive text summarization, is increasingly attractive in recent years. However, teaching a neural network to generate a human-readable ...
详细信息
The application of neural networks in natural language processing, including abstractive text summarization, is increasingly attractive in recent years. However, teaching a neural network to generate a human-readable summary that reflects the core idea of the original source text (i.e., semantically similar) remains a challenging problem. In this paper, we explore using generative adversarial networks to solve this problem. The proposed model contains three components: a generator that encodes the long input text into a shorter representation;a discriminator to teach the generator to create human-readable summaries and another discriminator to restrict the output of the generator to reflect the core idea of the input text. The main training process can be carried out in an adversarial learning process. To solve the non-differentiable problem caused by the words sampling process, we use the policy gradient algorithm to optimize the generator. We evaluate the proposed model on the CNN/Daily Mail summarization task. The experimental results show that the model outperforms previous state-of-the-art models.
Bayesian Active Learning has had significant impact to various NLP problems, but nevertheless its application to textsummarization has been explored very little. We introduce Bayesian Active summarization (BAS), as a...
详细信息
Bayesian Active Learning has had significant impact to various NLP problems, but nevertheless its application to textsummarization has been explored very little. We introduce Bayesian Active summarization (BAS), as a method of combining active learning methods with stateof-the-art summarization models. Our findings suggest that BAS achieves better and more robust performance, compared to random selection, particularly for small and very small data annotation budgets. More specifically, applying BAS with a summarization model like PEGASUS we managed to reach 95% of the performance of the fully trained model, using less than 150 training samples. Furthermore, we have reduced standard deviation by 18% compared to the conventional random selection strategy. Using BAS we showcase it is possible to leverage large summarization models to effectively solve real-world problems with very limited annotated data.
summarization is flexible and allows the model to generate new words and phrases. However, the familiar words are more likely to be selected as abstract candidate words in the process of abstractivesummarization, cau...
详细信息
summarization is flexible and allows the model to generate new words and phrases. However, the familiar words are more likely to be selected as abstract candidate words in the process of abstractivesummarization, causing the generated abstract to diverge from the refer-ence. In our consideration, this is caused by representation degeneration of the pre-trained word embedding. Therefore, this paper proposes a general abstractivesummarization framework with dynamic word embedding representation correction (RepSum). The representation correction algorithm identifies the dimension most relevant to word frequency and eliminates the word frequency features. Then the distribution of word embeddings will be more even. As a result, the words will be selected as candidate words without frequency bias to improve the quality of the abstract. The experimental results illustrate that RepSum performs better than the benchmark model in summary quality, demonstrating our method's effectiveness.
This age of data-driven innovation has made automated relevant and important data extraction a necessity. Automated textsummarization has made it possible to extract relevant information from large amounts of data wi...
详细信息
ISBN:
(数字)9781728132457
ISBN:
(纸本)9781728132464
This age of data-driven innovation has made automated relevant and important data extraction a necessity. Automated textsummarization has made it possible to extract relevant information from large amounts of data without needing any supervision. But the extracted information could seem artificial at times and that's where the abstractivesummarization method tries to mimic the human way of summarizing by creating coherent summaries using novel words and sentences. Due to the difficult nature of this method, before deep learning, there hasn't been much progress. So, during this work, we have proposed an attention mechanism-based sequence-to-sequence network to generate abstractive summaries of Bengali text. We have also built our own large Bengali news dataset and applied our model on it to show indeed deep sequence-to-sequence neural networks can achieve good performance summarizing Bengali texts.
暂无评论