This paper presents findings from the Check-That! Lab Task 1B-English submission at CLEF 2023. The research developed a method for evaluating the check-worthiness of short English texts. The first iteration focused on...
ISBN:
(纸本)9783031717352;9783031717369
This paper presents findings from the Check-That! Lab Task 1B-English submission at CLEF 2023. The research developed a method for evaluating the check-worthiness of short English texts. The first iteration focused on identifying optimal model architectures and adaptation techniques, while the second iteration involved curating the dataset for improved results. The study included fine-tuning several GPT and BERT models, applying zero-shot, few-shot, and Chain-of-Thought prompting strategies, and utilizing dataset sampling techniques informed by quality and training dynamics metrics. Team achieved first place in the competition by fine-tuning the OpenAI GPT-3 curie model. Findings suggest that fine-tuned BERT models can perform comparably to GPT models, but dataset curation was pivotal in obtaining superior results across various model architectures.
In this paper, we introduce and study mathematical programming formulations for the Least Cost Directed Perfect Awareness Problem (LDPAP), an NP-hard optimization problem that arises in the context of influence market...
ISBN:
(数字)9783031605994
ISBN:
(纸本)9783031606014;9783031605994
In this paper, we introduce and study mathematical programming formulations for the Least Cost Directed Perfect Awareness Problem (LDPAP), an NP-hard optimization problem that arises in the context of influence marketing. In the LDPAP, we seek to identify influential members of a given social network that can disseminate a piece of information and trigger its propagation throughout the network. The objective is to minimize the cost of recruiting the initial spreaders while ensuring that the information reaches everyone. This problem has been previously modeled as two different integer programming formulations that were tested on a collection of 300 small synthetic instances. In this work, we propose two new integer programming models and three constraint programming formulations for the LDPAP. We also present preprocessing techniques capable of significantly reducing the sizes of these models. To investigate and compare the efficiency and effectiveness of our approaches, we perform a series of experiments using the existing small instances and a new publicly available benchmark of 14 large instances. Our findings yield new optimal solutions to 185 small instances that were previously unsolved, tripling the total number of instances with known optima. Regarding both small and large instances, our contributions include a comprehensive analysis of the experimental results and an evaluation of the performance of each formulation in distinct scenarios, further advancing our understanding of the LDPAP toward the design of exact approaches for the problem.
Metrics play a crucial role in evaluating the performance of machine learning models. In the context of Natural Language Processing (NLP) tasks, such as text summarization and machine translation, Natural Language Gen...
ISBN:
(纸本)9783031702419;9783031702426
Metrics play a crucial role in evaluating the performance of machine learning models. In the context of Natural Language Processing (NLP) tasks, such as text summarization and machine translation, Natural Language Generation (NLG) metrics such as Bleu and Rouge have been widely used. However, these metrics are based on n-gram matching and do not capture the semantic similarity between the generated and reference texts. To address this, BERTScore has emerged as a popular evaluation metric that uses a pre-trained Large Language Model (LLM) to measure semantic similarity between two sentences. Unlike n-grambased metrics, BERTScore uses the contextual and semantic embeddings of words, allowing flexible semantic evaluation. We outline a number of hypotheticals in which the dependence of BERTScore on token embedding cosine similarity may be exploited. The comparative distribution of BERTScores on a set of reference - prediction pairs mean that results often scale differently with training to traditional metrics, which requires more expertise when interpreting results. In this paper, we demonstrate an improvement to BERTScore, using accelerated Linear Sum Assignment approximations that reduce the mean score while maintaining accuracy. Linear Sum Assignment allows BERTScore to be more easily understood in the context of other NLG metrics, by changing the distribution of the metric.
Few-shot classification aims to classify query samples using very few labeled examples. Most existing methods follow the Prototypical Network to classify query samples by matching them to the nearest centroid. However...
ISBN:
(纸本)9789819984619;9789819984626
Few-shot classification aims to classify query samples using very few labeled examples. Most existing methods follow the Prototypical Network to classify query samples by matching them to the nearest centroid. However, scarce labeled examples tend to bias the centroids, which leads to query samples matching the wrong centroids. In this paper, we address the mismatching problem of centroids and query samples by optimizing the matching strategy. The idea is to combine the Shared Nearest Neighbor similarity with cosine similarity proportionally to calibrate the matching results of the latter. Furthermore, we also improve a bias-diminishing approach to increase the number of shared nearest neighbors between query samples and the centroid of their class. We validate the effectiveness of our method with extensive experiments on three few-shot classification datasets: miniImageNet, tieredImageNet, and CUB-200-2011 (CUB). Moreover, our method has achieved competitive performance across different settings and datasets.
Directed Greybox Fuzzing has proven effective in vulnerability detection areas such as bug reproduction and patch testing. However, existing directed fuzzers are often difficult to customize, lack modularity and have ...
ISBN:
(纸本)9783031641701;9783031641718
Directed Greybox Fuzzing has proven effective in vulnerability detection areas such as bug reproduction and patch testing. However, existing directed fuzzers are often difficult to customize, lack modularity and have limited binary support. This constrains their usability on complex software or when the source code is unavailable;a challenge encountered when fuzzing embedded systems. This article addresses these limitations by introducing the Directed Fuzzing Toolkit (DRIFT) as a platform for directed fuzzing within the modular framework LibAFL. DRIFT modularizes techniques from the state-of-the-art directed fuzzer AFLGo and adapts them for binary applications thereby augmenting LibAFL's highly customizable fuzzers with directed fuzzing capabilities. Additionally, by leveraging Ghidra's analysis, DRIFT achieves architecture agnostic static analysis, opening doors for DGF to tackle previously challenging scenarios. Our evaluation of DRIFT shows a 90% correlation in static analysis metrics over binary compared to its source-code counterpart. Fuzzing performance was also notable despite operating over emulation. In benchmarks, DRIFT's performance exceeds the original fuzzer with up to doubled bug discovery rates and 9-40x faster exploitation times of target bugs. These results are attributed to the toolkit's modular design and its integration with LibAFL. Additionally, DRIFT includes a profiling platform for DGF metrics and is incorporated with the Magma benchmark. Together, these features position DRIFT as a practical advancement in directed fuzzing within LibAFL.
Interpreting human behavior from entirely performed actions is called human action recognition (HAR). HAR applications rapidly expand into robotics, CCTV surveillance, self-driving vehicles, gaming, and video retrieva...
ISBN:
(纸本)9783031646072;9783031646089
Interpreting human behavior from entirely performed actions is called human action recognition (HAR). HAR applications rapidly expand into robotics, CCTV surveillance, self-driving vehicles, gaming, and video retrieval. Among different data modalities, skeleton data offers compact representation and computational efficiency. In recent years, much work has gone into developing a robust and accurate deep-learning framework for skeleton-based HAR. The paper reviews state-of-the-art methods for skeleton-based HAR. The survey also summarizes evaluation results on a large-scale benchmark dataset. Trends in action recognition research are discussed.
Topic modeling is a widely used technique in Natural Language Processing (NLP) for extracting and uncovering latent thematic patterns in a collection of documents. However, working with topic modeling algorithms requi...
ISBN:
(纸本)9783031702419;9783031702426
Topic modeling is a widely used technique in Natural Language Processing (NLP) for extracting and uncovering latent thematic patterns in a collection of documents. However, working with topic modeling algorithms requires technical skills that may hinder the novices. This demo paper presents CWordTM (CWordTM stands for ChineseWord TopicModeling, where "Word" refers to the Word, i.e., the Holy Bible / Scripture.), an open-source Python toolkit designed to simplify the topic modeling process and some other NLP tasks. Although the toolkit offers the Holy Bible as a sample corpus, it works on other possible text data.
Over the past few centuries maps have continued to play many different roles as tools used for guiding and coordinating all kinds of human activities. In modern times, maps are also used as the basis for a range of in...
ISBN:
(纸本)9783031616976;9783031616983
Over the past few centuries maps have continued to play many different roles as tools used for guiding and coordinating all kinds of human activities. In modern times, maps are also used as the basis for a range of interactive tools, applications and visualizations. Despite this, there are not many venues for coordinating map-related research, design, and professional practices. Therefore, this workshop aimed to bring together researchers, designers and practitioners interested in maps and map-like visualizations as the underlying physical, theoretical, or metaphorical framework for designing interfaces and interactions. This workshop created a common ground and a collaborative space for sharing design, research, and practical expertise to aid its participants with creating novel future map-based designs across a number of fields, including visualization, visual design, interaction design, user interface design, and cartography. In this paper, we summarise the contributions made to this workshop by its participants, and provide a brief overview of its activities and outcomes.
This paper introduces computer vision techniques to address the challenge of analyzing tabular data in digital documents. It details the research and experiments conducted by the authors aimed at automating the invoic...
详细信息
ISBN:
(纸本)9783031711145;9783031711152
This paper introduces computer vision techniques to address the challenge of analyzing tabular data in digital documents. It details the research and experiments conducted by the authors aimed at automating the invoice registration process within the accounting system of a particular private company. Based on the results obtained, the authors assert that in certain contexts, pattern recognition processing techniques can continue to be effectively utilized for this problem, providing commercial benefits while circumventing some of the limitations inherent in newer, high-performance deep learning approaches. The article shows an overview of the latest advancements in the field and presents the authors' proposed table structure recognition (TSR) algorithm, which is based on the typical profile-projection approach. It also features comparison of the developed TSR algorithm with selected methods described in the literature, undertaking a verification of its efficacy using a dataset of reals documents derived from the accounting department of a private company.
Multi-object tracking (MOT) is an important and representative task in the field of computer vision, while tracking-by-detection is the most mainstream paradigm for MOT, so that target detection quality, feature repre...
详细信息
ISBN:
(纸本)9783031500787;9783031500770
Multi-object tracking (MOT) is an important and representative task in the field of computer vision, while tracking-by-detection is the most mainstream paradigm for MOT, so that target detection quality, feature representation ability, and association algorithm greatly affect tracking performance. On the one hand, multiple pedestrians moving together in the same group maintain similar motion pattern, so that they can indicate each other's moving state. We extract groups from detections and maintain the group relationship of trajectories in tracking. We propose a state transition mechanism to smooth detection bias, recover missing detection and confront false detection. We also build a two-level group-detection association algorithm, which improves the accuracy of association. On the other hand, different areas of the tracking scene have diverse and varying impact on the detections' appearance feature, which weakens the appearance feature's representation ability. We propose a self-adaptive feature fusion strategy based on the tracking scene and the group structure, which can help us to get fusion feature with stronger representative ability to use in the trajectory-detection association to improve tracking performance. To summary, in this paper, we propose a novel Group Perception based Self-adaptive Fusion Tracking (GST) framework, including Group concept and Group Exploration Net, Group Perception based State Transition Mechanism, and Self-adaptive Feature Fusion Strategy. Experiments on the MOT17 dataset demonstrate the effectiveness of our method. The method achieves competitive results compared to the state-of-the-art methods.
暂无评论