With society’s increasing data production and the corresponding demand for systems that are capable of utilizing them, the big data domain has gained significant importance. However, besides the systems’ actual impl...
With society’s increasing data production and the corresponding demand for systems that are capable of utilizing them, the big data domain has gained significant importance. However, besides the systems’ actual implementation, their testing also needs to be considered. For this, oftentimes, proper test data sets are necessary. This publication discusses several different approaches how these can be provisioned and, further, highlights the respective advantages, disadvantages, and suitable application scenarios. In doing so, researchers and practitioners that are implementing big data applications and need to test them, or who are generally interested in the domain, are supported in their own considerations on how to obtain test data.
This document 1 is an extended abstract of a science of computer Programming paper, "Candidate Test Set Reduction for Adaptive Random Testing: An Overheads Reduction Technique," presented as a J1C2 (Journa...
详细信息
This document 1 is an extended abstract of a science of computer Programming paper, "Candidate Test Set Reduction for Adaptive Random Testing: An Overheads Reduction Technique," presented as a J1C2 (Journal publication first, Conference presentation following) at the 30th IEEE International Conference on software Analysis, Evolution and Reengineering (Saner 2023).The paper presents a candidate set reduction strategy to enhance the Fixed-Sized-Candidate-Set version of Adaptive Random Testing (FSCS-ART). The proposed method reduces the number of randomly-generated candidate test cases by retaining valuable, unused candidates from previous iterations. As the computational costs associated with a stored/retained candidate are less than the costs associated with a randomly-generating one, the overall computational overheads of FSCS-ART are reduced. The reported experimental studies show that the proposed method has a comparable failure-detection effectiveness to FSCS-ART, but less computational overheads.
作者:
Koana, TomohiroTechnische Universität Berlin
Faculty IV Institute of Software Engineering and Theoretical Computer Science Algorithmics and Computational Complexity Germany
In this work, we study the Induced Matching problem: Given an undirected graph G and an integer , is there an induced matching M of size at least ? An edge subset M is an induced matching in G if M is a matching such ...
详细信息
Social media has emerged as one of the most transformative developments on the internet,revolu-tionizing the way people communicate and ***,alongside its benefits,social media has also given rise to significant challe...
详细信息
Social media has emerged as one of the most transformative developments on the internet,revolu-tionizing the way people communicate and ***,alongside its benefits,social media has also given rise to significant challenges,one of the most pressing being *** issue has become a major concern in modern society,particularly due to its profound negative impacts on the mental health and well-being of its *** the Arab world,where social media usage is exceptionblly high,cyberbullying has become increasingly prevalent,necessitating urgent *** detection of harmful online behavior is critical to fostering safer digital environments and mitigating the adverse efcts of *** underscores the importance of developing advanced tools and systems to identify and address such behavior *** paper investigates the development of a robust cyberbullying detection and classifcation system tailored for Arabic comments on *** study explores the efectiveness of various deep learning models,including Bi-LSTM(Bidirectional Long Short Term Memory),LSTM(Long Short-Term Memory),CNN(Convolutional Neural Networks),and a hybrid CNN-LSTM,in classifying Arabic comments into binary classes(bullying or not)and multiclass categories.A comprehensive dataset of 20,000 Arabic YouTube comments was collected,preprocessed,and labeled to support these *** results revealed that the CNN and hybrid CNN-LSTM models achieved the highest accuracy in binary classification,reaching an impressive 91.9%.For multiclass dlassification,the LSTM and Bi-LSTM models outperformed others,achieving an accuracy of 89.5%.These findings highlight the efctiveness of deep learning approaches in the mitigation of cyberbullying within Arabic online communities.
We study the network untangling problem introduced by Rozenshtein, Tatti, and Gionis [DMKD 2021], which is a variant of Vertex Cover on temporal graphs-graphs whose edge set changes over discrete time steps. They intr...
详细信息
In various applications in Internet of Things like industrial monitoring, large amounts of floating-point time series data are generated at an unprecedented rate. Efficient compression algorithms can effectively reduc...
In various applications in Internet of Things like industrial monitoring, large amounts of floating-point time series data are generated at an unprecedented rate. Efficient compression algorithms can effectively reduce the size of data, enhance transmission performance and storage efficiency, and simultaneously lower storage costs. Therefore, there is a need for lightweight and efficient stream compression algorithms. In this paper, we propose a novel lossless floating-point data compression algorithm called Ant. The main idea is to encode double-precision floating-point numbers into integer form, calculate the delta between adjacent values, and then convert the delta into unsigned integers. This encoding method effectively reduces storage costs and improves data compression efficiency. Extensive experiments on real-world datasets demonstrate that our algorithm achieves compression speeds at least as fast as state-of-the-art streaming methods, and a 63% relative improvement in average compression rate.
Vertex Cover parameterized by the solution size k is the quintessential fixed-parameter tractable problem. FPT algorithms are most interesting when the parameter is small. Several lower bounds on k are well-known, suc...
详细信息
Due to the rapid development of image data and the necessity to analyze it to extract meaningful information, heterogeneous systems have gained prominence. One of the most critical aspects of distributed systems is lo...
详细信息
In dynamic meteorological prediction, accurate rainfall forecasting is a mystery. In a complex and dynamic natural environment with unpredictable sky movements, we propose an innovative methodology that forecasts week...
In dynamic meteorological prediction, accurate rainfall forecasting is a mystery. In a complex and dynamic natural environment with unpredictable sky movements, we propose an innovative methodology that forecasts weekly average rainfall patterns using the Gooseneck Barnacle Optimizer and the Least Squares Support Vector Machine (LSSVM). The significance of rainfall prediction is shown by its widespread implications, including public health. However, conventional single-model techniques and machine-learning methods must represent rainfall pattern changes, limiting our preparation. With a 2014–2018 dataset from reliable meteorological libraries, our research explores the possibility of this novel combination. The Gooseneck Barnacle Optimizer (GBO) serves as the bedrock of our methodology, introducing a novel evolutionary algorithm inspired by the intricate mating behaviors of gooseneck barnacles. GBO aptly captures the dynamic interplay of factors such as navigational sperm casting properties, food availability, food attractiveness, wind direction, and intertidal zone wave movement during mating, culminating in two crucial optimization stages: exploration and exploitation. In contrast to previous algorithms like the Barnacle Mating Optimizer (BMO), GBO stands out by more accurately emulating the unique mating behaviors of gooseneck barnacles. The prediction job is performed by integrating the Least Squares Support Vector Machines (LSSVM) algorithm with GBO as an objective function, using the optimized hyper-parameter values. The results suggest that the GBO algorithm exhibits superior performance compared to existing methodologies. This is accomplished by efficiently augmenting the original random population for a particular problem, resulting in a convergence towards the global optimum and producing significantly enhanced optimization outcomes.
The potato is a widely grown crop in many regions of the world. In recent decades, potato farming has gained incredible traction in the world. Potatoes are susceptible to several illnesses that stunt their development...
详细信息
暂无评论