Due to the characteristics of stream applications and the insufficiency of conventional processors when running stream programs, stream processors which support data-level parallelism become the research hotspot. This...
详细信息
Due to the characteristics of stream applications and the insufficiency of conventional processors when running stream programs, stream processors which support data-level parallelism become the research hotspot. This paper presents two means, stream partition (SP) and stream compression (SC), to optimize streams on Imagine. The results of simulation show that SP and SC can make stream applications take full advantage of the parallel clusters, pipelines and three-level memory hierarchy of the Imagine processor, and then reduce the execution time of stream programs.
Underwater acoustic classification is a challenging task due to complex background noise and complicated sound propagation patterns. How to represent the signals is important for the classification task. In this paper...
Underwater acoustic classification is a challenging task due to complex background noise and complicated sound propagation patterns. How to represent the signals is important for the classification task. In this paper, we propose a novel representation learning method for the underwater acoustic signals, leveraging the mask modeling-based self-supervised learning paradigm. Specifically, we first explore modifying the Swin Transformer architecture to learn general representation for the audio signals, accompanied with random masking on the log-mel spectrogram. The main goal of the pretext task is to predict the masked parts of Log-mel spectrogram and the gamma-stone spectrogram, so that the model can not only learn the local and global features but also learn complementary information. For downstream task, we utilize the labelled datasets to fine-tune the pre-trained model. On DeepShip datasets which consist of 47 hand 4 minof ship sounds in four categories, our model achieves state-of-the-art performance compared with competitive approaches. Our method obtains a classification accuracy of 78.03%, which is better than the separable convolution autoencoder (SCAE) and using the constant-Q transform spectrogram. This work demonstrates the potential of the masked modeling based self-supervised learning for understanding and interpretation of underwater acoustic signals.
To deal with the scalable and fast unbiased sampling problems in unstructured P2P systems, a sampling method based on multi-peer adaptive random walk (SMARW) is proposed. In the method, based on the multi-peer random ...
详细信息
Despite impressive capabilities and outstanding performance, deep neural networks (DNNs) have captured increasing public concern about their security problems, due to their frequently occurred erroneous behaviors. The...
详细信息
Despite impressive capabilities and outstanding performance, deep neural networks (DNNs) have captured increasing public concern about their security problems, due to their frequently occurred erroneous behaviors. Therefore, it is necessary to conduct a systematical testing for DNNs before they are deployed to real-world applications. Existing testing methods have provided fine-grained metrics based on neuron coverage and proposed various approaches to improve such metrics. However, it has been gradually realized that a higher neuron coverage does not necessarily represent better capabilities in identifying defects that lead to errors. Besides, coverage-guided methods cannot hunt errors due to faulty training procedure. So the robustness improvement of DNNs via retraining by these testing examples are unsatisfactory. To address this challenge, we introduce the concept of excitable neurons based on Shapley value and design a novel white-box testing framework for DNNs, namely DeepSensor. It is motivated by our observation that neurons with larger responsibility towards model loss changes due to small perturbations are more likely related to incorrect corner cases due to potential defects. By maximizing the number of excitable neurons concerning various wrong behaviors of models, DeepSensor can generate testing examples that effectively trigger more errors due to adversarial inputs, polluted data and incomplete training. Extensive experiments implemented on both image classification models and speaker recognition models have demonstrated the superiority of DeepSensor. Compared with the state-of-the-art testing approaches, DeepSensor can find more test errors due to adversarial inputs (∼ ×1.2), polluted data (∼ ×5) and incompletely-trained DNNs (∼ ×1.3). Additionally, it can help DNNs build larger l2-norm robustness bound (∼ ×3) via retraining according to CLEVER's certification. We further provide interpretable proofs for effectiveness of DeepSensor via excitable neuro
Regarding the non-negativity property of the magnitude spectrogram of speech signals, nonnegative matrix factorization (NMF) has obtained promising performance for speech separation by independently learning a diction...
详细信息
ISBN:
(纸本)9781479928941
Regarding the non-negativity property of the magnitude spectrogram of speech signals, nonnegative matrix factorization (NMF) has obtained promising performance for speech separation by independently learning a dictionary on the speech signals of each known speaker. However, traditional NM-F fails to represent the mixture signals accurately because the dictionaries for speakers are learned in the absence of mixture signals. In this paper, we propose a new transductive NMF algorithm (TNMF) to jointly learn a dictionary on both speech signals of each speaker and the mixture signals to be separated. Since TNMF learns a more descriptive dictionary by encoding the mixture signals than that learned by NMF, it significantly boosts the separation performance. Experiments results on a popular TIMIT dataset show that the proposed TNMF-based methods outperform traditional NMF-based methods for separating the monophonic mixtures of speech signals of known speakers.
As a new stage in the development of the cloud computing paradigm, serverless computing has the high-level abstraction characteristic of shielding underlying details. This makes it extremely challenging for users to c...
As a new stage in the development of the cloud computing paradigm, serverless computing has the high-level abstraction characteristic of shielding underlying details. This makes it extremely challenging for users to choose a suitable serverless platform. To address this, targeting the jointcloud computing scenario of heterogeneous serverless platforms across multiple clouds, this paper presents a jointcloud collaborative mechanism called FCloudless with cross-cloud detection of the full lifecycle performance of serverless platforms. Based on the benchmark metrics set that probe performance critical stages of the full lifecycle, this paper proposes a performance optimization algorithm based on detected performance data that takes into account all key stages that affect the performance during the lifecycle of a function and predicts the overall performance by combining the scores of local stages and dynamic weights. We evaluate FCloudless on AWS, AliYun, and Azure. The experimental results show that FCloudless can detect the underlying performance of serverless platforms hidden in the black box and its optimization algorithm can select the optimal scheduling strategy for various applications in a jointcloud environment. FCloudless reduces the runtime by 23.3% and 24.7% for cold and warm invocations respectively under cost constraints.
This paper introduces a new deep learning approach to approximately solve the Covering Salesman Problem (CSP). In this approach, given the city locations of a CSP as input, a deep neural network model is designed to d...
详细信息
Recent years have seen the wide application of NLP models in crucial areas such as finance, medical treatment, and news media, raising concerns of the model robustness and vulnerabilities. In this paper, we propose a ...
详细信息
Deduplication is a data redundancy elimination technique, designed to save system storage resources by reducing redundant data in cloud storage systems. With the development of cloud computing technology, deduplicatio...
详细信息
ISBN:
(数字)9781728190747
ISBN:
(纸本)9781728183824
Deduplication is a data redundancy elimination technique, designed to save system storage resources by reducing redundant data in cloud storage systems. With the development of cloud computing technology, deduplication has been increasingly applied to cloud data centers. However, traditional technologies face great challenges in big data deduplication to properly weigh the two conflicting goals of deduplication throughput and high duplicate elimination ratio. This paper proposes a similarity clustering-based deduplication strategy (named SCDS), which aims to delete more duplicate data without significantly increasing system overhead. The main idea of SCDS is to narrow the query range of fingerprint index by data partitioning and similarity clustering algorithms. In the data preprocessing stage, SCDS uses data partitioning algorithm to classify similar data together. In the data deletion stage, the similarity clustering algorithm is used to divide the similar data fingerprint superblock into the same cluster. Repetitive fingerprints are detected in the same cluster to speed up the retrieval of duplicate fingerprints. Experiments show that the deduplication ratio of SCDS is better than some existing similarity deduplication algorithms, but the overhead is only slightly higher than some high throughput but low deduplication ratio methods.
暂无评论