作者:
Zhou, ZhengyuLiu, WeiweiSchool of Computer Science
National Engineering Research Center for Multimedia Software Institute of Artificial Intelligence Hubei Key Laboratory of Multimedia and Network Communication Engineering Wuhan University Wuhan China
Goodness-of-fit testing, a classical statistical tool, has been extensively explored in the batch setting, where the sample size is ***, practitioners often prefer methods that adapt to the complexity of a problem rat...
Goodness-of-fit testing, a classical statistical tool, has been extensively explored in the batch setting, where the sample size is ***, practitioners often prefer methods that adapt to the complexity of a problem rather than fixing the sample size *** batch tests are generally unsuitable for streaming data, as valid inference after data peeking requires multiple testing corrections, resulting in reduced statistical *** address this issue, we delve into the design of consistent sequential goodness-of-fit *** the principle of testing by betting, we reframe this task as selecting a sequence of payoff functions that maximize the wealth of a fictitious bettor, betting against the null in a repeated *** conduct experiments to demonstrate the adaptability of our sequential test across varying difficulty levels of problems while maintaining control over type-I errors. Copyright 2024 by the author(s)
With the development of Blockchain, cloud computing, and artificialintelligence, smart transportation is highly likely to drive urban transportation in the direction of intelligence and digitalization. However, when ...
详细信息
Learning with inaccurate supervision is often encountered in weakly supervised learning, and researchers have invested a considerable amount of time and effort in designing specialized algorithms for different forms o...
详细信息
Learning with inaccurate supervision is often encountered in weakly supervised learning, and researchers have invested a considerable amount of time and effort in designing specialized algorithms for different forms of annotations in inaccurate supervision. In fact, different forms of these annotations share the fundamental characteristic that they all still incorporate some portion of correct labeling information. This commonality can serve as a lever, enabling the creation of a cohesive framework designed to tackle the challenges associated with various forms of annotations in learning with inaccurate supervision. In this paper, we propose a unified label refinement framework named ULAREF, i.e., a Unified LAbel REfinement Framework for learning with inaccurate supervision, which is capable of leveraging label refinement to handle inaccurate supervision. Specifically, our framework trains the predictive model with refined labels through global detection of reliability and local enhancement using an enhanced model fine-tuned by a proposed consistency loss. Also, we theoretically justify that the enhanced model in local enhancement can achieve higher accuracy than the predictive model on the detected unreliable set under mild assumptions. Copyright 2024 by the author(s)
In the realm of medical image analysis, self-supervised learning (SSL) techniques have emerged to alleviate labeling demands, while still facing the challenge of training data scarcity owing to escalating resource req...
详细信息
In order to improve the prediction accuracy of PM2.5, a combined prediction model based on empirical mode decomposition, temporal convolutional network and frequency-enhanced gated attention was proposed. The eigenmod...
详细信息
Soft pseudo-labels, generated by the softmax predictions of the trained networks, offer a probabilistic rather than binary form, and have been shown to improve the performance of deep neural networks in supervised lea...
详细信息
Soft pseudo-labels, generated by the softmax predictions of the trained networks, offer a probabilistic rather than binary form, and have been shown to improve the performance of deep neural networks in supervised learning. Most previous methods adopt classification loss to train a classifier as the soft-pseudo-label generator and fail to fully exploit their potential due to the misalignment with the target of soft-pseudo-label generation, aimed at capturing the knowledge in the data rather than making definitive classifications. Nevertheless, manually designing an effective objective function for a soft-pseudo-label generator is challenging, primarily because datasets typically lack ground-truth soft labels, complicating the evaluation of the soft pseudo-label accuracy. To deal with this problem, we propose a novel framework that alternately trains the predictive model and the soft-pseudo-label generator guided by a meta-network-parameterized label enhancement objective. The parameters of the objective function are optimized based on the feedback from both the performance of the predictive model and the soft-pseudo-label generator in the learning task. Additionally, the framework offers versatility across different learning tasks by allowing direct modifications to the task loss. Experiments on the benchmark datasets validate the effectiveness of the proposed framework. Source code is available at https://***/palm-ml/SEAL. Copyright 2024 by the author(s)
Optical machine learning has emerged as an important research area that,by leveraging the advantages inherent to optical signals,such as parallelism and high speed,paves the way for a future where optical hardware can...
详细信息
Optical machine learning has emerged as an important research area that,by leveraging the advantages inherent to optical signals,such as parallelism and high speed,paves the way for a future where optical hardware can process data at the speed of *** this work,we present such optical devices for data processing in the form of single-layer nanoscale holographic perceptrons trained to perform optical inference *** experimentally show the functionality of these passive optical devices in the example of decryptors trained to perform optical inference of single or whole classes of keys through symmetric and asymmetric *** decryptors,designed for operation in the near-infrared region,are nanoprinted on complementary metal-oxide-semiconductor chips by galvo-dithered two-photon nanolithography with axial nanostepping of 10 nm achieving a neuron density of>500 million neurons per square *** power-efficient commixture of machine learning and on-chip integration may have a transformative impact on optical decryption3,sensing4,medical diagnostics5 and computing6,7.
作者:
Ma, XinsongZou, XinLiu, WeiweiSchool of Computer Science
National Engineering Research Center for Multimedia Software Institute of Artificial Intelligence Hubei Key Laboratory of Multimedia and Network Communication Engineering Wuhan University Wuhan China
Out-of-distribution (OOD) detection task plays the key role in reliable and safety-critical applications. Existing researches mainly devote to designing or training the powerful score function but overlook investigati...
Out-of-distribution (OOD) detection task plays the key role in reliable and safety-critical applications. Existing researches mainly devote to designing or training the powerful score function but overlook investigating the decision rule based on the proposed score function. Different from previous work, this paper aims to design a decision rule with rigorous theoretical guarantee and well empirical performance. Specifically, we provide a new insight for the OOD detection task from a hypothesis testing perspective and propose a novel generalized Benjamini Hochberg (g-BH) procedure with empirical p-values to solve the testing problem. Theoretically, the g-BH procedure controls false discovery rate (FDR) at pre-specified level. Furthermore, we derive an upper bound of the expectation of false positive rate (FPR) for the g-BH procedure based on the tailed generalized Gaussian distribution family, indicating that the FPR of g-BH procedure converges to zero in probability. Finally, the extensive experimental results verify the superiority of g-BH procedure over the traditional threshold-based decision rule on several OOD detection benchmarks. Copyright 2024 by the author(s)
The pre-training-then-fine-tuning paradigm has been widely used in deep *** to the huge computation cost for pre-training,practitioners usually download pre-trained models from the Internet and fine-tune them on downs...
详细信息
The pre-training-then-fine-tuning paradigm has been widely used in deep *** to the huge computation cost for pre-training,practitioners usually download pre-trained models from the Internet and fine-tune them on downstream datasets,while the downloaded models may suffer backdoor *** from previous attacks aiming at a target task,we show that a backdoored pre-trained model can behave maliciously in various downstream tasks without foreknowing task *** can restrict the output representations(the values of output neurons)of trigger-embedded samples to arbitrary predefined values through additional training,namely neuron-level backdoor attack(NeuBA).Since fine-tuning has little effect on model parameters,the fine-tuned model will retain the backdoor functionality and predict a specific label for the samples embedded with the same *** provoke multiple labels in a specific task,attackers can introduce several triggers with predefined contrastive *** the experiments of both natural language processing(NLP)and computer vision(CV),we show that NeuBA can well control the predictions for trigger-embedded instances with different trigger *** findings sound a red alarm for the wide use of pre-trained ***,we apply several defense methods to NeuBA and find that model pruning is a promising technique to resist NeuBA by omitting backdoored neurons.
Self-supervised learning and knowledge distillation intersect to achieve exceptional performance on downstream tasks across diverse network capacities. This paper introduces MIM-HD, which implements enhancements for m...
详细信息
暂无评论