The Hopfield network is an example of an artificialneural network used to implement associative memories. A binary digit represents the neuron's state of a traditional Hopfield neural network. Inspired by the hum...
详细信息
The Hopfield network is an example of an artificialneural network used to implement associative memories. A binary digit represents the neuron's state of a traditional Hopfield neural network. Inspired by the human brain's ability to cope simultaneously with multiple sensorial inputs, this paper presents three multi-modal Hopfield-type neuralnetworks that treat multi-dimensional data as a single entity. In the first model, called the vector-valued Hopfield neural network, the neuron's state is a vector of binary digits. Synaptic weights are modeled as finite impulse response (FIR) filters in the second model, yielding the so-called convolutional associative memory. Finally, the synaptic weights are modeled by linear time-varying (LTV) filters in the third model. Besides their potential applications for multi-modal intelligence, the new associative memories may also be used for signal and imageprocessing and solve optimization and classification tasks.
Vision transformers have become popular as a possible substitute to convolutional neuralnetworks (CNNs) for a variety of computer vision applications. These transformers, with their ability to focus on global relatio...
详细信息
Vision transformers have become popular as a possible substitute to convolutional neuralnetworks (CNNs) for a variety of computer vision applications. These transformers, with their ability to focus on global relationships in images, offer large learning capacity. However, they may suffer from limited generalization as they do not tend to model local correlation in images. Recently, in vision transformers hybridization of both the convolution operation and self-attention mechanism has emerged, to exploit both the local and global image representations. These hybrid vision transformers, also referred to as CNN-Transformer architectures, have demonstrated remarkable results in vision applications. Given the rapidly growing number of hybrid vision transformers, it has become necessary to provide a taxonomy and explanation of these hybrid architectures. This survey presents a taxonomy of the recent vision transformer architectures and more specifically that of the hybrid vision transformers. Additionally, the key features of these architectures such as the attention mechanisms, positional embeddings, multi-scale processing, and convolution are also discussed. In contrast to the previous survey papers that are primarily focused on individual vision transformer architectures or CNNs, this survey uniquely emphasizes the emerging trend of hybrid vision transformers. By showcasing the potential of hybrid vision transformers to deliver exceptional performance across a range of computer vision tasks, this survey sheds light on the future directions of this rapidly evolving architecture.
The paper examines the integration of AI in visual communication design, particularly focusing on the enhanced impact of images when combined with text. It explores the evolution of visual communication design in the ...
详细信息
Convolutional neuralnetworks (CNN) have become a common choice for industrial quality control, as well as other critical applications in the Industry 4.0. When these CNNs behave in ways unexpected to human users or d...
详细信息
Convolutional neuralnetworks (CNN) have become a common choice for industrial quality control, as well as other critical applications in the Industry 4.0. When these CNNs behave in ways unexpected to human users or developers, severe consequences can arise, such as economic losses or an increased risk to human life. Concept extraction techniques can be applied to increase the reliability and transparency of CNNs through generating global explanations for trained neural network models. The decisive features of image datasets in quality control often depend on the feature's scale;for example, the size of a hole or an edge. However, existing concept extraction methods do not correctly represent scale, which leads to problems interpreting these models as we show herein. To address this issue, we introduce the Scale-Preserving Automatic Concept Extraction (SPACE) algorithm, as a state-of-the-art alternative concept extraction technique for CNNs, focused on industrial applications. SPACE is specifically designed to overcome the aforementioned problems by avoiding scale changes throughout the concept extraction process. SPACE proposes an approach based on square slices of input images, which are selected and then tiled before being clustered into concepts. Our method provides explanations of the models' decision-making process in the form of human-understandable concepts. We evaluate SPACE on three image classification datasets in the context of industrial quality control. Through experimental results, we illustrate how SPACE outperforms other methods and provides actionable insights on the decision mechanisms of CNNs. Finally, code for the implementation of SPACE is provided.
With the rise of deep learning, the cross collision between artificial intelligence and art, represented by image style transfer, has attracted high attention in the fields of graphic and image technology and art. Bas...
详细信息
Cross-modal image-text retrieval has gained increasing attention due to its ability to combine computer vision with natural language processing. Previously, image and text features were extracted and concatenated to f...
详细信息
Cross-modal image-text retrieval has gained increasing attention due to its ability to combine computer vision with natural language processing. Previously, image and text features were extracted and concatenated to feed the transformer-based retrieval network. However, these approaches implicitly aligned the image and text modalities since the self-attention mechanism computes attention coefficients for all input features. In this paper, we propose cross-modal Semantic Alignments Module (SAM) to establish an explicit alignment through enhancing an inter-modal relationship. Firstly, visual and textual representations were extracted from an image and text pair. Secondly, we constructed a bipartite graph by representing the image regions and words in the sentence as nodes, and the relationship between them as edges. Then our proposed SAM allows the model to compute attention coefficients based on the edges in the graph. This process helps explicitly align the two modalities. Finally, a binary classifier was used to determine whether the given image-text pair is aligned. We reported extensive experiments on MS-COCO and Flickr30K test sets, showing that SAM could capture the joint representation between the two modalities and could be applied to the existing retrieval networks.
Purpose With the help of basic physics, the application of computer algorithms in the form of recent advances such as machine learning and neural networking in textile Industry has been discussed in this review. Scien...
详细信息
Purpose With the help of basic physics, the application of computer algorithms in the form of recent advances such as machine learning and neural networking in textile Industry has been discussed in this review. Scientists have linked the underlying structural or chemical science of textile materials and discovered several strategies for completing some of the most time-consuming tasks with ease and precision. Since the 1980s, computer algorithms and machine learning have been used to aid the majority of the textile testing process. With the rise in demand for automation, deep learning, and neuralnetworks, these two now handle the majority of testing and quality control operations in the form of imageprocessing. Design/methodology/approach The state-of-the-art of artificial intelligence (AI) applications in the textile sector is reviewed in this paper. Based on several research problems and AI-based methods, the current literature is evaluated. The research issues are categorized into three categories based on the operation processes of the textile industry, including yarn manufacturing, fabric manufacture and coloration. Findings AI-assisted automation has improved not only machine efficiency but also overall industry operations. AI's fundamental concepts have been examined for real-world challenges. Several scientists conducted the majority of the case studies, and they confirmed that image analysis, backpropagation and neural networking may be specifically used as testing techniques in textile material testing. AI can be used to automate processes in various circumstances. Originality/value This research conducts a thorough analysis of artificialneural network applications in the textile sector.
Emerging optoelectronic synapses hold immense potential for advancing neuromorphic computing systems. However, achieving precise control over selective responses in optoelectronic memory and clarifying tunable synapti...
详细信息
Emerging optoelectronic synapses hold immense potential for advancing neuromorphic computing systems. However, achieving precise control over selective responses in optoelectronic memory and clarifying tunable synaptic weights has remained challenging. This study reports an optoelectronic synapse utilizing oxygen plasma-assisted defect engineering in tellurene for artificialneuralnetworks. Through DFT calculations and experimental analyses, we demonstrate that tellurene conductance can be modulated by controlling plasma-defined defect engineering, allowing a transition from short-term to long-term synaptic plasticity, largely determined by intrinsic large-lattice-relaxation effects. Our artificial synapses exhibit high linearity, a broad dynamic range, and tunable synaptic weights. Additionally, our optoelectronic synapses display selective sensitivity to multi-spectral light and achieve a pattern recognition accuracy of up to 96.7% across five typical datasets, surpassing even the ideal synapse. These tunable spectral responses, combined with high-performance neuromorphic applications using spike coding, establish a foundation for developments in brain-inspired machine learning, robotics, and real-time data processing.
Instrument tone recognition systems have over time had the highest application value and significance in information retrieval. Notably, the traditional systems and methods often rely on convolutional neuralnetworks ...
详细信息
As deep learning models scale, they become increasingly competitive from domains spanning from computer vision to natural language processing;however, this happens at the expense of efficiency since they require incre...
详细信息
As deep learning models scale, they become increasingly competitive from domains spanning from computer vision to natural language processing;however, this happens at the expense of efficiency since they require increasingly more memory and computing power. The power efficiency of the biological brain outperforms any large-scale deep learning (DL) model;thus, neuromorphic computing tries to mimic the brain operations, such as spike-based information processing, to improve the efficiency of DL models. Despite the benefits of the brain, such as efficient information transmission, dense neuronal interconnects, and the co-location of computation and memory, the available biological substrate has severely constrained the evolution of biological brains. Electronic hardware does not have the same constraints;therefore, while modeling spiking neuralnetworks (SNNs) might uncover one piece of the puzzle, the design of efficient hardware backends for SNNs needs further investigation, potentially taking inspiration from the available work done on the artificialneuralnetworks (ANNs) side. As such, when is it wise to look at the brain while designing new hardware, and when should it be ignored? To answer this question, we quantitatively compare the digital hardware acceleration techniques and platforms of ANNs and SNNs. As a result, we provide the following insights: (i) ANNs currently process static data more efficiently, (ii) applications targeting data produced by neuromorphic sensors, such as event-based cameras and silicon cochleas, need more investigation since the behavior of these sensors might naturally fit the SNN paradigm, and (iii) hybrid approaches combining SNNs and ANNs might lead to the best solutions and should be investigated further at the hardware level, accounting for both efficiency and loss optimization.
暂无评论