Spiking neuralnetworks (SNNs) are artificialneural network models that show significant advantages in terms of power and energy when realizing deep learning applications. However, the data-intensive nature of machin...
详细信息
ISBN:
(纸本)9781665494663
Spiking neuralnetworks (SNNs) are artificialneural network models that show significant advantages in terms of power and energy when realizing deep learning applications. However, the data-intensive nature of machine learning applications imposes a challenging problem to neural network implementations in terms of latency, energy efficiency and memory bottleneck. Therefore, we introduce a scalable deep SNN to address the problem of latency and energy efficiency. We integrate a Computing-In-Memory (CIM) architecture built with a fabricated memristor crossbar array to reduce the memory bandwidth in vector-matrix multiplication, a key operation in deep learning. By applying an inter-spike interval (ISI) encoding scheme to the input signals, we demonstrate the spatiotemporal information processing capability of our designed architecture. The memristor crossbar array has an enhanced heat dissipation layer that reduces the resistance variation of the memristors by similar to 30%. We further develop a time-to-first-spike (TTFS) method to classify the outputs. The designed circuits and architecture can achieve very high accuracies with both digit recognition and the MNIST dataset Our architecture can classify handwritten digits while consuming merely 2.9mW of power with an inference speed of 2 mu s/image. Only 2.51pJ of energy per synaptic connection makes it suitable to apply in deep learning accelerators.
artificial intelligence techniques are now widely used in various agricultural applications, including the detection of devastating diseases such as late blight (Phytophthora infestans) and early blight (Alternaria so...
详细信息
artificial intelligence techniques are now widely used in various agricultural applications, including the detection of devastating diseases such as late blight (Phytophthora infestans) and early blight (Alternaria solani) affecting potato (Solanum tuberorsum L.) crops. In this paper, we present a mobile application for detecting potato crop diseases based on deep neuralnetworks. The images were taken from the PlantVillage dataset with a batch of 1000 images for each of the three identified classes (healthy, early blight-diseased, late blight-diseased). An exploratory analysis of the architectures used for early and late blight diagnosis in potatoes was performed, achieving an accuracy of 98.7%, with MobileNetv2. Based on the results obtained, an offline mobile application was developed, supported on devices with Android 4.1 or later, also featuring an information section on the 27 diseases affecting potato crops and a gallery of symptoms. For future work, segmentation techniques will be used to highlight the damaged region in the potato leaf by evaluating its extent and possibly identifying different types of diseases affecting the same plant.
Recently, the application of artificial Intelligence (AI) in the Internet of Things (IoT) devices is increasing. As these devices are limited in processing and storing massive computations of AI applications, research...
详细信息
Recently, the application of artificial Intelligence (AI) in the Internet of Things (IoT) devices is increasing. As these devices are limited in processing and storing massive computations of AI applications, researchers are searching for methods to overcome these limitations. One of these applications is Convolutional neural Network (CNN) processing, which is common in object detection and image classification. A neural network consists of layers with a large number of neurons and requires high processing power to run. A CNN can be partitioned into segments and offloaded as tasks of IoT devices to cloudlet servers on the edge. By utilizing the edge servers available in the environment, the total latency in the system and energy consumption by IoT devices can be optimized. Making decisions about offloading CNN segmented layers and allocating resources to each of them is the challenge. In this paper, we propose a method based on deep reinforcement learning that divides the offloading and resource allocation problem into two minor problems. This algorithm updates the offloading policy based on information from the environment, and with the help of the Salp Swarm Algorithm (SSA), optimizes resource allocation. The proposed method is tested for different deep-learning tasks of IoT devices under different capacities of cloudlet servers. The simulation results show the proposed algorithm has the least cost in terms of latency and power consumption and on average has improved 92%, 17%, and 12% compared to full local, full offload, and Jointly Resource allocation and computation Offloading PSO (JROPSO) methods respectively.
Transformer design is the de facto standard for natural language processing tasks. The success of the transformer design in natural language processing has lately piqued the interest of researchers in the domain of co...
Transformer design is the de facto standard for natural language processing tasks. The success of the transformer design in natural language processing has lately piqued the interest of researchers in the domain of computer vision. When compared to Convolutional neuralnetworks (CNNs), Vision Transformers (ViTs) are becoming more popular and dominant solutions for many vision problems. Transformer-based models outperform other types of networks, such as convolutional and recurrent neuralnetworks, in a range of visual benchmarks. We evaluate various vision transformer models in this work by dividing them into distinct jobs and examining their benefits and drawbacks. ViTs can overcome several possible difficulties with convolutional neuralnetworks (CNNs). The goal of this survey is to show the first use of ViTs in CV. In the first phase, we categorize various CV applications where ViTs are appropriate. image classification, object identification, image segmentation, video transformer, image denoising, and NAS are all CV applications. Our next step will be to analyze the state-of-the-art in each area and identify the models that are currently available. In addition, we outline numerous open research difficulties as well as prospective research possibilities.
The global significance of diagnosing liver diseases is heightened, particularly in under-resourced regions with limited healthcare facilities. Traditional diagnostic methods, characterized by time and labor-intensive...
详细信息
The global significance of diagnosing liver diseases is heightened, particularly in under-resourced regions with limited healthcare facilities. Traditional diagnostic methods, characterized by time and labor-intensive processes, have led to a growing demand for telemedicine-based solutions. The incorporation of artificial Intelligence is deemed essential to enhance the efficiency and accuracy of diagnostic models. This review explores the seamless integration of diverse data modalities, including clinical records, demographics, laboratory values, biopsy specimens, and imaging data, emphasizing the importance of combining both types for comprehensive liver disease diagnosis. The study rigorously examines various approaches, focusing on pre-processing and feature engineering in non-image data diagnostic model development. Additionally, it analyzes studies employing Convolutional neuralnetworks for cutting-edge solutions in image data interpretation. The paper provides insights into existing liver disease datasets, encompassing both image and non-image data, offering a comprehensive understanding of crucial research data sources. Emphasis is placed on performance evaluation metrics and their correlation in assessing diagnostic model efficiency. The review also explores open-source software tools dedicated to computer-aided liver analysis, enhancing exploration in liver disease diagnostics. Serving as a concise handbook, it caters to novice and experienced researchers alike, offering essential insights, summarizing the latest research, and providing a glimpse into emerging trends, challenges, and future trajectories in liver disease diagnosis.
Transformer design is the de facto standard for natural language processing tasks. The success of the transformer design in natural language processing has lately piqued the interest of researchers in the domain of co...
详细信息
Adversarial attacks pose a significant challenge to deploying deep learning models in safety-critical applications. Maintaining model robustness while ensuring interpretability is vital for fostering trust and compreh...
详细信息
ISBN:
(数字)9798331515942
ISBN:
(纸本)9798331515959
Adversarial attacks pose a significant challenge to deploying deep learning models in safety-critical applications. Maintaining model robustness while ensuring interpretability is vital for fostering trust and comprehension in these models. This study investigates the impact of Saliency-guided Training (SGT) on model robustness, a technique aimed at improving the clarity of saliency maps to deepen understanding of the model’s decision-making process. Experiments were conducted on standard benchmark datasets using various deep learning architectures trained with and without SGT. Findings demonstrate that SGT enhances both model robustness and interpretability. Additionally, we propose a novel approach combining SGT with standard adversarial training to achieve even greater robustness while preserving saliency map quality. Our strategy is grounded in the assumption that preserving salient features crucial for correctly classifying adversarial examples enhances model robustness, while masking non-relevant features improves interpretability. Our technique yields significant gains, achieving a 35% and 20% improvement in robustness against PGD attack with noise magnitudes of 0.2 and 0.02 for the MNIST and CIFAR-10 datasets, respectively, while producing high-quality saliency maps.
Generative Adversarial networks or GANS, are another way of achieving generative modeling using various deep learning methods like convoluted neuralnetworks. They have a wide range of applications like :- image to im...
详细信息
Spiking neuralnetworks are a promising concept not only in terms of better simulation of biological neuralnetworks but also in terms of overcoming the current disadvantages of artificialneuralnetworks, such as hig...
详细信息
The prolonged waiting time at supermarket checkout lines poses a significant challenge to the shopping experience, impacting customer satisfaction and operational efficiency. This paper presents a prototype that addre...
The prolonged waiting time at supermarket checkout lines poses a significant challenge to the shopping experience, impacting customer satisfaction and operational efficiency. This paper presents a prototype that addresses this issue by utilizing computer vision and deep learning. The model, incorporating convolutional neuralnetworks such as YOLO v4 tiny and YOLO v5 small, along with tools like OpenCV and Roboflow for dataset management, achieves a remarkable 98% mean average precision for two-class detection. It efficiently detects, classifies, tracks, and counts items on a mobile supermarket conveyor belt. Additionally, we introduce a versatile framework designed for seamless integration into real-world applications. It comprises a customizable monitoring application and simulator that facilitates synthetic image data generation. Managing diverse items in a supermarket presents a major challenge for data gathering, labeling, and training. In that sense, the importance of customizable monitoring and simulation tools is highlighted, emphasizing their practical role. Our findings demonstrate the feasibility of maintaining a minimal 0% to 2.85% precision tradeoff while using half of the data as synthetic for two-class detection, indicating potential practicality in supermarkets with proper scaling. In summary, this study brings tangible benefits to both customers and retailers, offering a potential to streamline, speed up, and cut costs in the supermarket checkout process.
暂无评论