The proceedings contain 17 papers. The special focus in this conference is on Digital Transformation in Education and artificialintelligenceapplications. The topics include: artificialintelligence in Elementar...
ISBN:
(纸本)9783031620577
The proceedings contain 17 papers. The special focus in this conference is on Digital Transformation in Education and artificialintelligenceapplications. The topics include: artificialintelligence in Elementary Math Education: Analyzing Impact on Students Achievements;Gamification in Learning Process Enhanced with AI;impact of the e-Schools Programme on the Use of the e-Class Register;students’ Digital Learning Behavior Using the Mandatory and Non-mandatory Platforms in an Online Learning Environment;Assessing the Impact of Large-Scale ICT Investment in Education Through Measuring the Digital Maturity of Schools;opportunities for the Professional Development of Teachers in Digital Competences Related to the Use of artificialintelligence in Education in Croatia;crop-Guided Neural Network Segmentation of High-Resolution Skin Lesion Images;artificialintelligence-Based Control of Autonomous Vehicles in Simulation: A CNN vs. RL Case Study;fuzzy-Based Knowledge Design and Delivery Model for Personalised Learning;the Development of Assistive robotics: A Comprehensive Analysis Integrating machine Learning, Robotic vision, and Collaborative Human Assistive Robots;development of a Dynamic Multi-object Planning Framework for Autonomous Mobile Robots;application of artificialintelligence in the Economic and Legal Affairs of Companies;Impact of AI Tools on Software Development Code Quality;a Systematic Review of robotics’ Transformative Role in Education;application of the Decision Tree in the Business Process.
Point cloud registration aims at accurately aligning and integrating multiple point cloud data by scanning the same object from different viewpoints into a more comprehensive 3-D model or scene. It has significant app...
详细信息
Point cloud registration aims at accurately aligning and integrating multiple point cloud data by scanning the same object from different viewpoints into a more comprehensive 3-D model or scene. It has significant applications in the field of computer vision and robotics. In recent years, with the continuous impact of machinevision technology, more and more registration methods based on deep network models have emerged. However, most deep learning-based methods perform poorly in low overlap scenarios. Therefore, this article proposes a novel network architecture to pursue better performance with low overlapping regions. A spatial rotation feature encoder with point attention (PASR) is addressed to improve the rotation invariance of point clouds and enhance the network's perception of local feature extraction. Then this article introduces the overlap attention prediction (OAP) module for the estimation process of point cloud overlap factors. On this basis, the cross-attention mechanism is introduced to regress the initial transformation between two input point clouds. In addition, by combining the previous overlap factors, we constructed an iterative dual-branch similarity matrix learning (DBSML) integrated network which guides similarity estimation and further eliminates interference from non-overlapping points. Extensive experiments on ModelNet40 and our real datasets with noisy and partially overlapping point clouds show that the proposed method outperforms the traditional and mainstream learning-based methods, achieving the state-of-the-art performance. In particular, we also verify the effectiveness and superiority of the network model in coping with multiple registration task scenarios.
The demand for mobile roboticsapplications has grown considerably in recent years, especially due to the advent of industry 4.0, which has as one of its pillars the autonomous robotics field, the subject of this rese...
详细信息
The demand for mobile roboticsapplications has grown considerably in recent years, especially due to the advent of industry 4.0, which has as one of its pillars the autonomous robotics field, the subject of this research. In this context, autonomous mobile robots must interact with the world to achieve their goals. One of the main challenges regarding mobile robots is the navigation problem: a robot can face several problems according to the type of sensor that is chosen in each application. The use of computer vision as a navigation tool in robotics represents an interesting alternative for controlling the movement of a mobile robot, and represent several vision techniques that gained more space in the last few years. Therefore, this work's research proposes the development of a control center to assist navigation and location of mobile robots in closed environments using the global view technique. In addition to computer vision, wireless communication (WiFi) between the exchange and the robots has been investigated to date. The results obtained in the initial steps of the project's development were promising, in which data from an autonomous robot is compared with a human-guided robot. Through the algorithm developed for the project, it was possible to transform the collected data into the robot's kinematics necessary to take the correct path to the destination using multivalued logic as a control algorithm. The optimization of the trajectory between the origin and the destination is performed using the A* and Dijkstra algorithms for calculating the shortest path.
The role of technology is vital and can be observed through the 5th industrial revolution. As a matter of fact, the impact is so severe, it can be felt almost everywhere. As technology advances, one of the most promis...
详细信息
machine Learning applications Practical resource on the importance of machine Learning and Deep Learning applications in various technologies and real-world situations machine Learning applications discusses methodolo...
详细信息
ISBN:
(数字)9781394173358
ISBN:
(纸本)9781394173327
machine Learning applications Practical resource on the importance of machine Learning and Deep Learning applications in various technologies and real-world situations machine Learning applications discusses methodological advancements of machine learning and deep learning, presents applications in image processing, including face and vehicle detection, image classification, object detection, image segmentation, and delivers real-world applications in healthcare to identify diseases and diagnosis, such as creating smart health records and medical imaging diagnosis, and provides real-world examples, case studies, use cases, and techniques to enable the reader’s active learning. Composed of 13 chapters, this book also introduces real-world applications of machine and deep learning in blockchain technology, cyber security, and climate change. An explanation of AI and robotic applications in mechanical design is also discussed, including robot-assisted surgeries, security, and space exploration. The book describes the importance of each subject area and detail why they are so important to us from a societal and human perspective. Edited by two highly qualified academics and contributed to by established thought leaders in their respective fields, machine Learning applications includes information on: Content based medical image retrieval (CBMIR), covering face and vehicle detection, multi-resolution and multisource analysis, manifold and image processing, and morphological processing Smart medicine, including machine learning and artificialintelligence in medicine, risk identification, tailored interventions, and association rules AI and robotics application for transportation and infrastructure (e.g., autonomous cars and smart cities), along with global warming and climate change Identifying diseases and diagnosis, drug discovery and manufacturing, medical imaging diagnosis, personalized medicine, and smart health records With its practical approach to the subject, Ma
The Joint Photographic Experts Group (JPEG) AI learning-based image coding system is an ongoing joint standardization effort between International Organization for Standardization (ISO), International Electrotechnical...
详细信息
The Joint Photographic Experts Group (JPEG) AI learning-based image coding system is an ongoing joint standardization effort between International Organization for Standardization (ISO), International Electrotechnical Commission (IEC), and International Telecommunication Union - Telecommunication Sector (ITU-T) for the development of the first image coding standard based on machine learning (a subset of artificialintelligence), offering a single stream, compact compressed domain representation, targeting both human visualization and machine consumption. The main motivation for this upcoming standard is the excellent performance of tools based on deep neural networks, in image coding, computer vision, and image processing tasks. The JPEG AI aims to develop an image coding standard addressing the needs of a wide range of applications such as cloud storage, visual surveillance, autonomous vehicles and devices, image collection storage and management, live monitoring of visual data, and media distribution. This article presents and discusses the rationale behind the JPEG AI vision, notably how this new standardization initiative aims to shape the future of image coding, through relevant application-driven use cases. The JPEG AI requirements, the JPEG AI history, and current status are also presented, offering a glimpse of the development of the first learning-based image coding standard.
The paper presents the application of a computer vision approach to tracking the mobile robot's state. As an exemplary environment, we use a feedback control system for the trajectory planning and control. The sys...
详细信息
ISBN:
(纸本)9783031425073;9783031425080
The paper presents the application of a computer vision approach to tracking the mobile robot's state. As an exemplary environment, we use a feedback control system for the trajectory planning and control. The system in the feedback loop use images taken from a centrally placed camera and, based on this, calculates the robots states, i.e. position and angle of rotation. The solution is adopted for indoor experiments. The experimental part shows the application of trajectory planning for multiple robots to cover a given area. The robot state is calculated using the YOLO model. We show that current machine learning techniques are fast and accurate for such applications and do not require image preprocessing or camera calibration.
Facial recognition is a widely-used process that aims to detect and verify an individual's identity. This technique is employed in various applications, such as image and video analysis, surveillance, and security...
详细信息
Robot person following (RPF) is a crucial capability in human-robot interaction (HRI) applications, allowing a robot to persistently follow a designated person. In practical RPF scenarios, the person can often be occl...
详细信息
Robot person following (RPF) is a crucial capability in human-robot interaction (HRI) applications, allowing a robot to persistently follow a designated person. In practical RPF scenarios, the person can often be occluded by other objects or people. Consequently, it is necessary to re-identify the person when he/she reappears within the robot's field of view. Previous person re-identification (ReID) approaches to person following rely on a fixed feature extractor. Such an approach often fails to generalize to different viewpoints and lighting conditions in practical RPF environments. In other words, it suffers from the so-called domain shift problem where it cannot re-identify the person when his re-appearance is out of the domain modeled by the fixed feature extractor. To mitigate this problem, we propose a ReID framework for RPF where we use a feature extractor that is optimized online with both short-term and long-term experiences (i.e., recently and previously observed samples during RPF) using the online continual learning (OCL) framework. The long-term experiences are maintained by a memory manager to enable OCL to update the feature extractor. Our experiments demonstrate that even in the presence of severe appearance changes and distractions from visually similar people, the proposed method can still re-identify the person more accurately than the state-of-the-art methods.
This paper explores the application of active inference and the free energy principle (FEP) to enable active vision in physical robots, using pixel-level RGB camera observations. By adapting existing methodologies pre...
详细信息
ISBN:
(纸本)9783031723582;9783031723599
This paper explores the application of active inference and the free energy principle (FEP) to enable active vision in physical robots, using pixel-level RGB camera observations. By adapting existing methodologies previously limited to simulated environments, we introduce architectural improvements, including spatial softmax, to address the challenges of real-world application. Our model demonstrates proficiency in both exploratory and goal-directed behaviors within complex environments, achieving a dynamic understanding of visual scenes from pixel data. Our findings further demonstrate the potential of active inference and the FEP for tackling active vision in real-world robotics, and in bridging the gap between artificial and biological systems, offering a robust framework for developing more adaptive and aware robotic agents.
暂无评论