The AI-powered exercise tracking system presented in this project represents a new approach to revolutionize fitness routines using cutting-edge technologies. Leveraging advanced computer vision techniques, particular...
详细信息
ISBN:
(纸本)9798350373301;9798350373295
The AI-powered exercise tracking system presented in this project represents a new approach to revolutionize fitness routines using cutting-edge technologies. Leveraging advanced computer vision techniques, particularly poseestimation, the system offers real-time tracking and analysis of exercise repetitions during gym workouts. Focused initially on four fundamental exercises like Bicep Curls, Push-Ups, Squats, and Shoulder Press. The proposed system employs Machine Learning (ML) models to detect and classify correct and incorrect exercise forms. Integrating the power of Mediapipe's posedetection framework and speech recognition capabilities, the proposed system provides accurate feedback and guidance to users through a smart web application interface. Through the selection of a tech stack including Python, OpenCV, Streamlit, and Scikit-Learn, the system ensures seamless data processing, model training, and user interaction. By analyzing workout videos uploaded by users, the proposed system identifies and highlights areas of improvement in exercise technique, enhancing the effectiveness and safety of fitness routines. Moreover, the integration of speech recognition enables hands-free control, adding convenience and accessibility for users during their workouts. This research study not only aims to optimize individual fitness journeys but also contributes to advancing the field of fitness tracking through the application of Artificial Intelligence (AI) and computer vision technologies. By empowering users with personalized feedback and guidance, the proposed system promotes healthier habits, improved performance, and overall well-being.
In the field of industrial robotics, robotic arms have been significantly integrated, driven by their precise functionality and operational efficiency. We here propose a hybrid method for binpicking tasks using a coll...
详细信息
In the field of industrial robotics, robotic arms have been significantly integrated, driven by their precise functionality and operational efficiency. We here propose a hybrid method for binpicking tasks using a collaborative robot, or cobot combining the You Only Look Once version 5 (YOLOv5) convolutional neural network (CNN) model for object detection and pose estimation with traditional feature detection based on the features from accelerated segment test (FAST) technique, feature description using binary robust invariant scalable keypoints (BRISK) algorithms, and matching algorithms. By integrating these algorithms and utilizing a low-cost depth sensor camera for capturing depth and RGB images, the system enhances real -time object detection and pose estimation speed, facilitating accurate object manipulation by the robotic arm. Furthermore, the proposed method is implemented within the robot operating system (ROS) framework to provide a seamless platform for robotic control and integration. We compared our results with those of other methodologies, highlighting the superior objectdetection accuracy and processing speed of our hybrid approach. This integration of robotic arm, camera, and AI technology contributes to the development of industrial robotics, opening up new possibilities for automating challenging tasks and improving overall operational efficiency.
This paper introduces an improved algorithm for texture-less object detection and pose estimation in industrial scenes. In the template training stage, a multi-scale template training method is proposed to improve the...
详细信息
This paper introduces an improved algorithm for texture-less object detection and pose estimation in industrial scenes. In the template training stage, a multi-scale template training method is proposed to improve the sensitivity of LineMOD to template depth. When this method performs template matching, the test image is first divided into several regions, and then training templates with similar depth are selected according to the depth of each test image region. In this way, without traversing all the templates, the depth of the template used by the algorithm during template matching is kept close to the depth of the target object, which improves the speed of the algorithm while ensuring that the accuracy of recognition will not decrease. In addition, this paper also proposes a method called coarse positioning of objects. The method avoids a lot of useless matching operations, and further improves the speed of the algorithm. The experimental results show that the improved LineMOD algorithm in this paper can effectively solve the algorithm's template depth sensitivity problem.
Construction sites are dynamic, and the environment is changing fast, which means the collective safety equipment, such as fall protection barriers, should also be changed to keep it compliant with the construction co...
详细信息
To collect a human-annotated dataset for training deep convolutional neural networks is a very time-consuming and laborious process. To reduce this burden, we previously proposed an automated annotation by placing one...
详细信息
To collect a human-annotated dataset for training deep convolutional neural networks is a very time-consuming and laborious process. To reduce this burden, we previously proposed an automated annotation by placing one visual marker above the detection target object in the training phase. However, in this approach, occasionally the marker hides the object surface. To avoid this issue, we propose placing a pedestal with multiple markers at the bottom of the object. If we use multiple markers, the object can be annotated even when the object hides some of the markers. Besides that, the simple modification of placing the markers on the bottom allows the use of simple background masking to avoid the neural network learning the remaining markers in the training image as a feature of the object. Background masking can completely remove the markers during the training process. Experiments showed the proposed vision system using our automatic object annotation outperformed the vision system using manual annotation in terms of objectdetection, orientation estimation, and 2D position estimation while reducing the time required for dataset collection from 16.1 hours to 7.30 hours.
Real-time tracking of complex 3D objects has been shown to be a challenging task for industrial applications where robustness, accuracy and run-time performance are of critical importance. This paper presents a fully ...
详细信息
ISBN:
(纸本)9789898111210
Real-time tracking of complex 3D objects has been shown to be a challenging task for industrial applications where robustness, accuracy and run-time performance are of critical importance. This paper presents a fully automated object tracking system which is capable of overcoming some of the problems faced in industrial environments. This is achieved by combining a real-time tracking system with a fast objectdetection system for automatic initialization and re-initialization at run-time. This ensures robustness of objectdetection, and at the same time accuracy and speed of recursive tracking. For the initialization we build a compact representation of the object of interest using statistical learning techniques during an off-line learning phase, in order to achieve speed and reliability at run-time by imposing geometric and photometric consistency constraints. The proposed tracking system is based on a novel template management algorithm which is incorporated into the ESM algorithm. Experimental results demonstrate the robustness and high precision of tracking of complex industrial machines with poor textures under severe illumination conditions.
暂无评论