The increase in precision agriculture has promoted the development of picking robottechnology,and the visual recognition system at its core is crucial for improving the level of agricultural *** paper reviews the pro...
详细信息
The increase in precision agriculture has promoted the development of picking robottechnology,and the visual recognition system at its core is crucial for improving the level of agricultural *** paper reviews the progress of visual recognition tech-nology for picking robots,including image capture technology,target detection algorithms,spatial positioning strategies and scene *** article begins with a description of the basic structure and function of the vision system of the picking robot and em-phasizes the importance of achieving high-efficiency and high-accuracy recognition in the natural agricultural ***-sequently,various image processing techniques and vision algorithms,including color image analysis,three-dimensional depth percep-tion,and automatic object recognition technology that integrates machine learning and deep learning algorithms,were *** the same time,the paper also highlights the challenges of existing technologies in dynamic lighting,occlusion problems,fruit maturity di-versity,and real-time processing *** paper further discusses multisensor information fusion technology and discusses methods for combining visual recognition with a robotcontrol system to improve the accuracy and working rate of *** the same time,this paper also introduces innovative research,such as the application of convolutional neural networks(CNNs)for accurate fruit detection and the development of event-based vision systems to improve the response speed of the *** the end of this paper,the future development of visual recognition technology for picking robots is predicted,and new research trends are proposed,including the refinement of algorithms,hardware innovation,and the adaptability of technology to different agricultural *** purpose of this paper is to provide a comprehensive analysis of visual recognition technology for researchers and practitioners in the field of agricul-tural rob
This paper presents a data-driven predictive control method for optimizing the energy consumption of air-cooled data centers with unknown system model parameters. First, based on the measurable data of the studied sys...
详细信息
To address the issues of excessive maintenance and untimely maintenance of bearings, this paper proposes a performance evaluation method for bearing condition monitoring based on the combination of Principal Component...
详细信息
Domain adaptive semantic segmentation enables robust pixel- wise understanding in real-world driving scenes. Source-free domain adaptation, as a more practical technique, addresses the concerns of data privacy and sto...
详细信息
Domain adaptive semantic segmentation enables robust pixel- wise understanding in real-world driving scenes. Source-free domain adaptation, as a more practical technique, addresses the concerns of data privacy and storage limitations in typical unsupervised domain adaptation methods, making it especially relevant in the context of intelligent vehicles. It utilizes a well-trained source model and unlabeled target data to achieve adaptation in the target domain. However, in the absence of source data and target labels, current solutions cannot sufficiently reduce the impact of domain shift and fully leverage the information from the target data. In this paper, we propose an end-to-end source-free domain adaptation semantic segmentation method via Importance-Aware and Prototype-Contrast (IAPC) learning. The proposed IAPC framework effectively extracts domain-invariant knowledge from the well-trained source model and learns domain-specific knowledge from the unlabeled target domain. Specifically, considering the problem of domain shift in the prediction of the target domain by the source model, we put forward an importance-aware mechanism for the biased target prediction probability distribution to extract domain-invariant knowledge from the source model. We further introduce a prototype-contrast strategy, which includes a prototype-symmetric cross-entropy loss and a prototype-enhanced cross-entropy loss, to learn target intra-domain knowledge without relying on labels. A comprehensive variety of experiments on two domain adaptive semantic segmentation benchmarks demonstrates that the proposed end-to-end IAPC solution outperforms existing state-of-the-art methods. The source code is publicly available at https://***/yihong-97/Source-free-IAPC. IEEE
Multi-view stereo aims to recover the 3D model of a scene from a set of images. However, low-textured areas in the scene have always been a challenge in 3D reconstruction. In this work, we propose a segmentation-guide...
详细信息
Recognition and early warning of plant diseases is one of the keys to agricultural disaster prevention and mitigation. Deep learning-based image recognition methods give us a new idea for plant disease identification....
详细信息
In this paper, we propose LF-PGVIO, a Visual-Inertial-Odometry (VIO) framework for large Field-of-View (FoV) cameras with a negative plane using points and geodesic segments. The purpose of our research is to unleash ...
详细信息
In this paper, we propose LF-PGVIO, a Visual-Inertial-Odometry (VIO) framework for large Field-of-View (FoV) cameras with a negative plane using points and geodesic segments. The purpose of our research is to unleash the potential of point-line odometry with large-FoV omnidirectional cameras, even for cameras with negative-plane FoV. To achieve this, we propose an Omnidirectional Curve Segment Detection (OCSD) method combined with a camera model which is applicable to images with large distortions, such as panoramic annular images, fisheye images, and various panoramic images. The geodesic segment is sliced into multiple straight-line segments based on the radian and descriptors are extracted and recombined. Descriptor matching establishes the constraint relationship between 3D line segments in multiple frames. In our VIO system, line feature residual is also extended to support large-FoV cameras. Extensive evaluations on public datasets demonstrate the superior accuracy and robustness of LF-PGVIO compared to state-of-the-art methods. The source code will be made publicly available at https://***/flysoaryun/LF-PGVIO. IEEE
Chinese herbal oral liquid can leach a variety of effective ingredients from herbs and has become a major drug for clinical application. However, it is easy to produce or introduce foreign matters that are very faint ...
详细信息
vision sensors are widely applied in vehicles, robots, and roadside infrastructure. However, due to limitations in hardware cost and system size, camera Field-of-View (FoV) is often restricted and may not provide suff...
详细信息
vision sensors are widely applied in vehicles, robots, and roadside infrastructure. However, due to limitations in hardware cost and system size, camera Field-of-View (FoV) is often restricted and may not provide sufficient coverage. Nevertheless, from a spatiotemporal perspective, it is possible to obtain information beyond the camera's physical FoV from past video streams. In this paper, we propose the concept of online video inpainting for autonomous vehicles to expand the field of view, thereby enhancing scene visibility, perception, and system safety. To achieve this, we introduce the FlowLens architecture, which explicitly employs optical flow and implicitly incorporates a novel clip-recurrent transformer for feature propagation. FlowLens offers two key features: 1) FlowLens includes a newly designed Clip-Recurrent Hub with 3D-Decoupled Cross Attention (DDCA) to progressively process global information accumulated over time. 2) It integrates a multi-branch Mix Fusion Feed Forward Network (MixF3N) to enhance the precise spatial flow of local features. To facilitate training and evaluation, we derive the KITTI360 dataset with various FoV mask, which covers both outer- and inner FoV expansion scenarios. We also conduct quantitative assessments of beyond-FoV semantics across different models and perform qualitative comparisons of beyond-FoV object detection. We illustrate that employing FlowLens to reconstruct unseen scenes even enhances perception within the field of view by providing reliable semantic context. Extensive experiments and user studies involving offline and online video inpainting, as well as beyond-FoV perception tasks, demonstrate that FlowLens achieves state-of-the-art performance. The source code and dataset are made publicly available at https://***/MasterHow/FlowLens. IEEE
The deep convolution method based on MSDP signal imaging has been proven to be an effective means of monitoring the robot grinding process. This method has very high requirements on the quality of imaging and requires...
详细信息
暂无评论