咨询与建议

限定检索结果

文献类型

  • 23,001 篇 会议
  • 126 册 图书
  • 92 篇 期刊文献

馆藏范围

  • 23,218 篇 电子文献
  • 1 种 纸本馆藏

日期分布

学科分类号

  • 13,623 篇 工学
    • 11,108 篇 计算机科学与技术...
    • 3,479 篇 软件工程
    • 2,445 篇 机械工程
    • 1,716 篇 光学工程
    • 1,075 篇 电气工程
    • 1,014 篇 控制科学与工程
    • 785 篇 信息与通信工程
    • 412 篇 仪器科学与技术
    • 352 篇 生物工程
    • 251 篇 生物医学工程(可授...
    • 196 篇 电子科学与技术(可...
    • 114 篇 化学工程与技术
    • 108 篇 安全科学与工程
    • 100 篇 测绘科学与技术
    • 88 篇 建筑学
    • 87 篇 交通运输工程
    • 84 篇 土木工程
  • 3,494 篇 医学
    • 3,481 篇 临床医学
    • 81 篇 基础医学(可授医学...
  • 3,242 篇 理学
    • 1,939 篇 物理学
    • 1,640 篇 数学
    • 563 篇 统计学(可授理学、...
    • 500 篇 生物学
    • 249 篇 系统科学
    • 107 篇 化学
  • 522 篇 管理学
    • 311 篇 图书情报与档案管...
    • 224 篇 管理科学与工程(可...
    • 76 篇 工商管理
  • 276 篇 艺术学
    • 276 篇 设计学(可授艺术学...
  • 66 篇 法学
    • 63 篇 社会学
  • 38 篇 农学
  • 28 篇 教育学
  • 22 篇 经济学
  • 10 篇 军事学
  • 3 篇 文学

主题

  • 10,187 篇 computer vision
  • 3,967 篇 pattern recognit...
  • 3,005 篇 training
  • 2,007 篇 computational mo...
  • 1,818 篇 visualization
  • 1,815 篇 cameras
  • 1,516 篇 feature extracti...
  • 1,481 篇 shape
  • 1,455 篇 three-dimensiona...
  • 1,438 篇 image segmentati...
  • 1,287 篇 robustness
  • 1,205 篇 computer archite...
  • 1,155 篇 semantics
  • 1,147 篇 conferences
  • 1,107 篇 layout
  • 1,092 篇 computer science
  • 1,087 篇 object detection
  • 1,025 篇 benchmark testin...
  • 970 篇 codes
  • 922 篇 face recognition

机构

  • 136 篇 univ sci & techn...
  • 121 篇 univ chinese aca...
  • 118 篇 chinese univ hon...
  • 107 篇 carnegie mellon ...
  • 101 篇 tsinghua univers...
  • 101 篇 microsoft resear...
  • 95 篇 swiss fed inst t...
  • 93 篇 zhejiang univ pe...
  • 82 篇 university of sc...
  • 81 篇 zhejiang univers...
  • 80 篇 university of ch...
  • 77 篇 shanghai ai lab ...
  • 72 篇 shanghai jiao to...
  • 69 篇 national laborat...
  • 67 篇 microsoft res as...
  • 67 篇 alibaba grp peop...
  • 64 篇 adobe research
  • 61 篇 tsinghua univ pe...
  • 60 篇 peking univ peop...
  • 59 篇 univ oxford oxfo...

作者

  • 81 篇 van gool luc
  • 72 篇 timofte radu
  • 64 篇 zhang lei
  • 47 篇 luc van gool
  • 40 篇 yang yi
  • 40 篇 li stan z.
  • 37 篇 loy chen change
  • 34 篇 chen chen
  • 33 篇 xiaoou tang
  • 32 篇 liu yang
  • 32 篇 qi tian
  • 31 篇 tian qi
  • 31 篇 sun jian
  • 30 篇 murino vittorio
  • 30 篇 pascal fua
  • 29 篇 darrell trevor
  • 29 篇 li fei-fei
  • 28 篇 li xin
  • 28 篇 ying shan
  • 27 篇 vasconcelos nuno

语言

  • 23,137 篇 英文
  • 53 篇 其他
  • 22 篇 中文
  • 5 篇 土耳其文
  • 2 篇 日文
检索条件"任意字段=IEEE Conference on Computer Vision and Pattern Recognition Workshops"
23219 条 记 录,以下是621-630 订阅
排序:
Efficient Deformable ConvNets: Rethinking Dynamic and Sparse Operator for vision Applications
Efficient Deformable ConvNets: Rethinking Dynamic and Sparse...
收藏 引用
ieee/CVF conference on computer vision and pattern recognition (CVPR)
作者: Xiong, Yuwen Li, Zhiqi Chen, Yuntao Wang, Feng Zhu, Xizhou Luo, Jiapeng Wang, Wenhai Lu, Tong Li, Hongsheng Qiao, Yu Lu, Lewei Zhou, Jie Dai, Jifeng Univ Toronto Toronto ON Canada Shanghai AI Lab OpenGVLab Shanghai Peoples R China Nanjing Univ Nanjing Peoples R China Chinese Acad Sci CAIR HKISI Beijing Peoples R China Tsinghua Univ Beijing Peoples R China SenseTime Res Hong Kong Peoples R China Chinese Univ Hong Kong Hong Kong Peoples R China
We introduce Deformable Convolution v4 (DCNv4), a highly efficient and effective operator designed for a broad spectrum of vision applications. DCNv4 addresses the limitations of its predecessor, DCNv3, with two key e...
来源: 评论
Panda-70M: Captioning 70M Videos with Multiple Cross-Modality Teachers
Panda-70M: Captioning 70M Videos with Multiple Cross-Modalit...
收藏 引用
ieee/CVF conference on computer vision and pattern recognition (CVPR)
作者: Chen, Tsai-Shien Siarohin, Aliaksandr Menapace, Willi Deyneka, Ekaterina Chao, Hsiang-wei Jeon, Byung Eun Fang, Yuwei Lee, Hsin-Ying Ren, Jian Yang, Ming-Hsuan Tulyakov, Sergey Snap Inc Santa Monica CA 90405 USA Univ Calif Merced Merced CA 95343 USA Univ Trento Trento Italy Snap Santa Monica CA USA
The quality of the data and annotation upper-bounds the quality of a downstream model. While there exist large text corpora and image-text pairs, high-quality video-text data is much harder to collect. First of all, m... 详细信息
来源: 评论
OmniSeg3D: Omniversal 3D Segmentation via Hierarchical Contrastive Learning
OmniSeg3D: Omniversal 3D Segmentation via Hierarchical Contr...
收藏 引用
ieee/CVF conference on computer vision and pattern recognition (CVPR)
作者: Ying, Haiyang Yin, Yixuan Zhang, Jinzhi Wang, Fan Yu, Tao Huang, Ruqi Fang, Lu Tsinghua Univ Beijing Peoples R China Alibaba Grp Beijing Peoples R China
Towards holistic understanding of 3D scenes, a general 3D segmentation method is needed that can segment diverse objects without restrictions on object quantity or categories, while also reflecting the inherent hierar... 详细信息
来源: 评论
Egocentric Whole-Body Motion Capture with FisheyeViT and Diffusion-Based Motion Refinement
Egocentric Whole-Body Motion Capture with FisheyeViT and Dif...
收藏 引用
ieee/CVF conference on computer vision and pattern recognition (CVPR)
作者: Wang, Jian Cao, Zhe Luvizon, Diogo Liu, Lingjie Sarkar, Kripasindhu Tang, Danhang Beeler, Thabo Theobalt, Christian MPI Informat & Saarland Informat Campus Saarbrucken Germany Google Mountain View CA USA Univ Penn Philadelphia PA USA Saarbrucken Res Ctr Visual Com Interact & Artific Saarbrucken Germany
In this work, we explore egocentric whole-body motion capture using a single fisheye camera, which simultaneously estimates human body and hand motion. This task presents significant challenges due to three factors: t... 详细信息
来源: 评论
HRVDA: High-Resolution Visual Document Assistant
HRVDA: High-Resolution Visual Document Assistant
收藏 引用
ieee/CVF conference on computer vision and pattern recognition (CVPR)
作者: Liu, Chaohu Yin, Kun Cao, Haoyu Jiang, Xinghua Li, Xin Liu, Yinsong Jiang, Deqiang Sun, Xing Xu, Linli Univ Sci & Technol China Sch Comp Sci & Technol Hefei Anhui Peoples R China State Key Lab Cognit Intelligence Hefei Anhui Peoples R China Tencent YouTu Lab Shanghai Peoples R China
Leveraging vast training data, multimodal large language models (MLLMs) have demonstrated formidable general visual comprehension capabilities and achieved remarkable performance across various tasks. However, their p... 详细信息
来源: 评论
What, when, and where? Self-Supervised Spatio-Temporal Grounding in Untrimmed Multi-Action Videos from Narrated Instructions
What, when, and where? Self-Supervised Spatio-Temporal Groun...
收藏 引用
ieee/CVF conference on computer vision and pattern recognition (CVPR)
作者: Chen, Brian Shvetsova, Nina Rouditchenko, Andrew Kondermann, Daniel Thomas, Samuel Chang, Shih-Fu Feris, Rogerio Glass, James Kuehne, Hilde Columbia Univ New York NY 10027 USA Goethe Univ Frankfurt Frankfurt Germany Univ Bonn Bonn Germany MIT CSAIL Cambridge MA USA Qual Match GmbH Heidelberg Germany IBM Res AI Yorktown Hts NY USA MIT IBM Watson Lab Cambridge MA USA
Spatio-temporal grounding describes the task of localizing events in space and time, e.g., in video data, based on verbal descriptions only. Models for this task are usually trained with human-annotated sentences and ... 详细信息
来源: 评论
Leveraging Cross-Modal Neighbor Representation for Improved CLIP Classification
Leveraging Cross-Modal Neighbor Representation for Improved ...
收藏 引用
ieee/CVF conference on computer vision and pattern recognition (CVPR)
作者: Yi, Chao Ren, Lu Zhan, De-Chuan Ye, Han-Jia Nanjing Univ Natl Key Lab Novel Software Technol Nanjing Peoples R China Nanjing Univ Sch Artificial Intelligence Nanjing Peoples R China
CLIP showcases exceptional cross-modal matching capabilities due to its training on image-text contrastive learning tasks. However, without specific optimization for unimodal scenarios, its performance in single-modal... 详细信息
来源: 评论
Single-Model and Any-Modality for Video Object Tracking
Single-Model and Any-Modality for Video Object Tracking
收藏 引用
ieee/CVF conference on computer vision and pattern recognition (CVPR)
作者: Wu, Zongwei Zheng, Jilai Ren, Xiangxuan Vasluianu, Florin-Alexandru Ma, Chao Paudel, Danda Pani Luc Van Gool Timofte, Radu Univ Wurzburg Comp Vision Lab CAIDAS & IFI Wurzburg Germany Shanghai Jiao Tong Univ AI Inst Shanghai Peoples R China Sofia Univ INSAIT Sofia Bulgaria Swiss Fed Inst Technol CVL Zurich Switzerland
In the realm of video object tracking, auxiliary modalities such as depth, thermal, or event data have emerged as valuable assets to complement the RGB trackers. In practice, most existing RGB trackers learn a single ...
来源: 评论
360Loc: A Dataset and Benchmark for Omnidirectional Visual Localization with Cross-device Queries
360Loc: A Dataset and Benchmark for Omnidirectional Visual L...
收藏 引用
ieee/CVF conference on computer vision and pattern recognition (CVPR)
作者: Huang, Huajian Liu, Changkun Zhu, Yipeng Cheng, Hui Braud, Tristan Yeung, Sai-Kit Hong Kong Univ Sci & Technol Hong Kong Peoples R China Sun Yat Sen Univ Guangzhou Peoples R China
Portable 360 degrees cameras are becoming a cheap and efficient tool to establish large visual databases. By capturing omnidirectional views of a scene, these cameras could expedite building environment models that ar... 详细信息
来源: 评论
Semantically Grounded Visual Embeddings for Zero-Shot Learning
Semantically Grounded Visual Embeddings for Zero-Shot Learni...
收藏 引用
ieee/CVF conference on computer vision and pattern recognition (CVPR)
作者: Nawaz, Shah Cavazza, Jacopo Del Bue, Alessio Ist Italiano Tecnol IIT Pattern Anal & Comp Vis PAVIS Genoa Italy Ist Italiano Tecnol IIT Visual Geometry & Modelling VGM Genoa Italy Deutsch Elektronen Synchrotron DESY Hamburg Germany
Zero-shot learning methods rely on fixed visual and semantic embeddings, extracted from independent vision and language models, both pre-trained for other large-scale tasks. This is a weakness of current zero-shot lea... 详细信息
来源: 评论