咨询与建议

限定检索结果

文献类型

  • 11,267 篇 会议
  • 14 篇 期刊文献

馆藏范围

  • 11,281 篇 电子文献
  • 0 种 纸本馆藏

日期分布

学科分类号

  • 7,859 篇 工学
    • 7,418 篇 计算机科学与技术...
    • 799 篇 机械工程
    • 390 篇 电气工程
    • 377 篇 软件工程
    • 224 篇 控制科学与工程
    • 68 篇 光学工程
    • 32 篇 信息与通信工程
    • 26 篇 生物工程
    • 10 篇 生物医学工程(可授...
    • 8 篇 化学工程与技术
    • 7 篇 电子科学与技术(可...
    • 6 篇 交通运输工程
    • 5 篇 安全科学与工程
    • 3 篇 仪器科学与技术
    • 2 篇 力学(可授工学、理...
    • 2 篇 材料科学与工程(可...
    • 2 篇 动力工程及工程热...
    • 2 篇 航空宇航科学与技...
  • 3,103 篇 医学
    • 3,102 篇 临床医学
    • 4 篇 基础医学(可授医学...
  • 297 篇 理学
    • 199 篇 系统科学
    • 69 篇 物理学
    • 27 篇 生物学
    • 24 篇 数学
    • 9 篇 统计学(可授理学、...
    • 7 篇 化学
  • 23 篇 管理学
    • 14 篇 图书情报与档案管...
    • 9 篇 管理科学与工程(可...
    • 4 篇 工商管理
  • 6 篇 法学
    • 6 篇 社会学
  • 2 篇 农学
  • 1 篇 教育学
  • 1 篇 艺术学

主题

  • 5,461 篇 computer vision
  • 2,564 篇 training
  • 2,118 篇 pattern recognit...
  • 1,632 篇 computational mo...
  • 1,454 篇 visualization
  • 1,325 篇 three-dimensiona...
  • 1,070 篇 semantics
  • 972 篇 codes
  • 968 篇 benchmark testin...
  • 930 篇 computer archite...
  • 885 篇 deep learning
  • 831 篇 task analysis
  • 729 篇 feature extracti...
  • 541 篇 conferences
  • 530 篇 neural networks
  • 526 篇 face recognition
  • 503 篇 transformers
  • 480 篇 object detection
  • 478 篇 image segmentati...
  • 469 篇 cameras

机构

  • 169 篇 univ sci & techn...
  • 146 篇 tsinghua univ pe...
  • 142 篇 univ chinese aca...
  • 142 篇 carnegie mellon ...
  • 132 篇 chinese univ hon...
  • 122 篇 peng cheng lab p...
  • 102 篇 zhejiang univ pe...
  • 96 篇 sensetime res pe...
  • 95 篇 swiss fed inst t...
  • 90 篇 shanghai ai lab ...
  • 86 篇 tsinghua univers...
  • 86 篇 stanford univ st...
  • 84 篇 shanghai jiao to...
  • 80 篇 zhejiang univers...
  • 79 篇 alibaba grp peop...
  • 79 篇 univ hong kong p...
  • 76 篇 peng cheng labor...
  • 76 篇 tech univ munich...
  • 74 篇 australian natl ...
  • 73 篇 peking univ peop...

作者

  • 67 篇 timofte radu
  • 60 篇 van gool luc
  • 50 篇 zhang lei
  • 43 篇 yang yi
  • 36 篇 loy chen change
  • 36 篇 tao dacheng
  • 31 篇 liu yang
  • 30 篇 zhou jie
  • 30 篇 chen chen
  • 30 篇 tian qi
  • 29 篇 sun jian
  • 28 篇 zha zheng-jun
  • 27 篇 qi tian
  • 27 篇 boxin shi
  • 26 篇 li xin
  • 26 篇 vasconcelos nuno
  • 26 篇 pollefeys marc
  • 24 篇 liu xiaoming
  • 24 篇 zheng wei-shi
  • 24 篇 luo ping

语言

  • 11,273 篇 英文
  • 7 篇 其他
  • 1 篇 中文
检索条件"任意字段=2020 IEEE/CVF Conference on Computer Vision and Pattern Recognition, CVPR 2020"
11281 条 记 录,以下是351-360 订阅
排序:
SC-Tune: Unleashing Self-Consistent Referential Comprehension in Large vision Language Models
SC-Tune: Unleashing Self-Consistent Referential Comprehensio...
收藏 引用
ieee/cvf conference on computer vision and pattern recognition (cvpr)
作者: Yue, Tongtian Cheng, Jie Guo, Longteng Dai, Xingyuan Zhao, Zijia He, Xingjian Xiong, Gang Lv, Yisheng Liu, Jing CASIA Lab Cognit & Decis Intelligence Complex Syst Beijing Peoples R China CASIA State Key Lab Multimodal Artificial Intelligence Beijing Peoples R China Univ Chinese Acad Sci Sch Artificial Intelligence Beijing Peoples R China
Recent trends in Large vision Language Models (LVLMs) research have been increasingly focusing on advancing beyond general image understanding towards more nuanced, object-level referential comprehension. In this pape...
来源: 评论
DVMNet: Computing Relative Pose for Unseen Objects Beyond Hypotheses
DVMNet: Computing Relative Pose for Unseen Objects Beyond Hy...
收藏 引用
ieee/cvf conference on computer vision and pattern recognition (cvpr)
作者: Zhao, Chen Zhang, Tong Dang, Zheng Salzmann, Mathieu Ecole Polytech Fed Lausanne Lausanne Switzerland ClearSpace SA Renens Switzerland
Determining the relative pose of an object between two images is pivotal to the success of generalizable object pose estimation. Existing approaches typically approximate the continuous pose representation with a larg... 详细信息
来源: 评论
Random Entangled Tokens for Adversarially Robust vision Transformer
Random Entangled Tokens for Adversarially Robust Vision Tran...
收藏 引用
ieee/cvf conference on computer vision and pattern recognition (cvpr)
作者: Gong, Huihui Dong, Mingjing Mao, Siqi Camtepe, Seyit Nepal, Surya Xu, Chang Univ Sydney Sydney NSW Australia CSIRO Data61 Eveleigh Australia City Univ Hong Kong Hong Kong Peoples R China Univ New South Wales Sydney NSW Australia
vision Transformers (ViTs) have emerged as a compelling alternative to Convolutional Neural Networks ( CNNs) in the realm of computer vision, showcasing tremendous potential. However, recent research has unveiled a su... 详细信息
来源: 评论
Segment and Caption Anything
Segment and Caption Anything
收藏 引用
ieee/cvf conference on computer vision and pattern recognition (cvpr)
作者: Huang, Xiaoke Wang, Jianfeng Tang, Yansong Zhang, Zheng Hue, Han Lu, Jiwen Wang, Lijuan Liu, Zicheng Tsinghua Univ Shenzhen Int Grad Sch Shenzhen Peoples R China Microsoft Shanghai Peoples R China Tsinghua Univ Dept Automat Beijing Peoples R China Adv Micro Devices Inc Beijing Peoples R China
We propose a method to efficiently equip the Segment Anything Model ( SAM) with the ability to generate regional captions. SAM presents strong generalizability to segment anything while is short for semantic understan... 详细信息
来源: 评论
MM-Narrator: Narrating Long-form Videos with Multimodal In-Context Learning
MM-Narrator: Narrating Long-form Videos with Multimodal In-C...
收藏 引用
ieee/cvf conference on computer vision and pattern recognition (cvpr)
作者: Zh, Chaoyi Lin, Kevin Yang, Zhengyuan Wang, Jianfeng Li, Linjie Lin, Chung-Ching Liu, Zicheng Wang, Lijuan Univ Sydney Sydney NSW Australia Microsoft Corp Redmond WA 98052 USA Adv Micro Devices Inc Santa Clara CA USA
We present MM-Narrator, a novel system leveraging GPT-4 with multimodal in-context learning for the generation of audio descriptions (AD). Unlike previous methods that primarily focused on downstream fine-tuning with ... 详细信息
来源: 评论
VideoGrounding-DINO: Towards Open-Vocabulary Spatio-Temporal Video Grounding
VideoGrounding-DINO: Towards Open-Vocabulary Spatio-Temporal...
收藏 引用
ieee/cvf conference on computer vision and pattern recognition (cvpr)
作者: Wasim, Syed Talal Naseer, Muzammal Khan, Salman Yang, Ming-Hsuan Khan, Fahad Shahbaz Mohamed Bin Zayed Univ AI Abu Dhabi U Arab Emirates Australian Natl Univ Canberra Australia Univ Calif Merced Merced CA USA Google Res Mountain View CA USA Linkoping Univ Linkoping Sweden
Video grounding aims to localize a spatio-temporal section in a video corresponding to an input text query. This paper addresses a critical limitation in current video grounding methodologies by introducing an Open-Vo... 详细信息
来源: 评论
Emergent Open-Vocabulary Semantic Segmentation from Off-the-shelf vision-Language Models
Emergent Open-Vocabulary Semantic Segmentation from Off-the-...
收藏 引用
ieee/cvf conference on computer vision and pattern recognition (cvpr)
作者: Luo, Jiayun Khandelwal, Siddhesh Sigal, Leonid Li, Boyang Nanyang Technol Univ Singapore Singapore Univ British Columbia Vector Inst AI Vancouver BC Canada
From image-text pairs, large-scale vision-language models (VLMs) learn to implicitly associate image regions with words, which prove effective for tasks like visual question answering. However, leveraging the learned ... 详细信息
来源: 评论
EgoThink: Evaluating First-Person Perspective Thinking Capability of vision-Language Models
EgoThink: Evaluating First-Person Perspective Thinking Capab...
收藏 引用
ieee/cvf conference on computer vision and pattern recognition (cvpr)
作者: Cheng, Sijie Guo, Zhicheng Wu, Jingwen Fang, Kechen Li, Peng Liu, Huaping Liu, Yang Tsinghua Univ Dept Comp Sci & Technol Beijing Peoples R China Tsinghua Univ Inst AI Ind Res AIR Beijing Peoples R China Univ Toronto Dept Elect & Comp Engn Toronto ON Canada Tsinghua Univ Zhili Coll Beijing Peoples R China 01 Ai Beijing Peoples R China
vision-language models (VLMs) have recently shown promising results in traditional downstream tasks. Evaluation studies have emerged to assess their abilities, with the majority focusing on the third-person perspectiv... 详细信息
来源: 评论
Generative Bias for Robust Visual Question Answering
Generative Bias for Robust Visual Question Answering
收藏 引用
ieee/cvf conference on computer vision and pattern recognition (cvpr)
作者: Cho, Jae Won Kim, Dong-Jin Ryu, Hyeonggon Kweon, In So Korea Adv Inst Sci & Technol Daejeon South Korea Hanyang Univ Seoul South Korea
The task of Visual Question Answering (VQA) is known to be plagued by the issue of VQA models exploiting biases within the dataset to make its final prediction. Various previous ensemble based debiasing methods have b... 详细信息
来源: 评论
MMA: Multi-Modal Adapter for vision-Language Models
MMA: Multi-Modal Adapter for Vision-Language Models
收藏 引用
ieee/cvf conference on computer vision and pattern recognition (cvpr)
作者: Yang, Lingxiao Zhang, Ru-Yuan Wang, Yanchen Xie, Xiaohua Sun Yat Sen Univ Guangzhou Peoples R China Shanghai Jiao Tong Univ Shanghai Peoples R China Stanford Univ Stanford CA USA
Pre-trained vision-Language Models (VLMs) have served as excellent foundation models for transfer learning in diverse downstream tasks. However, tuning VLMs for few-shot generalization tasks faces a discrimination - g... 详细信息
来源: 评论