咨询与建议

限定检索结果

文献类型

  • 20,860 篇 会议
  • 105 篇 期刊文献
  • 43 册 图书

馆藏范围

  • 21,007 篇 电子文献
  • 1 种 纸本馆藏

日期分布

学科分类号

  • 13,620 篇 工学
    • 11,056 篇 计算机科学与技术...
    • 2,652 篇 机械工程
    • 2,252 篇 软件工程
    • 914 篇 光学工程
    • 885 篇 电气工程
    • 529 篇 控制科学与工程
    • 477 篇 信息与通信工程
    • 216 篇 测绘科学与技术
    • 135 篇 生物工程
    • 127 篇 生物医学工程(可授...
    • 98 篇 电子科学与技术(可...
    • 92 篇 仪器科学与技术
    • 46 篇 安全科学与工程
    • 40 篇 建筑学
    • 40 篇 化学工程与技术
    • 39 篇 土木工程
    • 37 篇 交通运输工程
    • 35 篇 力学(可授工学、理...
    • 33 篇 航空宇航科学与技...
  • 3,494 篇 医学
    • 3,489 篇 临床医学
    • 32 篇 基础医学(可授医学...
  • 2,247 篇 理学
    • 1,145 篇 物理学
    • 1,081 篇 数学
    • 401 篇 生物学
    • 384 篇 统计学(可授理学、...
    • 245 篇 系统科学
    • 46 篇 化学
  • 343 篇 管理学
    • 176 篇 管理科学与工程(可...
    • 168 篇 图书情报与档案管...
    • 34 篇 工商管理
  • 31 篇 法学
  • 19 篇 农学
  • 15 篇 教育学
  • 8 篇 经济学
  • 5 篇 艺术学
  • 2 篇 军事学
  • 1 篇 文学

主题

  • 8,141 篇 computer vision
  • 2,886 篇 training
  • 2,841 篇 pattern recognit...
  • 1,809 篇 computational mo...
  • 1,715 篇 visualization
  • 1,493 篇 cameras
  • 1,433 篇 three-dimensiona...
  • 1,433 篇 feature extracti...
  • 1,366 篇 shape
  • 1,360 篇 face recognition
  • 1,243 篇 image segmentati...
  • 1,135 篇 robustness
  • 1,124 篇 semantics
  • 992 篇 computer archite...
  • 985 篇 object detection
  • 982 篇 layout
  • 959 篇 benchmark testin...
  • 935 篇 codes
  • 900 篇 computer science
  • 898 篇 object recogniti...

机构

  • 174 篇 univ sci & techn...
  • 158 篇 univ chinese aca...
  • 153 篇 carnegie mellon ...
  • 145 篇 chinese univ hon...
  • 109 篇 microsoft resear...
  • 103 篇 zhejiang univ pe...
  • 99 篇 swiss fed inst t...
  • 95 篇 tsinghua univers...
  • 90 篇 microsoft res as...
  • 90 篇 tsinghua univ pe...
  • 88 篇 shanghai ai lab ...
  • 81 篇 zhejiang univers...
  • 77 篇 alibaba grp peop...
  • 74 篇 hong kong univ s...
  • 73 篇 university of sc...
  • 72 篇 peking univ peop...
  • 72 篇 university of ch...
  • 68 篇 shanghai jiao to...
  • 66 篇 univ oxford oxfo...
  • 65 篇 google res mount...

作者

  • 80 篇 van gool luc
  • 70 篇 zhang lei
  • 58 篇 timofte radu
  • 48 篇 yang yi
  • 47 篇 luc van gool
  • 46 篇 xiaoou tang
  • 44 篇 tian qi
  • 43 篇 darrell trevor
  • 42 篇 loy chen change
  • 42 篇 sun jian
  • 41 篇 qi tian
  • 40 篇 li stan z.
  • 38 篇 li fei-fei
  • 37 篇 chen xilin
  • 36 篇 shan shiguang
  • 35 篇 zhou jie
  • 35 篇 vasconcelos nuno
  • 35 篇 liu yang
  • 35 篇 torralba antonio
  • 34 篇 liu xiaoming

语言

  • 20,982 篇 英文
  • 10 篇 中文
  • 7 篇 其他
  • 5 篇 土耳其文
  • 2 篇 日文
  • 2 篇 葡萄牙文
检索条件"任意字段=2016 IEEE Conference on Computer Vision and Pattern Recognition, CVPR 2016"
21008 条 记 录,以下是91-100 订阅
排序:
PIN: Positional Insert Unlocks Object Localisation Abilities in VLMs
PIN: Positional Insert Unlocks Object Localisation Abilities...
收藏 引用
ieee/CVF conference on computer vision and pattern recognition (cvpr)
作者: Dorkenwald, Michael Barazani, Nimrod Snoek, Cees G. M. Asano, Yuki M. Univ Amsterdam Amsterdam Netherlands
vision-Language Models (VLMs), such as Flamingo and GPT-4V, have shown immense potential by integrating large language models with vision systems. Nevertheless, these models face challenges in the fundamental computer... 详细信息
来源: 评论
Iterated Learning Improves Compositionality in Large vision-Language Models
Iterated Learning Improves Compositionality in Large Vision-...
收藏 引用
ieee/CVF conference on computer vision and pattern recognition (cvpr)
作者: Zheng, Chenhao Zhang, Jieyu Kembhavi, Aniruddha Krishna, Ranjay Univ Washington Seattle WA 98195 USA Univ Michigan Ann Arbor MI 48109 USA Allen Inst Artificial Intelligence Seattle WA USA
A fundamental characteristic common to both human vision and natural language is their compositional nature. Yet, despite the performance gains contributed by large vision and language pretraining, recent investigatio...
来源: 评论
Video2Game: Real-time, Interactive, Realistic and Browser-Compatible Environment from a Single Video
Video2Game: Real-time, Interactive, Realistic and Browser-Co...
收藏 引用
ieee/CVF conference on computer vision and pattern recognition (cvpr)
作者: Xia, Hongchi Lin, Zhi-Hao Ma, Wei-Chiu Wang, Shenlong Univ Illinois Champaign IL 61820 USA Shanghai Jiao Tong Univ Shanghai Peoples R China Cornell Univ Ithaca NY USA
Creating high-quality and interactive virtual environments, such as games and simulators, often involves complex and costly manual modeling processes. In this paper, we present Video2Game, a novel approach that automa... 详细信息
来源: 评论
Animate Anyone: Consistent and Controllable Image-to-Video Synthesis for Character Animation
Animate Anyone: Consistent and Controllable Image-to-Video S...
收藏 引用
ieee/CVF conference on computer vision and pattern recognition (cvpr)
作者: Hu, Li Alibaba Grp Inst Intelligent Comp Hangzhou Peoples R China
Character Animation aims to generating character videos from still images through driving signals. Currently, diffusion models have become the mainstream in visual generation research, owing to their robust generative... 详细信息
来源: 评论
StyleCineGAN: Landscape Cinemagraph Generation using a Pre-trained StyleGAN
StyleCineGAN: Landscape Cinemagraph Generation using a Pre-t...
收藏 引用
ieee/CVF conference on computer vision and pattern recognition (cvpr)
作者: Choi, Jongwoo Seo, Kwanggyoon Ashtari, Amirsaman Noh, Junyong Korea Adv Inst Sci & Technol Visual Media Lab Daejeon South Korea
We propose a method that can generate cinemagraphs automatically from a still landscape image using a pre-trained StyleGAN. Inspired by the success of recent unconditional video generation, we leverage a powerful pre-... 详细信息
来源: 评论
eTraM: Event-based Traffic Monitoring Dataset
eTraM: Event-based Traffic Monitoring Dataset
收藏 引用
ieee/CVF conference on computer vision and pattern recognition (cvpr)
作者: Verma, Aayush Atul Chakravarthi, Bharatesh Vaghela, Arpitsinh Wei, Hua Yang, Yezhou Arizona State Univ Tempe AZ 85287 USA
Event cameras, with their high temporal and dynamic range and minimal memory usage, have found applications in various fields. However, their potential in static traffic monitoring remains largely unexplored. To facil... 详细信息
来源: 评论
Unsupervised Video Domain Adaptation with Masked Pre-Training and Collaborative Self-Training
Unsupervised Video Domain Adaptation with Masked Pre-Trainin...
收藏 引用
ieee/CVF conference on computer vision and pattern recognition (cvpr)
作者: Reddy, Arun Paul, William Rivera, Corban Shah, Ketul de Melo, Celso M. Chellappa, Rama Johns Hopkins Univ Baltimore MD 21218 USA Johns Hopkins Univ Dept Elect & Comp Engn Baltimore MD USA DEVCOM US Army Res Lab Aberdeen Proving Ground MD USA
In this work, we tackle the problem of unsupervised domain adaptation (UDA) for video action recognition. Our approach, which we call UNITE, uses an image teacher model to adapt a video student model to the target dom... 详细信息
来源: 评论
SpikingResformer: Bridging ResNet and vision Transformer in Spiking Neural Networks
SpikingResformer: Bridging ResNet and Vision Transformer in ...
收藏 引用
ieee/CVF conference on computer vision and pattern recognition (cvpr)
作者: Shi, Xinyu Hao, Zecheng Yu, Zhaofei Peking Univ Inst Artificial Intelligence Beijing Peoples R China Peking Univ Sch Comp Sci Beijing Peoples R China
The remarkable success of vision Transformers in Artificial Neural Networks (ANNs) has led to a growing interest in incorporating the self-attention mechanism and transformer-based architecture into Spiking Neural Net... 详细信息
来源: 评论
Dual Pose-invariant Embeddings: Learning Category and Object-specific Discriminative Representations for recognition and Retrieval
Dual Pose-invariant Embeddings: Learning Category and Object...
收藏 引用
ieee/CVF conference on computer vision and pattern recognition (cvpr)
作者: Sarkar, Rohan Kak, Avinash Purdue Univ Elect & Comp Engn W Lafayette IN 47907 USA
In the context of pose-invariant object recognition and retrieval, we demonstrate that it is possible to achieve significant improvements in performance if both the category-based and the object-identity-based embeddi... 详细信息
来源: 评论
CONFORM: Contrast is All You Need For High-Fidelity Text-to-Image Diffusion Models
CONFORM: Contrast is All You Need For High-Fidelity Text-to-...
收藏 引用
ieee/CVF conference on computer vision and pattern recognition (cvpr)
作者: Meral, Tuna Han Salih Simsar, Enis Tombari, Federico Yanardag, Pinar Virginia Tech Blacksburg VA USA Swiss Fed Inst Technol Zurich Switzerland TUM Munich Germany Google Menlo Pk CA USA
Images produced by text-to-image diffusion models might not always faithfully represent the semantic intent of the provided text prompt, where the model might overlook or entirely fail to produce certain objects. Exis... 详细信息
来源: 评论