咨询与建议

限定检索结果

文献类型

  • 267 篇 会议
  • 154 篇 期刊文献

馆藏范围

  • 421 篇 电子文献
  • 0 种 纸本馆藏

日期分布

学科分类号

  • 282 篇 工学
    • 184 篇 计算机科学与技术...
    • 164 篇 软件工程
    • 111 篇 信息与通信工程
    • 28 篇 生物工程
    • 27 篇 电子科学与技术(可...
    • 24 篇 电气工程
    • 23 篇 控制科学与工程
    • 21 篇 仪器科学与技术
    • 19 篇 化学工程与技术
    • 11 篇 机械工程
    • 8 篇 生物医学工程(可授...
    • 6 篇 光学工程
    • 5 篇 建筑学
    • 4 篇 土木工程
    • 3 篇 材料科学与工程(可...
  • 176 篇 理学
    • 137 篇 物理学
    • 56 篇 数学
    • 31 篇 生物学
    • 19 篇 化学
    • 16 篇 统计学(可授理学、...
    • 8 篇 系统科学
  • 44 篇 管理学
    • 37 篇 图书情报与档案管...
    • 7 篇 管理科学与工程(可...
  • 11 篇 法学
    • 11 篇 社会学
  • 8 篇 医学
    • 7 篇 临床医学
    • 6 篇 基础医学(可授医学...
    • 5 篇 药学(可授医学、理...
  • 7 篇 文学
    • 6 篇 中国语言文学
    • 5 篇 外国语言文学
  • 4 篇 教育学
    • 4 篇 教育学
  • 3 篇 农学
  • 2 篇 艺术学

主题

  • 59 篇 speech recogniti...
  • 51 篇 training
  • 33 篇 acoustics
  • 31 篇 speech
  • 20 篇 speech processin...
  • 19 篇 feature extracti...
  • 18 篇 hidden markov mo...
  • 18 篇 signal processin...
  • 16 篇 computational mo...
  • 15 篇 conferences
  • 14 篇 speech enhanceme...
  • 13 篇 predictive model...
  • 13 篇 decoding
  • 12 篇 machine translat...
  • 11 篇 speech synthesis
  • 10 篇 training data
  • 10 篇 neural networks
  • 10 篇 data models
  • 9 篇 transformers
  • 9 篇 self-supervised ...

机构

  • 70 篇 national enginee...
  • 51 篇 human language t...
  • 45 篇 center for langu...
  • 31 篇 human language t...
  • 21 篇 center for langu...
  • 21 篇 center for langu...
  • 13 篇 center for langu...
  • 11 篇 iflytek research
  • 10 篇 center for langu...
  • 9 篇 ict cluster sing...
  • 9 篇 human language t...
  • 8 篇 national enginee...
  • 8 篇 center for langu...
  • 8 篇 human language t...
  • 7 篇 center for langu...
  • 7 篇 human language t...
  • 7 篇 university of sc...
  • 7 篇 xiaomi corp.
  • 6 篇 university of sc...
  • 6 篇 state key labora...

作者

  • 49 篇 ling zhen-hua
  • 47 篇 khudanpur sanjee...
  • 35 篇 dehak najim
  • 32 篇 ai yang
  • 29 篇 sanjeev khudanpu...
  • 23 篇 dredze mark
  • 22 篇 zhen-hua ling
  • 19 篇 povey daniel
  • 18 篇 villalba jesús
  • 18 篇 van durme benjam...
  • 18 篇 daniel povey
  • 18 篇 yang ai
  • 17 篇 post matt
  • 16 篇 hermansky hynek
  • 16 篇 lu ye-xin
  • 15 篇 zelasko piotr
  • 14 篇 du hui-peng
  • 13 篇 raj desh
  • 13 篇 gu jia-chen
  • 13 篇 watanabe shinji

语言

  • 364 篇 英文
  • 57 篇 其他
检索条件"机构=Center for Language and Speech Processing & Human Language Technology"
421 条 记 录,以下是81-90 订阅
排序:
PQLM - Multilingual Decentralized Portable Quantum language Model
PQLM - Multilingual Decentralized Portable Quantum Language ...
收藏 引用
International Conference on Acoustics, speech, and Signal processing (ICASSP)
作者: Shuyue Stella Li Xiangyu Zhang Shu Zhou Hongchao Shu Ruixing Liang Hexin Liu Leibny Paola Garcia Center for Language and Speech Processing Johns Hopkins University Department of Physics Hong Kong University of Science and Technology School of Electrical and Electronic Engineering Nanyang Technological University Human Language Technology Center of Excellence Johns Hopkins University
With careful manipulation, malicious agents can reverse engineer private information encoded in pre-trained language models. Security concerns motivate the development of quantum pre-training. In this work, we propose... 详细信息
来源: 评论
Is ChatGPT a Good Multi-Party Conversation Solver?
arXiv
收藏 引用
arXiv 2023年
作者: Tan, Chao-Hong Gu, Jia-Chen Ling, Zhen-Hua National Engineering Research Center of Speech and Language Information Processing University of Science and Technology of China Hefei China
Large language Models (LLMs) have emerged as influential instruments within the realm of natural language processing;nevertheless, their capacity to handle multi-party conversations (MPCs) – a scenario marked by the ... 详细信息
来源: 评论
speech RECONSTRUCTION FROM SILENT TONGUE AND LIP ARTICULATION BY PSEUDO TARGET GENERATION AND DOMAIN ADVERSARIAL TRAINING
arXiv
收藏 引用
arXiv 2023年
作者: Zheng, Rui-Chen Ai, Yang Ling, Zhen-Hua National Engineering Research Center of Speech and Language Information Processing University of Science and Technology of China Hefei China
This paper studies the task of speech reconstruction from ultrasound tongue images and optical lip videos recorded in a silent speaking mode, where people only activate their intra-oral and extra-oral articulators wit... 详细信息
来源: 评论
Stage-Wise and Prior-Aware Neural speech Phase Prediction
Stage-Wise and Prior-Aware Neural Speech Phase Prediction
收藏 引用
IEEE Spoken language technology Workshop
作者: Fei Liu Yang Ai Hui-Peng Du Ye-Xin Lu Rui-Chen Zheng Zhen-Hua Ling National Engineering Research Center of Speech and Language Information Processing University of Science and Technology of China Hefei P. R. China
This paper proposes a novel Stage-wise and Prior-aware Neural speech Phase Prediction (SP-NSPP) model, which predicts the phase spectrum from input amplitude spectrum by two-stage neural networks. In the initial prior... 详细信息
来源: 评论
Lightweight Audio-Visual Wake Word Spotting with Diverse Acoustic Knowledge Distillation
收藏 引用
IEEE Transactions on Circuits and Systems for Video technology 2025年
作者: Li, Ke-Wei Chen, Hang Du, Jun Zhou, Heng-Shun Siniscalchi, Sabato Marco Niu, Shu-Tong Xiong, Shi-Fu University of Science and Technology of China National Engineering Research Center of Speech and Language Information Processing Anhui Hefei China IFlytek Research Anhui Hefei China University of Palermo Italy
Audio-Visual Wake Word Spotting (AVWWS) aims to accurately detect user-defined keywords by leveraging the complementary nature of different modalities in challenging acoustic environments. However, two primary challen... 详细信息
来源: 评论
Voice Attribute Editing with Text Prompt
arXiv
收藏 引用
arXiv 2024年
作者: Sheng, Zhengyan Ai, Yang Liu, Li-Juan Pan, Jia Ling, Zhen-Hua National Engineering Research Center of Speech and Language Information Processing University of Science and Technology of China Hefei China iFLYTEK Research Hefei China
Despite recent advancements in speech generation with text prompt providing control over speech style, voice attributes in synthesized speech remain elusive and challenging to control. This paper introduces a novel ta... 详细信息
来源: 评论
APNet2: High-quality and High-efficiency Neural Vocoder with Direct Prediction of Amplitude and Phase Spectra
arXiv
收藏 引用
arXiv 2023年
作者: Du, Hui-Peng Lu, Ye-Xin Ai, Yang Ling, Zhen-Hua National Engineering Research Center of Speech and Language Information Processing University of Science and Technology of China Hefei China
In our previous work, we proposed a neural vocoder called APNet, which directly predicts speech amplitude and phase spectra with a 5 ms frame shift in parallel from the input acoustic features, and then reconstructs t... 详细信息
来源: 评论
Source-Filter-Based Generative Adversarial Neural Vocoder for High Fidelity speech Synthesis
arXiv
收藏 引用
arXiv 2023年
作者: Lu, Ye-Xin Ai, Yang Ling, Zhen-Hua National Engineering Research Center of Speech and Language Information Processing University of Science and Technology of China Hefei China
This paper proposes a source-filter-based generative adversarial neural vocoder named SF-GAN, which achieves high-fidelity waveform generation from input acoustic features by introducing F0-based source excitation sig... 详细信息
来源: 评论
ZERO-SHOT PERSONALIZED LIP-TO-speech SYNTHESIS WITH FACE IMAGE BASED VOICE CONTROL
arXiv
收藏 引用
arXiv 2023年
作者: Sheng, Zheng-Yan Ai, Yang Ling, Zhen-Hua National Engineering Research Center of Speech and Language Information Processing University of Science and Technology of China Hefei China
Lip-to-speech (Lip2speech) synthesis, which predicts corresponding speech from talking face images, has witnessed significant progress with various models and training strategies in a series of independent studies. Ho... 详细信息
来源: 评论
Incorporating Ultrasound Tongue Images for Audio-Visual speech Enhancement through Knowledge Distillation
arXiv
收藏 引用
arXiv 2023年
作者: Zheng, Rui-Chen Ai, Yang Ling, Zhen-Hua National Engineering Research Center of Speech and Language Information Processing University of Science and Technology of China Hefei China
Audio-visual speech enhancement (AV-SE) aims to enhance degraded speech along with extra visual information such as lip videos, and has been shown to be more effective than audio-only speech enhancement. This paper pr... 详细信息
来源: 评论