default search action
Vincent Wan
Person information
Refine list
refinements active!
zoomed in on ?? of ?? records
view refined list in
export refined list as
2020 – today
- 2023
- [c37]Lev Finkelstein, Chun-an Chan, Vincent Wan, Heiga Zen, Rob Clark:
FiPPiE: A Computationally Efficient Differentiable method for Estimating Fundamental Frequency From Spectrograms. SSW 2023: 218-224 - 2022
- [c36]Lev Finkelstein, Heiga Zen, Norman Casagrande, Chun-an Chan, Ye Jia, Tom Kenter, Alexey Petelin, Jonathan Shen, Vincent Wan, Yu Zhang, Yonghui Wu, Rob Clark:
Training Text-To-Speech Systems From Synthetic Data: A Practical Approach For Accent Transfer Tasks. INTERSPEECH 2022: 4571-4575 - [i2]Lev Finkelstein, Heiga Zen, Norman Casagrande, Chun-an Chan, Ye Jia, Tom Kenter, Alexey Petelin, Jonathan Shen, Vincent Wan, Yu Zhang, Yonghui Wu, Rob Clark:
Training Text-To-Speech Systems From Synthetic Data: A Practical Approach For Accent Transfer Tasks. CoRR abs/2208.13183 (2022)
2010 – 2019
- 2019
- [c35]Tom Kenter, Vincent Wan, Chun-an Chan, Rob Clark, Jakub Vit:
CHiVE: Varying Prosody in Speech Synthesis with a Linguistically Driven Dynamic Hierarchical Conditional Variational Network. ICML 2019: 3331-3340 - [i1]Vincent Wan, Chun-an Chan, Tom Kenter, Jakub Vit, Rob Clark:
CHiVE: Varying Prosody in Speech Synthesis with a Linguistically Driven Dynamic Hierarchical Conditional Variational Network. CoRR abs/1905.07195 (2019) - 2017
- [c34]Vincent Wan, Yannis Agiomyrgiannakis, Hanna Silén, Jakub Vít:
Google's Next-Generation Real-Time Unit-Selection Synthesizer Using Sequence-to-Sequence LSTM-Based Autoencoders. INTERSPEECH 2017: 1143-1147 - 2016
- [j7]Sarah A. Cassidy, Björn Stenger, L. Van Dongen, Kayoko Yanagisawa, Robert Anderson, Vincent Wan, Simon Baron-Cohen, Roberto Cipolla:
Expressive visual text-to-speech as an assistive technology for individuals with autism spectrum conditions. Comput. Vis. Image Underst. 148: 193-200 (2016) - 2014
- [j6]Vincent Wan, Javier Latorre, Kayoko Yanagisawa, Norbert Braunschweiler, Langzhou Chen, Mark J. F. Gales, Masami Akamine:
Building HMM-TTS Voices on Diverse Data. IEEE J. Sel. Top. Signal Process. 8(2): 296-306 (2014) - [c33]Vincent Wan, Javier Latorre, Kayoko Yanagisawa, Mark J. F. Gales, Yannis Stylianou:
Cluster adaptive training of average voice models. ICASSP 2014: 280-284 - [c32]Xie Chen, Mark J. F. Gales, Kate M. Knill, Catherine Breslin, Langzhou Chen, K. K. Chin, Vincent Wan:
An initial investigation of long-term adaptation for meeting transcription. INTERSPEECH 2014: 954-958 - [c31]BalaKrishna Kolluru, Vincent Wan, Javier Latorre, Kayoko Yanagisawa, Mark J. F. Gales:
Generating multiple-accent pronunciations for TTS using joint sequence model interpolation. INTERSPEECH 2014: 1273-1277 - [c30]Javier Latorre, Vincent Wan, Kayoko Yanagisawa:
Voice expression conversion with factorised HMM-TTS models. INTERSPEECH 2014: 1514-1518 - [c29]Javier Latorre, Kayoko Yanagisawa, Vincent Wan, BalaKrishna Kolluru, Mark J. F. Gales:
Speech intonation for TTS: study on evaluation methodology. INTERSPEECH 2014: 2957-2961 - 2013
- [c28]Robert Anderson, Björn Stenger, Vincent Wan, Roberto Cipolla:
Expressive Visual Text-to-Speech Using Active Appearance Models. CVPR 2013: 3382-3389 - [c27]Vincent Wan, Robert Anderson, Art Blokland, Norbert Braunschweiler, Langzhou Chen, BalaKrishna Kolluru, Javier Latorre, Ranniery Maia, Björn Stenger, Kayoko Yanagisawa, Yannis Stylianou, Masami Akamine, Mark J. F. Gales, Roberto Cipolla:
Photo-realistic expressive text to talking head synthesis. INTERSPEECH 2013: 2667-2669 - [c26]Robert Anderson, Björn Stenger, Vincent Wan, Roberto Cipolla:
An expressive text-driven 3D talking head. SIGGRAPH Posters 2013: 80 - [c25]Kayoko Yanagisawa, Javier Latorre, Vincent Wan, Mark J. F. Gales, Simon King:
Noise robustness in HMM-TTS speaker adaptation. SSW 2013: 119-124 - 2012
- [j5]Thomas Hain, Lukás Burget, John Dines, Philip N. Garner, Frantisek Grézl, Asmaa El Hannani, Marijn Huijbregts, Martin Karafiát, Mike Lincoln, Vincent Wan:
Transcribing Meetings With the AMIDA Systems. IEEE Trans. Speech Audio Process. 20(2): 486-498 (2012) - [c24]Florian Eyben, Sabine Buchholz, Norbert Braunschweiler, Javier Latorre, Vincent Wan, Mark J. F. Gales, Kate M. Knill:
Unsupervised clustering of emotion and voice styles for expressive TTS. ICASSP 2012: 4009-4012 - [c23]Langzhou Chen, Mark J. F. Gales, Vincent Wan, Javier Latorre, Masami Akamine:
Exploring Rich Expressive Information from Audiobook Data Using Cluster Adaptive Training. INTERSPEECH 2012: 959-962 - [c22]Javier Latorre, Vincent Wan, Mark J. F. Gales, Langzhou Chen, K. K. Chin, Kate M. Knill, Masami Akamine:
Speech factorization for HMM-TTS based on cluster adaptive training. INTERSPEECH 2012: 971-974 - [c21]Vincent Wan, Javier Latorre, K. K. Chin, Langzhou Chen, Mark J. F. Gales, Heiga Zen, Kate M. Knill, Masami Akamine:
Combining multiple high quality corpora for improving HMM-TTS. INTERSPEECH 2012: 1135-1138 - 2011
- [c20]Roger C. F. Tucker, Dan Fry, Vincent Wan, Stuart N. Wrigley, Thomas Hain:
Extending Audio Notetaker to Browse WebASR Transcriptions. INTERSPEECH 2011: 3329-3330 - 2010
- [c19]Thomas Hain, Lukás Burget, John Dines, Philip N. Garner, Asmaa El Hannani, Marijn Huijbregts, Martin Karafiát, Mike Lincoln, Vincent Wan:
The AMIDA 2009 meeting transcription system. INTERSPEECH 2010: 358-361
2000 – 2009
- 2009
- [c18]Philip N. Garner, John Dines, Thomas Hain, Asmaa El Hannani, Martin Karafiát, Danil Korchagin, Mike Lincoln, Vincent Wan, Le Zhang:
Real-time ASR from meetings. INTERSPEECH 2009: 2119-2122 - 2008
- [c17]Thomas Hain, Asmaa El Hannani, Stuart N. Wrigley, Vincent Wan:
Automatic speech recognition for scientific purposes - webASR. INTERSPEECH 2008: 504-507 - [c16]James Carmichael, Vincent Wan, Phil D. Green:
Combining neural network and rule-based systems for dysarthria diagnosis. INTERSPEECH 2008: 2226-2229 - [c15]Vincent Wan, John Dines, Asmaa El Hannani, Thomas Hain:
Bob: A lexicon and pronunciation dictionary generator. SLT 2008: 217-220 - 2007
- [j4]Odette Scharenborg, Vincent Wan, Roger K. Moore:
Towards capturing fine phonetic variation in speech using articulatory features. Speech Commun. 49(10-11): 811-826 (2007) - [c14]Thomas Hain, Lukás Burget, John Dines, Giulia Garau, Martin Karafiát, David A. van Leeuwen, Mike Lincoln, Vincent Wan:
The 2007 AMI(DA) System for Meeting Transcription. CLEAR 2007: 414-428 - [c13]Thomas Hain, Vincent Wan, Lukás Burget, Martin Karafiát, John Dines, Jithendra Vepa, Giulia Garau, Mike Lincoln:
The AMI System for the Transcription of Speech in Meetings. ICASSP (4) 2007: 357-360 - [c12]Yago Pereiro-Estevan, Vincent Wan, Odette Scharenborg:
Finding Maximum Margin Segments in Speech. ICASSP (4) 2007: 937-940 - [c11]Odette Scharenborg, Mirjam Ernestus, Vincent Wan:
Segmentation of speech: child's play? INTERSPEECH 2007: 1953-1956 - [c10]Odette Scharenborg, Vincent Wan:
Can unquantised articulatory feature continuums be modelled? INTERSPEECH 2007: 2473-2476 - 2006
- [c9]Vincent Wan, Thomas Hain:
Strategies for Language Model Web-Data Collection. ICASSP (1) 2006: 1069-1072 - [c8]Thomas Hain, Lukás Burget, John Dines, Giulia Garau, Martin Karafiát, Mike Lincoln, Jithendra Vepa, Vincent Wan:
The AMI Meeting Transcription System: Progress and Performance. MLMI 2006: 419-431 - 2005
- [j3]Stuart N. Wrigley, Guy J. Brown, Vincent Wan, Steve Renals:
Speech and crosstalk detection in multichannel audio. IEEE Trans. Speech Audio Process. 13(1): 84-91 (2005) - [j2]Vincent Wan, Steve Renals:
Speaker verification using sequence discriminant support vector machines. IEEE Trans. Speech Audio Process. 13(2): 203-210 (2005) - [c7]Thomas Hain, John Dines, Giulia Garau, Martin Karafiát, Darren Moore, Vincent Wan, Roeland Ordelman, Steve Renals:
Transcription of conference room meetings: an investigation. INTERSPEECH 2005: 1661-1664 - [c6]Vincent Wan, James Carmichael:
Polynomial dynamic time warping kernel support vector machines for dysarthric speech recognition with sparse training data. INTERSPEECH 2005: 3321-3324 - [c5]Thomas Hain, Lukás Burget, John Dines, Iain McCowan, Giulia Garau, Martin Karafiát, Mike Lincoln, Darren Moore, Vincent Wan, Roeland Ordelman, Steve Renals:
The Development of the AMI System for the Transcription of Speech in Meetings. MLMI 2005: 344-356 - [c4]Thomas Hain, Lukás Burget, John Dines, Giulia Garau, Martin Karafiát, Mike Lincoln, Iain McCowan, Darren Moore, Vincent Wan, Roeland Ordelman, Steve Renals:
The 2005 AMI System for the Transcription of Speech in Meetings. MLMI 2005: 450-462 - 2003
- [b1]Vincent Wan:
Speaker verification using support vector machines. University of Sheffield, UK, 2003 - [c3]Vincent Wan, Steve Renals:
SVMSVM: support vector machine speaker verification methodology. ICASSP (2) 2003: 221-224 - [c2]Stuart N. Wrigley, Guy J. Brown, Vincent Wan, Steve Renals:
Feature selection for the classification of crosstalk in multi-channel audio. INTERSPEECH 2003: 469-472 - 2002
- [c1]Vincent Wan, Steve Renals:
Evaluation of kernel methods for speaker verification and identification. ICASSP 2002: 669-672
1990 – 1999
- 1992
- [j1]Michael E. Bratman, Brian Harvey, Vincent Wan, Alice ter Meulen:
Book reviews. Minds Mach. 2(2): 203-215 (1992)
Coauthor Index
manage site settings
To protect your privacy, all features that rely on external API calls from your browser are turned off by default. You need to opt-in for them to become active. All settings here will be stored as cookies with your web browser. For more information see our F.A.Q.
Unpaywalled article links
Add open access links from to the list of external document links (if available).
Privacy notice: By enabling the option above, your browser will contact the API of unpaywall.org to load hyperlinks to open access articles. Although we do not have any reason to believe that your call will be tracked, we do not have any control over how the remote server uses your data. So please proceed with care and consider checking the Unpaywall privacy policy.
Archived links via Wayback Machine
For web page which are no longer available, try to retrieve content from the of the Internet Archive (if available).
Privacy notice: By enabling the option above, your browser will contact the API of archive.org to check for archived content of web pages that are no longer available. Although we do not have any reason to believe that your call will be tracked, we do not have any control over how the remote server uses your data. So please proceed with care and consider checking the Internet Archive privacy policy.
Reference lists
Add a list of references from , , and to record detail pages.
load references from crossref.org and opencitations.net
Privacy notice: By enabling the option above, your browser will contact the APIs of crossref.org, opencitations.net, and semanticscholar.org to load article reference information. Although we do not have any reason to believe that your call will be tracked, we do not have any control over how the remote server uses your data. So please proceed with care and consider checking the Crossref privacy policy and the OpenCitations privacy policy, as well as the AI2 Privacy Policy covering Semantic Scholar.
Citation data
Add a list of citing articles from and to record detail pages.
load citations from opencitations.net
Privacy notice: By enabling the option above, your browser will contact the API of opencitations.net and semanticscholar.org to load citation information. Although we do not have any reason to believe that your call will be tracked, we do not have any control over how the remote server uses your data. So please proceed with care and consider checking the OpenCitations privacy policy as well as the AI2 Privacy Policy covering Semantic Scholar.
OpenAlex data
Load additional information about publications from .
Privacy notice: By enabling the option above, your browser will contact the API of openalex.org to load additional information. Although we do not have any reason to believe that your call will be tracked, we do not have any control over how the remote server uses your data. So please proceed with care and consider checking the information given by OpenAlex.
last updated on 2025-01-09 13:22 CET by the dblp team
all metadata released as open data under CC0 1.0 license
see also: Terms of Use | Privacy Policy | Imprint