User profiles for Shanbo Cheng
![]() | Shanbo ChengByteDance Seed Verified email at bytedance.com Cited by 682 |
Eliciting the translation ability of large language models via multilingual finetuning with translation instructions
Large-scale pretrained language models (LLMs), such as ChatGPT and GPT4, have shown
strong abilities in multilingual translation, without being explicitly trained on parallel corpora. …
strong abilities in multilingual translation, without being explicitly trained on parallel corpora. …
Acquiring knowledge from pre-trained model to neural machine translation
Pre-training and fine-tuning have achieved great success in natural language process field.
The standard paradigm of exploiting them includes two steps: first, pre-training a model, eg …
The standard paradigm of exploiting them includes two steps: first, pre-training a model, eg …
[PDF][PDF] Sogou neural machine translation systems for WMT17
We describe the Sogou neural machine translation systems for the WMT 2017 Chinese↔
English news translation tasks. Our systems are based on a multilayer encoder-decoder …
English news translation tasks. Our systems are based on a multilayer encoder-decoder …
Speech translation with large language models: An industrial practice
Given the great success of large language models (LLMs) across various tasks, in this paper,
we introduce LLM-ST, a novel and effective speech translation model constructed upon a …
we introduce LLM-ST, a novel and effective speech translation model constructed upon a …
G-dig: Towards gradient-based diverse and high-quality instruction data selection for machine translation
Large Language Models (LLMs) have demonstrated remarkable abilities in general scenarios.
Instruction finetuning empowers them to align with humans in various tasks. Nevertheless…
Instruction finetuning empowers them to align with humans in various tasks. Nevertheless…
Retaining key information under high compression ratios: Query-guided compressor for llms
The growing popularity of Large Language Models has sparked interest in context compression
for Large Language Models (LLMs). However, the performance of previous methods …
for Large Language Models (LLMs). However, the performance of previous methods …
Towards achieving human parity on end-to-end simultaneous speech translation via llm agent
In this paper, we present Cross Language Agent -- Simultaneous Interpretation, CLASI, a
high-quality and human-like Simultaneous Speech Translation (SiST) System. Inspired by …
high-quality and human-like Simultaneous Speech Translation (SiST) System. Inspired by …
Language tags matter for zero-shot neural machine translation
Multilingual Neural Machine Translation (MNMT) has aroused widespread interest due to its
efficiency. An exciting advantage of MNMT models is that they could also translate between …
efficiency. An exciting advantage of MNMT models is that they could also translate between …
Learning kernel-smoothed machine translation with retrieved examples
How to effectively adapt neural machine translation (NMT) models according to emerging
cases without retraining? Despite the great success of neural machine translation, updating …
cases without retraining? Despite the great success of neural machine translation, updating …
Language-aware interlingua for multilingual neural machine translation
Multilingual neural machine translation (NMT) has led to impressive accuracy improvements
in low-resource scenarios by sharing common linguistic information across languages. …
in low-resource scenarios by sharing common linguistic information across languages. …