Text Generation
generative translation
large language model
LLaMA

This repo releases the trained LLaMA-adapter weights in paper "GenTranslate: Large Language Models are Generative Multilingual Speech and Machine Translators".

Code: https://github.com/YUCHEN005/GenTranslate

Data: https://huggingface.co/datasets/PeacefulData/HypoTranslate

Model: This repo

Filename format: [data_source]_[src_language_code]_[tgt_language_code]_[task].pth

e.g. covost2_ar_en_st.pth

Note:

  • Language code look-up: Table 15 & 17 in https://arxiv.org/pdf/2402.06894.pdf
  • Source/target language refers to the translation task, so that the N-best hypotheses and ground-truth transcription are both in target language
  • For speech translation datasets (FLEURS, CoVoST-2, MuST-C), the task ID "mt" denotes cascaded ASR+MT system

If you consider this work would be related or useful for your research, please kindly consider to cite the work below. Thank you.

@inproceedings{hu2024gentranslate,
    title = "GenTranslate: Large Language Models are Generative Multilingual Speech and Machine Translators",
    author = "Hu, Yuchen and Chen, Chen and Yang, Chao-Han Huck and Li, Ruizhe and Zhang, Dong and Chen, Zhehuai and Chng, Eng Siong",
    booktitle = "Proceedings of the 62nd Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers)",
    publisher = "Association for Computational Linguistics",
    year = "2024"
}
Downloads last month

-

Downloads are not tracked for this model. How to track
Inference Providers NEW
This model is not currently available via any of the supported Inference Providers.
The model cannot be deployed to the HF Inference API: The model has no library tag.

Dataset used to train PeacefulData/GenTranslate