END-TO-END UZBEK-RUSSIAN SPEECH TRANSLATION WITH SELF-SUPERVISED PRETRAINING

dc.contributor.authorSukhrob Avezov Sobirovich
dc.date.accessioned2025-12-29T14:28:39Z
dc.date.issued2025-09-27
dc.description.abstractIn this article we study end-to-end Uzbek→Russian speech translation under realistic low-resource and code-switching conditions. We couple a wav2vec-style encoder pre-trained on unlabeled audio with a Transformer decoder, add multi-task ASR/CTC objectives, and distill from a strong cascade teacher. Script-aware tokenization and data augmentation reduce sparsity. On conversational and broadcast tests the model improves BLEU/chrF at fixed latency and yields fewer morphology and NE errors.
dc.formatapplication/pdf
dc.identifier.urihttps://webofjournals.com/index.php/1/article/view/5127
dc.identifier.urihttps://asianeducationindex.com/handle/123456789/22185
dc.language.isoeng
dc.publisherWeb of Journals Publishing
dc.relationhttps://webofjournals.com/index.php/1/article/view/5127/5162
dc.rightshttps://creativecommons.org/licenses/by-nc-nd/4.0
dc.sourceWeb of Teachers: Inderscience Research ; Vol. 3 No. 9 (2025): WOT; 79-82
dc.source2938-379X
dc.subjectEnd-to-end speech translation, Uzbek-Russian, self-supervised pretraining, wav2vec 2.0, XLS-R, knowledge distillation, code-switching, low-resource.
dc.titleEND-TO-END UZBEK-RUSSIAN SPEECH TRANSLATION WITH SELF-SUPERVISED PRETRAINING
dc.typeinfo:eu-repo/semantics/article
dc.typeinfo:eu-repo/semantics/publishedVersion
dc.typePeer-reviewed Article

item.page.files

item.page.filesection.original.bundle

pagination.showing.labelpagination.showing.detail
loading.default
thumbnail.default.alt
item.page.filesection.name
sobirovich_2025_end-to-end_uzbek-russian_speech_translat.pdf
item.page.filesection.size
325.68 KB
item.page.filesection.format
Adobe Portable Document Format

item.page.collections