Scalable reordering models for SMT based on multiclass SVM
Scalable reordering models for SMT based on multiclass SVM
In state-of-the-art phrase-based statistical machine translation systems, modelling phrase reorderings is an important need to enhance naturalness of the translated outputs, particularly when the grammatical structures of the language pairs differ significantly. Posing phrase movements as a classification problem, we exploit recent developments in solving large-scale multiclass support vector machines. Using dual coordinate descent methods for learning, we provide a mechanism to shrink the amount of training data required for each iteration. Hence, we produce significant computational saving while preserving the accuracy of the models. Our approach is a couple of times faster than maximum entropy approach and more memory-efficient (50% reduction). Experiments were carried out on an Arabic-English corpus with more than a quarter of a billion words. We achieve BLEU score improvements on top of a strong baseline system with sparse reordering features.
65-84
Alrajeh, Abdullah
64acb5ae-6e8e-44a0-9afa-edd1658cd0cd
Niranjan, Mahesan
5cbaeea8-7288-4b55-a89c-c43d212ddd4f
18 April 2015
Alrajeh, Abdullah
64acb5ae-6e8e-44a0-9afa-edd1658cd0cd
Niranjan, Mahesan
5cbaeea8-7288-4b55-a89c-c43d212ddd4f
Alrajeh, Abdullah and Niranjan, Mahesan
(2015)
Scalable reordering models for SMT based on multiclass SVM.
The Prague Bulletin of Mathematical Linguistics, 103 (1), .
(doi:10.1515/pralin-2015-0004).
Abstract
In state-of-the-art phrase-based statistical machine translation systems, modelling phrase reorderings is an important need to enhance naturalness of the translated outputs, particularly when the grammatical structures of the language pairs differ significantly. Posing phrase movements as a classification problem, we exploit recent developments in solving large-scale multiclass support vector machines. Using dual coordinate descent methods for learning, we provide a mechanism to shrink the amount of training data required for each iteration. Hence, we produce significant computational saving while preserving the accuracy of the models. Our approach is a couple of times faster than maximum entropy approach and more memory-efficient (50% reduction). Experiments were carried out on an Arabic-English corpus with more than a quarter of a billion words. We achieve BLEU score improvements on top of a strong baseline system with sparse reordering features.
Text
art-alrajeh-niranjan.pdf
- Other
Available under License Other.
More information
Published date: 18 April 2015
Organisations:
Vision, Learning and Control
Identifiers
Local EPrints ID: 376612
URI: http://eprints.soton.ac.uk/id/eprint/376612
PURE UUID: 4a25a1f8-9155-4252-9752-66106f9b68a7
Catalogue record
Date deposited: 05 May 2015 17:25
Last modified: 15 Mar 2024 03:29
Export record
Altmetrics
Contributors
Author:
Abdullah Alrajeh
Author:
Mahesan Niranjan
Download statistics
Downloads from ePrints over the past year. Other digital versions may also be available to download e.g. from the publisher's website.
Loading...
View more statistics