Deep Neural Transformer Model for Mono and Multi Lingual Machine Translation
dc.contributor.author | Khaber, Mohamed Islam | |
dc.contributor.author | Frahta, Nabila | |
dc.contributor.author | Moussaoui, Abdelouahab | |
dc.contributor.author | Saidi, Mohamed | |
dc.date.accessioned | 2024-03-12T17:34:45Z | |
dc.date.available | 2024-03-12T17:34:45Z | |
dc.date.issued | 2021-05-25 | |
dc.description.abstract | In recent years, the Transformers have emerged as the most relevant deep architecture, especially machine translation. These models, which are based on attention mechanisms, outperformed previous neural machine translation architectures in several tasks. This paper proposes a new architecture based on the transformer model for the monolingual and multilingual translation system. The tests were carried out on the IWSLT 2015 and 2016 dataset. The Transformers attention mechanism increases the accuracy to more than 92% that we can quantify by more than 4 BLEU points (a performance metric used in machine translation systems). | |
dc.identifier.isbn | 978-9931-9788-0-0 | |
dc.identifier.uri | http://dspace.univ-oeb.dz:4000/handle/123456789/18721 | |
dc.language.iso | en | |
dc.publisher | University of Oum El Bouaghi | |
dc.title | Deep Neural Transformer Model for Mono and Multi Lingual Machine Translation | |
dc.type | Article |
Files
Original bundle
1 - 1 of 1
No Thumbnail Available
- Name:
- Deep Neural Transformer Model for Mono and Multi Lingual Machine Translation.pdf
- Size:
- 441.28 KB
- Format:
- Adobe Portable Document Format
License bundle
1 - 1 of 1
No Thumbnail Available
- Name:
- license.txt
- Size:
- 1.71 KB
- Format:
- Item-specific license agreed upon to submission
- Description: