Abstract | ||
---|---|---|
The task of translating between programming languages differs from the challenge of translating natural languages in that programming languages are designed with a far more rigid set of structural and grammatical rules. Previous work has used a tree-to-tree encoder/decoder model to take advantage of the inherent tree structure of programs during translation. Neural decoders, however, by default do not exploit known grammar rules of the target language. In this paper, we describe a tree decoder that leverages knowledge of a languageu0027s grammar rules to exclusively generate syntactically correct programs. We find that this grammar-based tree-to-tree model outperforms the state of the art tree-to-tree model in translating between two programming languages on a previously used synthetic task. |
Year | Venue | Field |
---|---|---|
2018 | arXiv: Learning | Programming language,Language translation,Decision tree model,Exploit,Grammar,Natural language,Tree structure,Artificial intelligence,Encoder,Machine learning,Mathematics |
DocType | Volume | Citations |
Journal | abs/1807.01784 | 1 |
PageRank | References | Authors |
0.36 | 0 | 8 |
Name | Order | Citations | PageRank |
---|---|---|---|
Mehdi Drissi | 1 | 2 | 1.04 |
Olivia Watkins | 2 | 2 | 0.70 |
Aditya Khant | 3 | 1 | 0.36 |
Vivaswat Ojha | 4 | 1 | 0.36 |
Pedro Sandoval | 5 | 3 | 1.07 |
Rakia Segev | 6 | 1 | 0.36 |
Eric Weiner | 7 | 1 | 0.36 |
Robert Keller | 8 | 2 | 1.72 |