Projects per year
Abstract
Attention mechanisms in neural networks have proved useful for problems in which the input and output do not have fixed dimension. Often there exist features that are locally translation invariant and would be valuable for directing the model’s attention, but previous attentional architectures are not constructed to learn such features specifically. We introduce an attentional neural network that employs convolution on the input tokens to detect local time-invariant and
long-range topical attention features in a context-dependent way. We apply this architecture to the problem of extreme summarization of source code snippets into short, descriptive function name-like summaries. Using those features, the model sequentially generates a summary by marginalizing over two attention mechanisms: one that predicts the next summary token based on the attention weights of the input tokens and another that is able to copy a code token as-is directly into the summary. We demonstrate our convolutional attention neural network’s performance on 10 popular Java projects showing that it achieves better performance compared to previous attentional mechanisms.
long-range topical attention features in a context-dependent way. We apply this architecture to the problem of extreme summarization of source code snippets into short, descriptive function name-like summaries. Using those features, the model sequentially generates a summary by marginalizing over two attention mechanisms: one that predicts the next summary token based on the attention weights of the input tokens and another that is able to copy a code token as-is directly into the summary. We demonstrate our convolutional attention neural network’s performance on 10 popular Java projects showing that it achieves better performance compared to previous attentional mechanisms.
| Original language | English |
|---|---|
| Title of host publication | Proceedings of The 33rd International Conference on Machine Learning, PMLR |
| Place of Publication | New York, United States |
| Publisher | PMLR |
| Pages | 2091-2100 |
| Number of pages | 10 |
| Volume | 48 |
| Publication status | Published - 24 Jun 2016 |
| Event | 33rd International Conference on Machine Learning: ICML 2016 - New York, United States Duration: 19 Jun 2016 → 24 Jun 2016 https://icml.cc/Conferences/2016/ |
Publication series
| Name | Proceedings of Machine Learning Research |
|---|---|
| Publisher | PMLR |
| Volume | 48 |
| ISSN (Electronic) | 2640-3498 |
Conference
| Conference | 33rd International Conference on Machine Learning |
|---|---|
| Abbreviated title | ICML 2016 |
| Country/Territory | United States |
| City | New York |
| Period | 19/06/16 → 24/06/16 |
| Internet address |
Fingerprint
Dive into the research topics of 'A Convolutional Attention Network for Extreme Summarization of Source Code'. Together they form a unique fingerprint.Projects
- 1 Finished
-
Statistical Natural Language Processing Methods for Computer Program Source Code
Sutton, C. (Principal Investigator)
1/10/13 → 31/03/17
Project: Research