Publication Type

Conference Proceeding Article

Version

acceptedVersion

Publication Date

11-2019

Abstract

Many Natural Language Processing (NLP) tasks, such as sentiment analysis or syntactic parsing, have benefited from the development of word embedding models. In particular, regardless of the training algorithms, the learned embeddings have often been shown to be generalizable to different NLP tasks. In contrast, despite recent momentum on word embeddings for source code, the literature lacks evidence of their generalizability beyond the example task they have been trained for. In this experience paper, we identify 3 potential downstream tasks, namely code comments generation, code authorship identification, and code clones detection, that source code token embedding models can be applied to. We empirically assess a recently proposed code token embedding model, namely code2vec’s token embeddings. Code2vec was trained on the task of predicting method names, and while there is potential for using the vectors it learns on other tasks, it has not been explored in literature. Therefore, we fill this gap by focusing on its generalizability for the tasks we have identified. Eventually, we show that source code token embeddings cannot be readily leveraged for the downstream tasks. Our experiments even show that our attempts to use them do not result in any improvements over less sophisticated methods. We call for more research into effective and general use of code embeddings

Keywords

Code Embeddings, Distributed Representations, Big Code

Discipline

Software Engineering

Research Areas

Software and Cyber-Physical Systems

Publication

2019 34th ACM/IEEE International Conference on Automated Software Engineering: San Diego, November 11-15: Proceedings

First Page

1

Last Page

12

ISBN

9781728125084

Identifier

10.1109/ASE.2019.00011

Publisher

IEEE

City or Country

Piscataway, NJ

Copyright Owner and License

Authors

Additional URL

https://doi.org/10.1109/ASE.2019.00011

Share

COinS