📚 Publications

Below you can find a list of my publications, ordered chronologically.


Copy that! Editing Sequences by Copying Spans.
Sheena Panthaplackel, Miltiadis Allamanis, Marc Brockschmidt. AAAI 2021.
TLDR: Learn seq2seq models that can edit sequence by copying long spans.
Fast and Memory-Efficient Neural Code Completion.
A. Svyatkovskiy, S. Lee, A. Hadjitofi, M. Riechert, J. Franco, M. Allamanis. Mining Software Repositories 2021.
TLDR: Lightweight yet accurate code completion use neural reranking models.


CODIT: Code Editing with Tree-Based Neural Models.
S. Chakraborty, M. Allamanis, B. Ray. TSE 2020.
TLDR: Model code edits with tree-to-tree neural networks.
Flexeme: Untangling Commits using Lexical Flows.
Profir-Petru Pârțachi, Santanu Kumar Dash, Miltiadis Allamanis, Earl T. Barr. FSE 2020.
Typilus: Neural Type Hints.
M. Allamanis, E. T. Barr, S. Ducousso, Z. Gao. PLDI 2020.
TLDR: Use meta-learning to predict Python type annotations, including rare ones.


The Adverse Effects of Code Duplication in Machine Learning Models of Code.
M. Allamanis. SPLASH Onward! 2019.
TLDR: Automatically scraped code corpora commonly have many duplicates and evaluations are serverly affected by them.
CodeSearchNet Challenge: Evaluating the State of Semantic Code Search.
H. Husain, H. Wu, T. Gazit, M. Allamanis, M. Brockschmidt. 2019.
TLDR: A benchmark for natural language code search using human-provided annotations.
Generative Code Modeling with Graphs.
M. Brockschmidt, M. Allamanis, A. L. Gaunt, O. Polozov. ICLR 2019.
TLDR: Generate code expressions using asynchronous GNNs.
Learning units-of-measure from scientific code.
M. Danish, M. Allamanis, M. Brockschmidt, A. Rice, D. Orchard. SE 4 Science Workshop 2019.
Learning to Represent Edits.
P. Yin, G. Neubig, M. Allamanis, M. Brockschmidt, A. L. Gaunt. ICLR 2019.
TLDR: How can we represent edits in neural networks?
A Neural Approach to Decompiled Identifier Renaming.
J. Lacomis, P. Yin, E.J. Schwartz, M. Allamanis, C. Le Goues, G. Neubig, B. Vasilescu. ASE 2019.
Program Synthesis and Semantic Parsing with Learned Code Idioms.
R. Shin, M. Allamanis, M. Brockschmidt, O. Polozov. NeurIPS 2019.
TLDR: Use code idioms to improve program synthesis and semantic parsing.
Structured Neural Summarization.
P. Fernandes, M. Allamanis, M. Brockschmidt. ICLR 2019.
TLDR: A graph-to-sequence model for improved summarization of code and text.


Constrained Graph Variational Autoencoders for Molecule Design.
Q. Liu, M. Allamanis, M. Brockschmidt, A. L. Gaunt. NIPS 2018.
TLDR: VAEs for graph encoding and generation.
Deep Learning Type Inference.
V. Hellendoorn, C. Bird, E. T. Barr, M. Allamanis. FSE 2018.
TLDR: Sequence-based model to predict type annotations.
Learning to Represent Programs with Graphs.
M. Allamanis, M. Brockscmidt, M. Khademi. ICLR 2018.
TLDR: Represent programs as graphs and use GNNs to find bugs.
Mining Semantic Loop Idioms from Big Code.
M. Allamanis, E. T. Barr, C. Bird, M. Marron, C. Sutton. IEEE Transactions in Software Engineering 2018.
RefiNym: Using Names to Refine Types.
S. Dash, M. Allamanis, E. T. Barr. FSE 2018.
TLDR: Automatically refine types, such as strings, respecting type constraints by using data flow and identifier names.
A Survey of Machine Learning for Big Code and Naturalness.
M. Allamanis, E. T. Barr, P. Devanbu, C. Sutton. ACM Computing Surveys 2018.


Autofolding for Source Code Summarization.
J. Fowkes, P. Chanthirasegaran, R. Ranca, M. Allamanis, M. Lapata, C. Sutton. IEEE Transactions on Software Engineering 2017.
Learning Natural Coding Conventions.
M. Allamanis. PhD Dissertation 2017.
Learning Continuous Semantic Representations of Symbolic Expressions.
M. Allamanis, P. Chanthirasegaran, P. Kohli, C. Sutton. ICML 2017.
TLDR: Can we learn models that distiguish syntax from semantics of a math expression?
SmartPaste: Learning to Adapt Source Code.
M. Allamanis, M. Brockscmidt. 2017.
TLDR: Learn to adapt a snippet into new contexts.


A Convolutional Attention Network for Extreme Summarization of Source Code.
M. Allamanis, H. Peng, C. Sutton. ICML 2016.
TLDR: A 1D-CNN-to-sequence model to summarize source code.


A Bimodal Modelling of Source Code and Natural Language.
M. Allamanis, D. Tarlow, A. D. Gordon, Y. Wei. ICML 2015.
Suggesting Accurate Method and Class Names.
M. Allamanis, E. T. Barr, C. Bird, C. Sutton. FSE 2015.


Learning Natural Coding Conventions.
M. Allamanis, E. T. Barr, C. Bird, C. Sutton. FSE 2014.
TLDR: Coding conventions can be learned and suggested to developers.
Mining Idioms from Source Code.
M. Allamanis, C. Sutton. FSE 2014.
TLDR: Mine interesting syntactic patterns in code.


Mining Source Code Repositories at Massive Scale Using Language Modeling .
M. Allamanis, C. Sutton. MSR 2013.
Why, When, and What: Analyzing Stack Overflow Questions by Topic, Type, and Code.
M. Allamanis, C. Sutton. MSR 2013.


Evolution of a Location-based Online Social Network: Analysis and Models.
M. Allamanis, S. Scellato, C. Mascolo. IMC 2012.