Get to know the Amazon research community at ACL 2019!
Amazon’s research teams are looking forward to meeting you at ACL 2019. Come and visit us at the Amazon booth, and read on for more information about academic collaboration, career opportunities, and our teams.
Publications and Talks at ACL
- Lluis Marquez
4th Workshop on Representation Learning for NLP (RepL4NLP-2019) - Isabelle Augenstein, Spandana Gella, Sebastian Ruder, Katharina Kann, Burcu Can, Johannes Welbl, Alexis Conneau, Xiang Ren and Marek Rei
NLP for Conversational AI - Tsung-Hsien Wen, Pei-Hao Su, Yun-Nung Chen, Dilek Hakkani-Tur, Mike Lewis, Minh-Thang Luong, Anuj Kumar, and Tania Bedrax-Weiss
- Multi-task Learning with Task, Group, and Universe Feature Learning
Shiva Pentyala, Mengwen Liu and Markus Dreyer
- A Large-Scale Corpus for Conversation Disentanglement
Jonathan K. Kummerfeld, Sai R. Gouravajhala, Joseph J. Peper, Vignesh Athreya, Chulaka Gunasekara, Jatin Ganhotra, Siva Sankalp Patel, Lazaros C Polymenakos and Walter Lasecki
- A Wind of Change: Detecting and Evaluating Lexical Semantic Change across Times and Domains
Dominik Schlechtweg, Anna Hätty, Marco Del Tredici and Sabine Schulte im Walde
- Choosing Transfer Languages for Cross-Lingual Learning
Yu-Hsiang Lin, Chian-Yu Chen, Jean Lee, Zirui Li, Yuyan Zhang, Mengzhou Xia, Shruti Rijhwani, Junxian He, Zhisong Zhang, Xuezhe Ma, Antonios Anastasopoulos, Patrick Littell and Graham Neubig
- Cross-lingual Knowledge Graph Alignment via Graph Matching Neural Network
Kun Xu, liwei wang, Mo Yu, Yansong Feng, Yan Song, Zhiguo Wang and Dong Yu
- Deep Neural Model Inspection and Comparison via Functional Neuron Pathways
James Fiacco, Samridhi Choudhary and Carolyn Rose
- Determining Relative Argument Specificity and Stance for Complex Argumentative Structures
Esin Durmus, Faisal Ladhak and Claire Cardie
- Disentangled Representation Learning for Non-Parallel Text Style Transfer
Vineet John, Lili Mou, Hareesh Bahuleyan and Olga Vechtomova
- Inducing Document Structure for Aspect-based Summarization
Lea Frermann and Alexandre Klementiev
- Joint Entity Extraction and Assertion Detection for Clinical Text
Parminder Bhatia, Busra Celikkaya and Mohammed Khalilia
- Learning Transferable Feature Representations Using Neural Networks
Himanshu Sharad Bhatt, Shourya Roy, Arun Rajkumar and Sriranjani Ramakrishnan
- Lightweight and Efficient Neural Natural Language Processing with Quaternion Networks
Yi Tay, Aston Zhang, Anh Tuan Luu, Jinfeng Rao, Shuai Zhang, Shuohang Wang, Jie Fu and Siu Cheung Hui
- Simple and Effective Curriculum Pointer-Generator Networks for Reading Comprehension over Long Narratives
Yi Tay, Shuohang Wang, Anh Tuan Luu, Jie Fu, Minh C. Phan, Xingdi Yuan, Jinfeng Rao, Siu Cheung Hui and Aston Zhang
- Multimodal Abstractive Summarization for How2 Videos
Shruti Palaskar, Jindřich Libovický, Spandana Gella and Florian Metze
- Poetry to Prose Conversion in Sanskrit as a Linearisation Task: A case for Low-Resource Languages
Amrith Krishna, Vishnu Sharma, Bishal Santra, Aishik Chakraborty, Pavankumar Satuluri and Pawan Goyal
- Ranking Generated Summaries by Correctness: An Interesting but Challenging Application for Natural Language Inference
Tobias Falke, Leonardo F. R. Ribeiro, Prasetya Ajie Utama, Ido Dagan and Iryna Gurevych
- Robust Zero-Shot Cross-Domain Slot Filling with Example Values
Darsh Shah, Raghav Gupta, Amir Fayazi and Dilek Hakkani-Tur
- Span-Level Model for Relation Extraction
Kalpit Dixit and Yaser Al-Onaizan
- Training Neural Machine Translation To Apply Terminology Constraints
Georgiana Dinu, Prashant Mathur, Marcello Federico and Yaser Al-Onaizan
- Topic Modeling with Wasserstein Autoencoders
Feng Nan, Ran Ding, Ramesh Nallapati and Bing Xiang
- Translating Translationese: A Two-Step Approach to Unsupervised Machine Translation
Nima Pourdamghani, Nada Aldarrab, Marjan Ghazvininejad, Kevin Knight and Jonathan May
- Why Didn’t You Listen to Me? Comparing User Control of Human-in-the-Loop Topic Models
Varun Kumar, Alison Smith-Renner, Leah Findlater, Kevin Seppi and Jordan Boyd-Graber
- Multimodal and Multi-view Models for Emotion Recognition
Gustavo Aguilar, Viktor Rozgic, Weiran Wang and Chao Wang
Internships for PhD Students
We offer 3-6 month internships year-round, with opportunities in Aachen, Atlanta, Austin, Bangalore, Barcelona, Berlin, Boston, Cambridge, Cupertino, Graz, Haifa, Herzliya, Manhattan Beach, New York, Palo Alto, Pasadena, Pittsburgh, San Francisco, Shanghai, Seattle, Sunnyvale, Tel Aviv, Tübingen, Turin, and Vancouver. To apply, email your resume to ACL2019@amazon.com, and let us know if there are any specific locations, teams, or research leaders that you are interested in working with.
Job Opportunities for Graduating Students and Experienced Researchers
We are looking for results-driven individuals who apply advanced computer vision and machine learning techniques, love to work with data, are deeply technical, and highly innovative. If you long for the opportunity to invent and build solutions to challenging problems that directly impact the way Amazon transforms the consumer experience, we are the place for you. To apply, email your resume to ACL2019@amazon.com and let us know if there are any specific locations, teams, or research leaders that you are interested in working with.
Amazon Scholars is a new program for academic leaders to work with Amazon in a flexible capacity, ranging from part-time to full-time research roles. Learn more at amazon.jobs/scholars.
BAIR Lab Opening
Amazon & the University of California Berkeley ARtificial Intelligence Research (BAIR) Lab Partnered to open the BAIR Open Research Commons, a new industrial affiliate program launched to accelerate cutting-edge AI research. The BAIR Commons is designed to streamline collaborative, cutting-edge research by students, faculty, and corporate research scholars.
Amazon and NSF Collaborate to Accelerate Fairness in AI Research
NSF and Amazon are partnering to jointly support computational research focused on fairness in AI, with the goal of contributing to trustworthy AI systems that are readily accepted and deployed to tackle grand challenges facing society. NSF has long supported transformative research in artificial intelligence (AI) and machine learning (ML). The resulting innovations offer new levels of economic opportunity and growth, safety and security, and health and wellness.
Check out the details here.
Amazon Web Services (AWS) Research Grants
In partnership with Machine Learning@Amazon, AWS offers up to $20,000 in compute tokens each quarter to professors and students. Academics have used these grants for projects ranging from Hack End weekends to massive MRI imaging projects. AWS provides building blocks for developing applications ranging from Elastic MapReduce for Hadoop analytics to fast and scalable storage with Amazon DynamoDB. Learn more & apply here.
Amazon Research Awards
ARA is an unrestricted gift to recognize exceptional faculty, and fund projects leading toward a PhD degree or conducted as a part of post-doctoral work. Each selected proposal is assigned an Amazon research contact, as we believe that both sides benefit from direct interaction on the topic of their research. We invite ARA recipients to visit Amazon offices worldwide to give talks related to their work and meet with our research groups face-to-face. We encourage ARA recipients to publish the outcome of the project and commit any related code to open source code repositories. Learn more here.
Applications will be open on September 10th, 2019 with a submission deadline of October 4, 2019.
- Computer vision
- Fairness in artificial intelligence
- Knowledge management and data quality
- Machine learning algorithms and theory
- Natural language processing
- Online advertising
- Operations research and optimization
- Search and information retrieval
- Security, privacy and abuse prevention
Publishing at Amazon
Amazon is committed to innovating at the frontiers of machine learning and artificial intelligence. Our scientists are encouraged to engage in the research community in the form of written publications, open source code and public datasets. We have instituted a new, fast-track publication approval process, to help share our research efforts as quickly as possible, while maintaining the highest standards of quality. Check out some of our most recent publications here.
Diversity at Amazon
We are a company of builders working on behalf of a global customer base. Diversity is core to our leadership principles, as we seek diverse perspectives so that we can be “Right, A Lot”. We welcome people from all backgrounds and perspectives to innovate with us. Learn more at amazon.com/diversity.
Questions about career opportunities or academic partnerships? Contact us at ACL2019@amazon.com.
Meet Amazonians working in NLP
Natural Language Processing at Amazon
"I spoke to the future and it listened" - Gizmodo
Meet the team of world-class scientists behind Alexa.
Graduate Research Careers
Learn more about graduate research careers for PhD and Masters students.
Amazon Lex - Quickly Build Conversational Interfaces
With Amazon Lex, the same deep learning technologies that power Amazon Alexa are now available to any developer.
Learn more about Amazon Polly
Amazon Polly is a service that turns text into lifelike speech, making it easy to add voice to your website, mobile app, or device.
Introducing the Alexa Prize
The Alexa Prize is an annual competition for university students dedicated to accelerating the field of conversational AI.
2017 Alexa Prize Finals
Learn more about the Alexa Prize: https://developer.amazon.com/alexaprize