Research at Grammarly

Grammarly builds unmatched AI-enabled writing assistance that helps people across platforms and devices, and our large and growing customer base reflects our edge. Grammarly’s support goes far beyond writing mechanics, combining rule-based approaches to explore uncharted areas in advanced machine learning. Our newest features leverage the power of generative AI to solve real customer problems.
Photograph of the Grammarly Research team

Who We Are

Continuous innovation makes Grammarly a market leader.
Grammarly's research team
Our team comprises researchers, linguists, and other functional roles. With diverse backgrounds in natural language processing, machine learning, and computational linguistics, together we advance research that revolutionizes the way people digitally communicate and uphold Grammarly’s vision of improving people’s lives through improved communication.
Our Applied Research team has been engaging in research for more than ten years. Our work directly impacts Grammarly’s products, which help millions of people worldwide. We further contribute to society and the research community by publishing, releasing open-source code, and sharing unique models and datasets.
Group of people at a research conference
If you are interested in joining our team

Recent Publications

CoEdIT: Text Editing by Task-Specific Instruction Tuning
Authors: Vipul Raheja, Dhruv Kumar, Ryan Koo, Dongyeop Kang
Conference: EMNLP 2023
Speakerly™: A Voice-based Writing Assistant for Text Composition
Authors: Dhruv Kumar*, Vipul Raheja*, Alice Kaiser-Schatzlein, Robyn Perry, Apurva Joshi, Justin Hugues-Nuger, Samuel Lou and Navid Chowdhury
Conference: EMNLP 2023
DeTexD: A Benchmark Dataset for Delicate Text Detection
Authors: Serhii Yavnyi, Oleksii Sliusarenko, Jade Razzaghi, Olena Nahorna, Yichen Mo, Knar Hovakimyan, Artem Chernodub
Conference: WOAH @ ACL 2023
Gender-Inclusive Grammatical Error Correction through Augmentation
Authors: Gunnar Lund, Kostiantyn Omelianchuk, Igor Samokhin
Conference: BEA 2023 @ ACL 2023
Privacy- And Utility-Preserving NLP with Anonymized data: A Case Study of Pseudonymization
Authors: Oleksandr Yermilov, Vipul Raheja, Artem Chernodub
Conference: TrustNLP @ ACL 2023
UA-GEC: Grammatical Error Correction and Fluency Corpus for the Ukrainian Language
Authors: Oleksiy Syvokon, Olena Nahorna, Pavlo Kuchmiichuk, Nastasiia Osidach
Conference: UNLP @ EACL 2023
Writing Assistants Should Model Social Factors of Language
Authors: Vivek Kulkarni, Vipul Raheja
Conference: In2Writing @ CHI 2023
Improving Iterative Text Revision by Learning Where to Edit from other Revision Tasks
Authors: Zae Myung Kim, Wanyu Du, Vipul Raheja, Dhruv Kumar, Dongyeop Kang
Conference: EMNLP 2022
Read, Revise, Repeat: A System Demonstration for Human-in-the-loop Iterative Text Revision
Authors: Wanyu Du, Zae Myung Kim, Vipul Raheja, Dhruv Kumar, Dongyeop Kang
Conference: In2Writing @ ACL 2022
Understanding Iterative Revision from Human-Written Text
Authors: Wanyu Du, Vipul Raheja, Dhruv Kumar, Zae Myung Kim, Melissa Lopez, Dongyeop Kang
Conference: ACL 2022
Ensembling and Knowledge Distilling of Large Sequence Taggers for Grammatical Error Correction
Authors: Maks Tarnavskyi, Artem Chernodub, Kostiantyn Omelianchuk
Conference: ACL 2022

Open Source and Datasets

IteraTeR (2022), R3 (2022), DELIteraTeR (2022): Iterative Text Revision (IteraTeR) from Human-Written Text.

Pseudonymization (2023): This work investigates the effectiveness of different pseudonymization techniques, ranging from rule-based substitutions to using pre-trained Large Language Models (LLMs), on a variety of datasets and models used for two widely used NLP tasks: text classification and summarization.

Grammarly NLP researchers found evidence of bias in off-the-shelf grammatical error correction systems: they performed worse for sentences with the gender-neutral pronoun “they.” So, we developed a technique to create augmented datasets with many examples of the singular “they.” We hope others use it to make NLP systems more gender-inclusive.

GECToR: Grammatical Error Correction: Tag, Not Rewrite. The GitHub repository provides code for training and testing state-of-the-art models for grammatical error correction with the official PyTorch implementation of the paper titled “GECToR – Grammatical Error Correction: Tag, Not Rewrite”.

DeTexD (2023): A Benchmark Dataset for Delicate Text Detection.

Privacy-Preserving NLP (2023)

UA-GEC Corpus (2021): Grammatical Error Correction and Fluency Corpus for the Ukrainian Language

CoEdIT (2023): Text Editing by Task-Specific Instruction Tuning.

Internship Program

Illustration of a hand holding a bubble containing a person and a balanced scale.
Grammarly has had a successful PhD internship program for many years; see publications above. In 2023, we had five interns pursuing PhDs in machine learning and one in linguistics.

We are starting our 2024 summer internship program soon; please check your eligibility and apply here.

Look Inside

Our Engineering blog is a great place to get to know the work we do, such as the sensible solutions that hold things together and the imaginative trials and errors that unlock the potential of our products.

And it’s all written by our very own builders and makers, for builders and makers like you.

Let’s Stay in Touch!

Stay connected with Grammarly’s vibrant community—follow us on LinkedIn and sign up for our Engineering Digest for news and updates on online meetups, offline events, and more. We believe in the power of networking and look forward to sharing our journey with you!