Journal
- A Survey of Evaluation Metrics Used for NLG Systems - ACM Computing Surveys (2022)
Ananya B. Sai, Akash Kumar Mohankumar, Mitesh M. Khapra (link)
- Improving Dialog Evaluation with a Multi-reference Adversarial Dataset and Large Scale Pretraining TACL (2020)
Ananya B. Sai* , Akash Kumar Mohankumar* , Siddhartha Arora , Mitesh M. Khapra ( ∗ denotes equal contribution) (link)
Conference
- IndicMT Eval: A Dataset to Meta-Evaluate Machine Translation Metrics for Indian Languages - ACL (2023)
Ananya B. Sai, Tanay Dixit, Vignesh Nagarajan, Anoop Kunchukuttan, Pratyush Kumar, Mitesh Khapra, Raj Dabre (link)
- Bi-Phone: Modeling Inter Language Phonetic Influences in Text - ACL (2023)
Abhirut Gupta, Ananya B. Sai, Richard Sproat, Aravindan Raghuveer (link)
- Perturbation CheckLists for Evaluating NLG Evaluation Metrics - EMNLP (2021)
Ananya B. Sai, Tanay Dixit, Dev Yashpal Sheth, Sreyas Mohan, Mitesh M. Khapra (link)
- Re-evaluating ADEM: A Deeper Look at Scoring Dialogue Responses AAAI (2019)
Ananya B. Sai, Mithun Das Gupta, Mitesh M. Khapra, Mukundhan Srinivasan (link ) - ElimiNet: A Model for Eliminating Options for Reading Comprehension with Multiple Choice Questions IJCAI (2018)
Soham Parikh ∗ Ananya B. Sai ∗ , Preksha Nema ∗ , Mitesh M. Khapra ( ∗ denotes equal contribution) (link)
Workshop
- Closing the Gap in the Trade-off between Fair Representations and Accuracy in AAAI-DAI (2024)
Biswajit Rout, Ananya B. Sai, Arun Rajkumar
Preprints
- NL-Augmenter: A Framework for Task-Sensitive Natural Language Augmentation
K. D. Dhole, V. Gangal, S. Gehrmann, Ananya B. Sai, and et al. (link) - Frustratingly Poor Performance of Reading Comprehension Models on Non-adversarial Examples
S. Parikh, Ananya B. Sai, P. Nema, and M. M. Khapra (link)