Nishant Balepur
Ph.D. Student in Computer Science at University of Maryland, College Park
Email:
nbalepur[at]umd[dot]edu
Hi! My name is Nishant and I’m a second-year Ph.D. student at the University of Maryland, where I am fortunate to be advised by Professors Jordan Boyd-Grayber and Rachel Rudinger. I also collaborate closely with Professor Shi Feng at GWU. I am graciously supported by the NSF GRFP and a Cohere For AI Research Grant.
I semi-jokingly say that I work on bullying (evaluating weaknesses) and babysitting (alignment) in LLMs. I am currently excited about the following three research questions:
- How can we better communicate factual knowledge? [topic mining (ACL’23), expository text (EMNLP’23), fact transfer (EMNLP’23), debates]
- How can we align models that actually help users? [flashcards (EMNLP’24), mnemonics (EMNLP’24)]
- How can we design evaluations to expose model/dataset weaknesses? [process of elimination (ACL’24), mcqa artifacts (ACL’24), benchmark cheating (ACL’24), mcqa plausibility (EMNLP’24), reverse qa]
If you’ve encountered another “Balepur, N” during your literature search, you may be looking for my sister 😛
📝 Selected Publications
2024
2023
🥳 Research Highlights
Sep 19, 2024 | New preprint on Reverse Question Answering: several strong LLMs struggle to generate accurate questions for answers as simple as numbers! what?! |
---|---|
Sep 19, 2024 | Three papers accepted to EMNLP on Aligning LLMs for Mnemonic Generation (main), NLP-Powered Flashcard Scheduling (main), and MCQA Quality Analysis (findings)! |
May 16, 2024 | Presenting three papers on LLM MCQA reasoning at ACL 2024! We see if LLMs can answer MCQA questions without the question (main); 2) perform process of elimination (findings); and 3) cheat on MCQA leaderboards (KnowLM workshop) |
May 3, 2024 | Presented our work Aritfacts or Abduction for MASC 2024 at Hopkins (1 of 5 selected orals). Also extremely grateful to be selected for 1 of 3 best paper awards! |
Apr 22, 2024 | Awarded a Cohere For AI Research Grant for our NLP+Education work with KAR³L. Excited for this collaboration! |