Welcome to Manav’s Homepage!

Hello! I’m currently a second-year graduate student pursuing a Master’s degree in Intelligent Information Systems (MIIS) at the Language Technologies Institute within the School of Computer Science at Carnegie Mellon University (expected graduation: August 2025).

Before CMU, I completed an Integrated Dual Degree (B.Tech+M.Tech) from the Department of Electrical Engineering at Indian Institute of Technology Kharagpur, with a specialization in Signal Processing and Machine Learning. I graduated with an Institute Order of Merit for academic excellence (9.18 CGPA).

I have a passion for AI research and have worked on various projects related to natural language processing, multimodal AI, recommender systems, conversational grounding, and LLM applications at prestigious institutions such as CMU, Inria Paris, Sony Research India, MIT IDSS, UCL, Triomics, and the University of Alberta. My recent work focuses on vision-language models for radiology report generation, code review automation, conversational grounding, and domain-specific pre-training techniques.

Currently, I am working on my capstone project, ChartBench, a multi-task synthetic chart understanding benchmark that evaluates multimodal large language models on incremental chart code editing. I also serve as a Teaching Assistant for the NLP course at CMU’s Language Technologies Institute.

From February 2026, I will be joining Apple as a Machine Learning Engineer in the Instructional Products team.

If you have any ideas that you would like to collaborate on, hit me up!

News

Research Interests

My research interests lie primarily in the following broad domains:

  • Multimodal AI & Vision-Language Models: Developing methods for radiology report generation, chart understanding, and cross-modal reasoning.
  • Large Language Models (LLMs): Code review automation, reinforcement learning from verifiable feedback, domain-specific continual pre-training.
  • Conversational AI: Conversational grounding, multi-turn dialogue understanding, multi-agent interaction.
  • Domain-Specific NLP: Legal document analysis, clinical NLP, procedural text understanding.
  • Representation Learning: Graph neural networks for recommendation systems, contrastive learning.

Current Projects

  • ChartBench: Multi-task synthetic chart understanding benchmark for evaluating multimodal LLMs on incremental chart code editing (CMU Capstone Project).
  • CRScore++: Reinforcement learning framework combining verifiable tool feedback and AI feedback for code review comment generation.
  • Inference Algorithms for LLMs: Research on beam search, self-refinement, and reranking methods for language models.