Previously worked on problems in Graph Machine Learning and Reinforcement Learning, and continues to draw on that experience when designing structured attention mechanisms or analyzing inductive biases in large models.
Education
Master's student in Computer Science at the University of California, Santa Barbara; Bachelor's degree in Computer Science from Indian Institute of Technology (IIT) Palakkad.
Background
Research interests include Transformers and Large Language Models (LLMs), with a particular focus on model efficiency, routing mechanisms, and architectural innovations such as Mixture-of-Experts (MoE), adapter-based tuning, and speculative decoding. Broadly motivated by challenges at the intersection of language understanding, scaling laws, and structured reasoning in LLMs. Passionate about building scalable, interpretable learning systems that bridge theoretical foundations and real-world deployment.
Miscellany
Personal interests include programming, tech education, and revolutionary tech innovations.