Hey there! I'm currently a research intern at CMU LTI working with Graham Neubig. I recently graduated from UC Berkeley 🐻, where I was very fortunate to be advised by Alane Suhr as part of the Berkeley NLP Group. Before that, I spent a great summer interning with the AI Technology Group at the MIT Lincoln Lab.

Although the specific research directions I pursue may change over time, I'm consistently drawn to ideas that develop a deep understanding of the underlying mechanisms, capabilities, and limitations of language models. This approach can manifest in various forms, including interpretability, carefully designed evaluations, studies of generalization, linguistic analysis, and ablations on data impact.

In addition to research, I spent three fun semesters teaching the data structures & algos class (CS 61B) at Berkeley. In my free time, I enjoy being outdoors, playing video games with my friends, and sending funny cat videos to Angela.

Preprints & Publications

Beyond English: Exploring Long Chain-of-Thought Reasoning Across Languages
Josh Barua, Seun Eisape, Kayo Yin, Alane Suhr
Under Review
paper coming soon

Using Language Models to Disambiguate Lexical Choices in Translation
Josh Barua, Sanjay Subramanian, Kayo Yin, Alane Suhr
EMNLP 2024
paper| code

Improving Medical Visual Representations via Radiology Report Generation
Keegan Quigley, Miriam Cha, Josh Barua, Geeticka Chauhan, Seth Berkowitz, Steven Horng, Polina Golland
ArXiv
paper| cite