Jose Javier Gonzalez Ortiz

  • PhD Student, MIT

I am a Research Scientist at DataBricks Mosaic Research, where I work in the NLP pretraining team. I was part of the core team behind DBRX, a frontier LLM model developped by DataBricks. I am currently working on improving Large Language Model (LLM) pretraining efficiency and fine-tuning capabilities, and I am interested in building multi modal and adaptable foundational models.
I got my PhD in Computer Science at MIT CSAIL working with John Guttag and Adrian Dalca. During my PhD I worked in efficient deep learning methods, and for my thesis I focused on learning reconfigurable vision models, where multiple tasks are learned jointly in an amortized way. We achieved this using techniques such as hypernetworks and in-context learning. During my PhD I also worked on asynchronous large-scale distributed training and neural network pruning methods.
While at MIT, I co-organized Missing Semester, a lecture series on programmer tools often left untaught in modern CS curriculums. During my PhD, I interned at Facebook AI Research and Microsoft Research. Previously, I received a BSc in Telematics Engineering from ICAI in Spain, and interned at CERN Openlab.

Deep Learning Libraries

Publications

Projects

Awards & Honors