I am a final-year PhD student in the Department of Industrial Engineering & Management Sciences at Northwestern University, where I am fortunate to work with Diego Klabjan. My research focuses on large-language models (post-training), model adaptation, and robust machine learning.
My aspiration is to develop AI systems with human-level adaptability and computational properties aligned with human preferences. Such AI systems would enable effective collaboration between humans and machines to tackle evolving real-world challenges that neither could solve alone.
Recently, I worked at Google as a student researcher, where I developed memory-efficient optimizers for (pre-/post-) training large-language models. Also, I interned at Autodesk Research, where I developed a multi-agent system for long context modeling.
Earlier in my career, I was a Senior Deep Learning Researcher at ESTsoft, where I worked on distribution shifts, contextual bandit, scalable variational inference, and model compression. I built the company's first ML model, which had been successfully operated in the Korean stock market.
Before that, I obtained my bachelor’s and master’s degrees at Hanyang University, where I worked on human-machine interactions. My research on formalizing human-machine interactions to enable adaptive automation under safety constraints was featured among the top 50 most popular articles in IEEE Transactions on Human-Machine Systems.
For any inquiries, please feel free to contact me by email: taejong.joo [at] northwestern.edu
I am actively seeking research scientist roles in industry and postdoctoral positions. With current work eligibility in the U.S. and Germany, I am open to relocation, including but not limited to Bay Area, NYC, Berlin, Zurich, London, Paris. I am driven to tackle pressing challenges with high impacts, and I am seeking a role in a leading research lab dedicated to pioneering foundational advances in AI. I thrive in technically and culturally diverse teams and am eager to contribute my skills and perspectives. Thank you for your consideration!
For the full list of my publications, visit my Google Scholar.
We develop Graph of Agents, a multi-agent system that expands the context window of large language models by orders of magnitude without any additional training. By framing long context modeling as a compression problem, GoA dynamically builds an optimal multi-agent collaboration structure tailored to the input. This principled approach eliminates the need for complex prompt engineering and specialized multi-agent system designs.
We demonstrate that the convenience of in-context learning, which enables adaptation to a new task by conditioning on demonstrations in an input prompt, comes with a hidden cost. While it matches the efficiency of a Bayes-optimal estimator in few-shot settings, we prove its statistical efficiency fundamentally diminishes over long contexts. This reveals a critical trade-off between ICL's flexibility and its long-term statistical efficiency.
We prove that selectively promoting temporal consistency for confident predictions significantly enhances self-training performance under distribution shifts. This approach prevents the common issue of model collapse—where performance deteriorates after a few epochs of self-training—resulting in improved performances with attractive robustness properties.
We introduce a new approach for simultaneously addressing model calibration and model selection in unsupervised domain adaptation: estimating the average accuracy across subpopulations. For efficient and accurate subpopulation accuracy estimation, we formulate the high-dimensional importance weight estimation problem into a more tractable coordinate-wise convex optimization problem.
We propose a scalable variational inference framework using a last-layer Dirichlet model as a new alternative to Bayesian neural networks. Our approach significantly enhances uncertainty representation ability of deterministic neural networks while preserving their strong generalization performances and efficiency unlike Monte Carlo dropout and deep ensembles.
Guided by first principles and the elegance of Occam’s razor, I believe simplicity often reveals the deepest insights and leads to effective and versatile solutions (with far fewer headaches).
Outside of work, I enjoy experimenting in the kitchen as a self-proclaimed master chef (enthusiastically endorsed by my wife), playing tennis, splashing paint on canvas, and traveling.
Fun fact: My Erdős Number = 3: Taejong Joo -> Diego Klabjan -> Craig Tovey -> Paul Erdős.