Chaohao Yang ☕️
Chaohao Yang

Graduate Student

About Me

I am an M.S.E. student in Computer Science at Johns Hopkins University. My research centers on understanding and improving the real-world representations within deep learning models — from how they encode inputs to how architecture, loss functions, and training shape the resulting representations.

My work spans NLP and vision: I have evaluated LLM information extraction capabilities (180 citations), improved Word2Vec with learnable distance weighting, developed a memory-efficient quantized optimizer for LLM training (ICML 2025 Workshop), and am currently investigating semantic representations in latent diffusion models. I also gained industry experience training Llama-3-70B translation models with RAG at Tencent.

I aspire to pursue a Ph.D. to deepen my expertise in representation learning and generative modeling.

Download CV
Interests
  • Deep Learning
  • (Multi-modal) Large Language Model
  • Vision Foundation Model
Education
  • M.S.E. in Computer Science

    Johns Hopkins University

  • Visiting Student

    University of California, Berkeley

  • BE in Computer Science and Engineering

    The Chinese University of Hong Kong, Shenzhen

Recent Papers
(2025). Q-Adam-mini: Memory-Efficient 8-bit Quantized Optimizer for Large Language Model Training. In ICML 2025 Workshop ES-FoMo-III.