Technical Lead of MERaLiON, Singapore's national multimodal AI initiative. Working on multimodal large language models, speech processing, and language technologies for Southeast Asia.
I am a researcher and technical lead at the Institute for Infocomm Research (I2R), A*STAR, Singapore. I lead the development of MERaLiON—Singapore's national multimodal large language model initiative—building AI systems that understand both speech and language with a focus on Southeast Asian contexts.
I received my Ph.D. in Computer Science from Johns Hopkins University, where I was part of the Center for Language and Speech Processing (CLSP), advised by Kevin Duh. Before JHU, I was a research engineer at the Baidu-I2R Research Centre in Singapore.
Leading Singapore's flagship multimodal LLM programme
Models from 0.6B to 10B parameters
Designed for regional languages and accents
Unifying speech, language, and audio understanding in joint architectures
Recognition, encoding, and emotion detection from raw audio signals
Training and adaptation for diverse languages and domains
Information retrieval and transfer across language barriers
Quality estimation and evaluation for multilingual systems
Language technology for underrepresented regions and accents