April 2, 2025 - I was featured in an interview at AIHub for my presentation at AAAI2025 on aligning generative AI with technical standards. See post here.
February 20, 2025 - I'm excited to join the British Standards Institution (BSI) as a committee member and UK national expert for the AI working group, ART/1, mirroring CEN-CENELEC's Joint Task Force 21 towards developing standards for AI.
January 28, 2025 - Our frontier AI benchmark Humanity's Last Exam (HLE) established by The Center for AI Safety has been released! I contributed 5 hard mathematics and linguistics questions in this benchmark for evaluating advanced AI capabilities. Access HLE here.
January 22, 2025 - Our paper INCLUDE: Evaluating Multilingual Language Understanding with Regional Knowledge has been accepted as Spotlight at ICLR 2025! Access the paper here.
November 4, 2024 - I recently passed my PhD candidacy exam and my dissertation entitled Natural Language Generation with Expert Standards has been accepted at the AAAI Doctoral Consortium 2025 happening in Philadelphia, USA.
October 10, 2024 - I'm an invited speaker at Meta's Open Innovation AI Research Community (OIAIRC) Annual Research Workshop in London. I will present my research on integrating industry standards into LLMs.
September 20, 2024 - I'm happy to have 3 long papers (2 Main, 1 Findings) accepted at EMNLP 2024 covering works on standardized NLG, benchmarking LLMs with specialized dictionaries, and our SEACrowd Project.
July 23, 2024 - I'm happy to receive the Best Reviewer Award (Top ~2% of 7,437 reviewers) at ICML 2024 in Vienna, Austria.
May 8, 2024 - I'm happy to receive the Doctoral Recognition Award 2024 for my research in NLP by the University of Bath Doctoral College.
May 2, 2024 - Our new position paper Near to Mid-term Risks and Opportunities of Open Source Generative AI has been accepted as Oral (top 1.5% of submissions) for ICML 2024. Work led by University of Oxford and supported by Meta. Paper can be found here.
April 17, 2024 - Our new AI Safety Benchmark, containing 40k+ prompts for evaluating GenAI models for harms and hazard categories, is now published at IEEE Spectrum and on Arxiv. Work led by the MLCommons AI Safety Group.