Home CommunitiesEventsSTACK MeetupsBuilding AI Responsibly Through Guardrails and Interpretability
STACK Meetup

Building AI Responsibly Through Guardrails and Interpretability

28 August 2025, 11:00 - 12:30 GMT+0
The Big Place @ GovTech HQ, #10-01, 10 Pasir Panjang Rd, Mapletree Business City, Singapore 117438

Overview

As AI systems grow more advanced, ensuring their safety and predictability becomes increasingly critical. This STACK Meetup explores how safety testing and guardrails, and mechanistic interpretability, can reduce misinformation and bias. These approaches work together to ensure that AI functions safely and as intended, especially in high-stakes settings.

Get tips from our GovTech's AI Practice team on safeguarding LLM applications against safety risks. Our speaker will guide you through the Responsible AI journey through the steps of defining a customised safety risk taxonomy, evaluating safety risks, and implementing safeguards to mitigate them.

Also, hear from a researcher at the Singapore AI Safety Institute on mechanistic interpretability, an approach akin to a brain scan for AI systems. This field seeks to uncover the inner workings of AI systems to identify backdoors, misalignment and unintended behaviours. This understanding powers applications such as model editing, behaviour steering, and the design of more robust guardrails, helping ensure that AI operates predictably and can be audited effectively.

Who should attend: AI Researchers/Engineers, Research Engineers, Data Scientists, Software Engineers/Developers and Designers who use AI in their products or solutions

Recommended knowledge level: Conceptual understanding of LLMs is helpful and experience building with LLMs is a bonus

Play video

Event details

Timing displayed in agenda is based on Singapore Timezone (GMT +8)

7:00 PM - 7:05 PM
Introduction by STACK Community
7:05 PM - 7:10 PM
Introduction on Lorong AI
Speaker(s)
Mr Mehul Shah
Mr Mehul Shah
Engagement and UX Lead, AI Practice, GovTech Singapore
7:10 PM - 7:45 PM
Safeguarding LLM Applications with Testing and Guardrails
Speaker(s)
Ms Goh Jia Yi
Ms Goh Jia Yi
AI Engineer (Responsible AI), AI Practice, GovTech Singapore
7:45 PM - 8:15 PM
Mechanistic Interpretability: Understanding Models From the Inside Out
Speaker(s)
Mr Clement Neo
Mr Clement Neo
Research Engineer, and Lab Advisor, Singapore AI Safety Institute; Apart Research
8:15 PM - 8:30 PM
Q&A

Meetup Resources

YouTube Playlist

Watch back on past STACK Meetups:

STACK Meetup YouTube Playlist
Flickr

Look back at moments captured during STACK Meetups:

STACK Meetup Flickr Photos
Telegram

Join our Telegram group for more on the latest happenings in STACK Community:

STACK Community's Telegram Group
Meetup.com

Join our Meetup.com group for latest info on STACK Meetups:

STACK Community's Meetup.com Group