You are here:

/

/

Testing Real World Gen AI Systems

Testing Real World Gen AI Systems: Lessons from the Global AI Assurance Pilot

GenAI can have a massive, positive impact on our society and economy – if it is adopted at scale in public and private sector organisations. Using GenAI in such real-world situations, at scale, raises the quality and reliability bar significantly.

On May 30, we brought together almost 100 GenAI practitioners and ecosystem players together with our Global AI Assurance Pilot participants.

The focus?

Making AI reliable, trustable – some might even say boring and predictable – for adoption at scale.

We spotlighted three major initiatives:

Read all case studies here.
Read the main report from Pilot here

Here’s what we explored together during the interactive panel discussions:

  • What should we actually test in AI systems?
  • How do we work with insufficient test data?
  • Why “observability” across the app pipeline matters — especially in agentic workflows
  • How to scale automated testing with LLM evaluators with human feedback

This is how we build trusted AI, together — by testing the real-world, human-impacting cases, and sharing the processes and learnings with one another.

Related Events

No posts found!

Thank you for completing the form. Your submission was successful.

Preview all the questions

1

Your organisation’s background – Could you briefly share your organisation’s background (e.g. sector, goods/services offered, customers), AI solution(s) that has/have been developed/used/deployed in your organisation, and what it is used for (e.g. product recommendation, improving operation efficiency)?

2

Your AI Verify use case – Could you share the AI model and use case that was tested with AI Verify? Which version of AI Verify did you use?

3

Your reasons for using AI Verify – Why did your organisation decide to use AI Verify?

4

Your experience with AI Verify – Could you share your journey in using AI Verify? For example, preparation work for the testing, any challenges faced, and how were they overcome? How did you find the testing process? Did it take long to complete the testing?

5

Your key learnings and insights – Could you share key learnings and insights from the testing process? For example, 2 to 3 key learnings from the testing process? Any actions you have taken after using AI Verify?

6

Your thoughts on trustworthy AI – Why is demonstrating trustworthy AI important to your organisation and to any other organisations using AI systems? Would you recommend AI Verify? How does AI Verify help you demonstrate trustworthy AI?
Enter your name and email address below to download the Discussion Paper by Aicadium and IMDA.
Disclaimer: By proceeding, you agree that your information will be shared with the authors of the Discussion Paper.