Dell Technologies is a leading American multinational technology company that develops, manufactures, sells, and supports computers and related products and services. The company helps to shape the future of digital transformation by enabling businesses and consumers worldwide to innovate, grow, and succeed.
With the rise in global AI adoption, particularly in high-risk sectors and use cases, building trust is critical to ensure that both consumers and organisations benefit from adopting AI.
How can we evaluate the safety and reliability of generative AI applications in real-world settings? In this post, we share a case study of third-party testing conducted on Singapore Airlines’ retrieval-augmented generation (RAG) search assistant—conducted in partnership with IMDA and Resaro. The post sets out a structured, risk-informed testing methodology, offering practical insights for developers, deployers, and policymakers working to raise the bar for AI safety.
Your organisation’s background – Could you briefly share your organisation’s background (e.g. sector, goods/services offered, customers), AI solution(s) that has/have been developed/used/deployed in your organisation, and what it is used for (e.g. product recommendation, improving operation efficiency)?
Your AI Verify use case – Could you share the AI model and use case that was tested with AI Verify? Which version of AI Verify did you use?
Your experience with AI Verify – Could you share your journey in using AI Verify? For example, preparation work for the testing, any challenges faced, and how were they overcome? How did you find the testing process? Did it take long to complete the testing?
Your key learnings and insights – Could you share key learnings and insights from the testing process? For example, 2 to 3 key learnings from the testing process? Any actions you have taken after using AI Verify?