The Proposed Model Framework for Generative AI proposes nine dimensions to create a trusted environment – one that enables end-users to use Generative AI confidently and safely, while allowing space for cutting-edge innovation. It offers practical suggestions that apply as initial steps, expanding on existing Model Governance Framework that covers Traditional AI. It also aims to facilitate international conversations among policymakers, industry, and the research community, to enable trusted development globally.
The AI Verify Foundation and IMDA welcomes comments and views from the international community, which can be sent to [email protected]. This will support the finalisation of the Model AI Governance Framework for Generative AI in mid-2024.
The Catalogue of Large Language Model (LLM) Evaluations sets out a comprehensive taxonomy that organises different domains of LLM evaluations to provide organisations a holistic overview of the available tests today.
It seeks to contribute to global discussions on safety standards by recommending a minimum baseline set of safety evaluations that LLM developers should conduct prior to LLM release.
The AI Verify Foundation welcomes initial comments and feedback on this draft release, which can be sent to [email protected]. We are currently establishing a more convenient method to receive and incorporate feedback from the community, and will update in due course.
Singapore launches AI Verify Foundation to shape the future of international AI standards through collaboration.
This crosswalk will allow companies to use AI Verify to achieve the desired outcomes of both AI Verify testing framework and US NIST AI Risk Management Framework in promoting trustworthy and responsible AI.
The development of the crosswalk is an important step towards harmonisation of international AI governance frameworks to reduce industry’s cost to meet multiple requirements. The joint effort also signals Singapore’s and the US’ common goal of balancing AI innovation, maximising benefits of AI technology while mitigating technology risks.
We invite interested organisations to use AI Verify to meet the common outcomes of both frameworks and share your experience with us.
Singapore launches AI Verify Foundation to shape the future of international AI standards through collaboration.
Your organisation’s background – Could you briefly share your organisation’s background (e.g. sector, goods/services offered, customers), AI solution(s) that has/have been developed/used/deployed in your organisation, and what it is used for (e.g. product recommendation, improving operation efficiency)?
Your AI Verify use case – Could you share the AI model and use case that was tested with AI Verify? Which version of AI Verify did you use?
Your experience with AI Verify – Could you share your journey in using AI Verify? For example, preparation work for the testing, any challenges faced, and how were they overcome? How did you find the testing process? Did it take long to complete the testing?
Your key learnings and insights – Could you share key learnings and insights from the testing process? For example, 2 to 3 key learnings from the testing process? Any actions you have taken after using AI Verify?