I'm thrilled to be speaking at this year's TrustCon!
Alongside Jen Weedon and Sarah A., we'll be doing a mock red teaming exercise from end-to-end: from threat modeling, to chatbot testing, to framing up recommendations. We've designed two scenarios for participants to choose from: a "Virtual Therapist" chatbot, or "Ask the Historian", an AI assistant geared towards K-12 educational contexts. We'll be looking for different types of bias, safety, and factuality, and helping participants understand the high-level steps in thinking about how different types of users engage with these types of technologies, varied testing strategies to simulate real user behavior, and how to evaluate results. Join us if you are interested in learning how to red team AI models!
Additionally, I'll be supporting Sujata Mukherjee and Rachel Fagen for a Language Equity Roundtable. The internet, despite its global reach, remains largely an English-centric space. This digital divide excludes billions of users, hindering their access to information and online communities. This roundtable will explore the critical issue of language equity in AI-powered content moderation systems. We will delve into the technical challenges of developing AI models that effectively and fairly address harmful content across diverse languages, including low-resource languages. The discussion will focus on practical outcomes such as:
▶️ Language Equity Metrics: Developing and discussing measurable criteria for assessing the fairness and inclusivity of AI systems across different languages, drawing inspiration from existing benchmarks for English language models.
▶️ Data Diversity: Strategies for building robust and representative datasets for multilingual AI models, mitigating biases, and addressing data scarcity.
▶️ Technical Solutions: Exploring techniques like de-biasing algorithms, transfer learning, and explainable AI to enhance the fairness and accuracy of multilingual content moderation.
▶️ Policy Implications: Examining the role of policy in promoting the development and deployment of equitable AI-powered content moderation systems, including data governance, algorithmic accountability, and transparency requirements.
As a new Board member for the Integrity Institute, I am looking forward to a gathering of our members tomorrow! And, as a Strategic Advisor for All Tech Is Human, I'm really excited for a meet-up the following day with our members there.
A huge thanks to Trust & Safety Professional Association for their efforts in organizing this wonderful convening. Please let me know if you are in town - it would be great to meet you!