Skip to main content

My Responsible AI Cookbook

· 6 min read
Nitya Narasimhan

Welcome to the kickoff of my latest project - a Responsible AI Cookbook. This is a personal learning project to explore the increasingly critical area of responsible AI - from core concepts and principles, to developer tools and workflows - with specific focus on the generative AI applications ecosystem. In this first post, I'll set the stage by defining a roadmap for what I want to learn, and explaining why I want to use the #30Days model for building a cookbook for developers.

What This Post Covers:

  • What is Responsible AI and why learn about it?
  • What is my #30DaysOf Learning Roadmap?
  • What does the Cookbook format provide?
  • How does this relate to Generative AI?
  • Resource Collection: Responsible AI For Developers

Deploy with Responsible AI


1. Background

As an AI Advocate at Microsoft, I've spent the last six months exploring different aspects of the Generative AI landscape - from Prompt Engineering and Fine Tuning fundamentals to Code-first Development using Azure AI Studio. But it wasn't till I started working on a Responsible AI Hub For Developers that I realized how responsible AI became critical to every step of the end-to-end development workflow (LLM Ops) for generative AI solutions.

And while there was a lot of information about responsible AI, it was scattered across different locations - from open-source toolboxes for data scientists, to cloud-based integrations for operationalization, and community-based forums for discussing real-world case studies and best practices. I needed a way to explore this in a structured manner, and build myself a "cookbook" of useful recipes that I could revisit to find the right tool or process for a given task.

2. What is Responsible AI?

Responsible AI is a fundamental approach to designing, developing, assessing, and deploying, AI solutions in a safe, trustworthy and ethical manner. Microsoft has an organization-wide Responsible AI standard and playbook anchored by six principles:

  1. Fairness - AI systems should treat all people fairly.
  2. Reliability and Safety - AI systems should perform reliably and safely (even under unplanned conditions).
  3. Privacy and Security - AI systems should be secure and respect privacy (think system security and user data privacy)
  4. Inclusiveness - AI systems should empower everyone and engage people (think about accessible experiences)
  5. Transparency - AI systems should be understandable. (think about interpretability and explainability for decisions)
  6. Accountability - People should be accountable for AI systems (think about oversight and guardrails for ensuring compliance)

Principles of Responsible AI

My perspective here is that we need to shift left in the application lifecycle when it comes to detecting and mitigating or eliminating potential harms that can adversely impact responsible AI usage. And that means raising developer awareness of these issues and documenting actionable recipes that can be integrated into their design thinking and developer workflows.

3. My #30Days Roadmap

My #30DaysOf platform and process started off as a way to create themed month content to give learners a deep-dive into a given topic in the span of four weeks. However, I want to repurpose the format for self-paced learning by keeping the same structure (roadmap) but without the time pressure for completion (30 days). The Cookbook format seemed to work best:

  1. The #30Days site becomes a developer cookbook for a given topic.
  2. The Roadmap is the table-of-contents, organized as top-level collections.
  3. The Collection is a set of recipes focused on a particular sub-topic.
  4. Each Recipe has an explainer (what is it) and an exercise (how do I use it).

My immediate goal is to build this cookbook for my own use - all mistakes and misconceptions are mine and the content is intended for learning purposes only and not for production use. With that in mind, my first steps are to find all relevant learning resources and organize them into a hierarchical roadmap (tree of nodes) with the following top-level nodes:

  • Core Concepts - understand key terminology and building blocks
  • Developer Tools - explore key tools and workflows for developers
  • Generative AI - understand Responsible AI usage for LLM Ops
  • Advanced Topics - covers research, best practices & new projects.

4. Generative AI and LLM Ops

Traditional AI applications focused on training models with domain-specific datasets, and deploying predictive AI endpoints that applications could use for decision-making. Modern generative AI applications focus on pre-trained "large language models" trained on massive internet-scale data, with endpoints that applications can use for generating original content (text, code, images, audio, video etc.).

This unlocks a whole new range of user experiences driven by natural language interactions (chat), driving a paradigm shift from MLOps to LLMOps. In my day job as an AI Advocate at Microsoft, I focus on how we can use the Azure AI platform to streamline the end-to-end development of Generative AI applications - and use this figure to explain what the LLM Ops (or end-to-end application lifecycle) would involve.

LLM Ops Workflow

When we think about Responsible AI in the context of this workflow, we have to plan a responsible generative AI solution with 4 steps:

  • Identify potential harms relevant to your solution
  • Evaluate presence of harms in generated outputs
  • Mitigate harms at multiple levels to minimize impact
  • Operate solution responsibly with an ops readiness plan.

With the Azure AI platform, this is captured in the approach to generative AI evaluations, the responsible AI guidelines for Azure AI services and the content filtering systemfor deployments.

Evaluation

The figure above gives us a high-level sense for where these mechanisms fit against the three stages in the LLM Ops diagram above. And just last week, the team announced new tools for building secure and trustworthy generative AI applications that enhance the workflow even more to tackle challenges like prompt injection attacks, vulnerability to jailbreaking, detection of "hallucinations" and more.

5. Next Steps

My next step is to build a roadmap (table of contents with links to relevant resources) to help me plan out my learning journey. Stay tuned!