Skip to content

A sample of using Dapr as an abstraction over several AI services.

License

Notifications You must be signed in to change notification settings

philliphoff/dapr-ai-bindings

Folders and files

NameName
Last commit message
Last commit date

Latest commit

 

History

47 Commits
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 

Repository files navigation

Dapr AI Bindings

A sample using Dapr extensibility to provide an abstraction over multiple AI services.

Background

An application that intends to make use of AI services must make many decisions.

  • Which AI service to use?
  • Which language model and settings to use?
  • Which SDK to use, if not making invoking services directly?

Each of these decisions has the potential to tightly couple the application to a specific AI service, model, or SDK, which can make it difficult for developers to respond to changes in the AI landscape as the field quickly evolves. Given the interest in consuming AI services in applications combined with a rapidly expanding marketplace offering services, developers may fear becoming locked into a single service, model, API, or platform.

Dapr provides an abstraction layer for different types of fundamental services such as state, pub-sub, secrets, and configuration, called "building blocks", that provide applications a high degree of flexibility in which (and how) services are used. While not an official building block today, Dapr can still be used to provide a similar abstraction layer for AI services.

Architecture

Dapr applications communicate with external services through a set of configured components, accessible via the Dapr sidecar process/container via HTTP or gRPC, with components of the same category conforming to the same interface. In this way, applications are shielded from the particulars of any individual component and can easily switch components with only configuration changes and not application (i.e. source) changes.

While the Dapr runtime provides a large selection of built-in components for various services, developers can create "pluggable" components (using Dapr 1.9 onward) to access services not yet part of the runtime (or might be internal to the enterprise). Pluggable components can be created (today) for the state store, pub-sub, and input/output binding building blocks.

This sample uses a set of output binding pluggable components to create an abstraction layer atop several different AI services: Open AI, Azure Open AI, and other Azure AI offerings such as Cognitive Services. Because each component implements the same set of operations, using the same request/response data, the application can switch between them no code changes. Configuration files are then used to allow components the ability to connect with their respective services and control how the components function.

These AI components can be used by applications directly, but there are common patterns of use that lend themselves to higher-level orchestration. For example, "chat" based language models typically require clients pass the entire history of a user's session for each completion, in order to provide the illusion of a continuous conversation. This implies a need to accumulate and manage session data, which would become a constant chore across applications. Conveniently, Dapr has a state store building block that can be used for that purpose. This is used by an "AI engine" component, another output binding itself, to expose an API that represents an AI service that maintains state. The engine uses its component configuration to determine which underlying AI service and state store is used.

The application itself is a basic web API with two endpoints, /complete and /summarize, that act as a proxies for requests to the AI binding components to enable easy testing. Those endpoints support an optional component query parameter to demonstrate the application switching between components with no code changes.

Architecture Diagram

Prerequisites

  • Dapr 1.10 or later
  • .NET 7 or later
  • Linux, MacOS, or Windows using WSL
  • A running Redis DB (such as the default Dapr state store container)

To use the Chat GPT binding, you must have an Open AI account and an API key.

To use the Azure AI binding, you must have an Azure Cognitive Services instance and its access key.

To use the Azure Open AI binding, you must have an Azure Open AI instance and its access key. This sample also expects that you've deployed two language models to the service, one named davinci and the other named gpt-35. (The exact models used for those deployments is not as important, but you could use text-davinci-003 and gpt-35-turbo, respectively.)

To use the .http files to send requests, install the REST Client VS Code extension.

Running in Docker Compose

The docker-compose.yaml in the repo root will deploy the application, Dapr sidecar, and the AI components sidecar at the same time, which is useful to get everything running and debugging is not a concern. The application will then be exposed locally on port 5050.

  1. Configure the secrets used by the components:

    Create a JSON file called secrets.json in ./src/DaprAI.Application/components/compose to hold the access keys. It should look something like:

    {
       "azure-ai-endpoint": "https://<name>.cognitiveservices.azure.com",
       "azure-ai-key": "<Azure Cognitive Services key>",
    
       "azure-open-ai-endpoint": "https://<name>.openai.azure.com"
       "azure-open-ai-key": "<Azure Open AI API key>"
    
       "open-api-endpoint": "https://api.openai.com"
       "open-api-key": "<Open AI API key>",
    
       "ai-store-host": "<redis host>",
       "ai-store-password": "<redis password>"
    }

    Note that the Redis configuration will most likely be redis with an empty password.

  2. Start the composition:

    docker compose up

    This should build the container images for the application and components, and start them alongside the Dapr sidecar. At this point, the application should be available at http://localhost:5050.

Running Locally

  1. Configure the Dapr components.

    Create a JSON file called secrets.json in the repo root folder to hold the access keys. It should look something like:

    {
       "azure-ai-endpoint": "https://<name>.cognitiveservices.azure.com",
       "azure-ai-key": "<Azure Cognitive Services key>",
    
       "azure-open-ai-endpoint": "https://<name>.openai.azure.com"
       "azure-open-ai-key": "<Azure Open AI API key>"
    
       "open-api-endpoint": "https://api.openai.com"
       "open-api-key": "<Open AI API key>",
    
       "ai-store-host": "<redis host>",
       "ai-store-password": "<redis password>"
    }
  2. Build and run the pluggable components.

    cd src/DaprAI.PluggableComponents
    dotnet run
  3. Build and run the application.

    dapr run -f ./dapr.yaml
  4. Send a completion request:

    POST http://localhost:5111/complete HTTP/1.1
    content-type: application/json
    
    {
        "user": "How are you?"
    }

    See the response from Chat GPT:

    HTTP/1.1 200 OK
    Connection: close
    Content-Type: application/json; charset=utf-8
    Date: Mon, 06 Mar 2023 22:01:23 GMT
    Server: Kestrel
    Transfer-Encoding: chunked
    
    {
      "assistant": "\n\nI'm doing well, thanks for asking. How are you?"
    }
  5. Send a summarization request:

    POST http://localhost:5111/summarize HTTP/1.1
    content-type: application/json
    
    {
        "text": "Jupiter is the fifth planet from the Sun and the largest in the Solar System. It is a gas giant with a mass one-thousandth that of the Sun, but two-and-a-half times that of all the other planets in the Solar System combined. Jupiter is one of the brightest objects visible to the naked eye in the night sky, and has been known to ancient civilizations since before recorded history. It is named after the Roman god Jupiter.[19] When viewed from Earth, Jupiter can be bright enough for its reflected light to cast visible shadows,[20] and is on average the third-brightest natural object in the night sky after the Moon and Venus."
    }

    See the response from Chat GPT:

    HTTP/1.1 200 OK
    Connection: close
    Content-Type: application/json; charset=utf-8
    Date: Mon, 20 Mar 2023 21:39:05 GMT
    Server: Kestrel
    Transfer-Encoding: chunked
    
    {
    "summary": "Jupiter is the largest planet in the Solar System, and it is a gas giant. With a mass two-and-a-half times that of all the other planets in the Solar System combined, it is also the fifth planet from the Sun. Jupiter has been known to ancient civilizations since before recorded history, and it is"
    }

    By default, requests will be made to Open AI's Chat GPT. You can use the query parameter to specify which binding to use:

    • Open AI's Chat GPT (e.g. ?component=open-ai-gpt)
    • Open AI's Davinci (e.g. ?component=open-ai-davinci)
    • Azure Open AI Davinci (e.g. ?component=azure-open-ai-davinci)
    • Azure Open AI Chat GPT 3.5 (e.g. ?component=azure-open-ai-gpt-3)
    • Azure Open AI Chat GPT 4 (e.g. ?component=azure-open-ai-gpt-4)
    • Azure AI (e.g. ?component=azure-ai)

    Not all operations (i.e. prompt or summarize) are supported by all bindings.

Reference

AI Engine

The "AI engine" consists of an output binding supporting the following operations:

  • createChat: create a chat session
  • completeText: request a "completion" of (i.e a response to) a user's text prompt
  • getChat: request the details of an existing chat session
  • getChats: request the IDs of all chat sessions
  • summarizeText: request the summarization of a text document (provided inline or via an URL)
  • terminateChat: delete a chat session

Component configuration metadata:

Name Required Description
aiName Yes The name of the component to use as the underlying AI service
storeName Yes The name of the component to use as the underlying state store

Create Chat

The createChat operation enables the engine to track the prompts/responses of a chat session. The engine the provides the underlying AI service with that history on each subsequent request for a text completion.

Creation expects a request of the form:

{
   "instanceId": "<A unique ID for the session>",
   "system": "<If provided, instructions for how the assistant should respond to user prompts>"
}

Text Completion

The completeText operation continues a chat session with the next user prompt.

Completion expects a request of the form:

{
   "instanceId": "<If provided, the ID of the session>",
   "user": "The user's prompt (e.g. 'How are you?')>"
}

Note that, if the instanceId is omitted, completions will be provided with no historical context (i.e. it is a straight pass through to the underlying AI service).

Text Summarization

The summarizeText operation expect a request of the form:

{
   "text": "<If provided, the actual text to summarize>",
   "url": "<If provided, a URL for the text document to summarize>"
}

The engine makes the request via the underlying AI service and returns a response of the form:

{
   "summary": "<The AI summarization of the text>"
}

Note that summarization is performed with no historical context (i.e. it is a straight pass through to the underlying AI service).

Terminate Chat

The terminateChat operation enables the engine to remove the history of a chat session.

Deletion expects a request of the form:

{
   "instanceId": "<The ID of the session>"
}

AI Building Block

The "AI building block" consists of an output binding supporting the following operations:

  • completeText: request a "completion" of (i.e a response to) a user's text prompt
  • summarizeText: request the summarization of a text document (provided inline or via an URL)

Note that components are not required to support all defined operations.

Text Completion

Components supporting the completeText operation expect a request of the form:

{
   "history":
   {
      "items":
      [
         { "role": "system", message: "You are a helpful AI assistant." }
      ]
   },
   "user": "<The user's prompt (e.g. 'How are you?')>"
}

Note that some language models may ignore past chat history, even when provided.

The components make the request of their respective service and are then expected to return a response of the form:

{
   "assistant": "<The AI assistant response to the user's prompt>"
}

Text Summarization

Components supporting the summarizeText operation expect a request of the form:

{
   "text": "<If provided, the actual text to summarize>",
   "url": "<If provided, a URL for the text document to summarize>"
}

The components make the request of their respective service and are then expected to return a response of the form:

{
   "summary": "<The AI summarization of the text>"
}

AI Building Block Components

Open AI

The Open AI component (bindings.open-ai) uses the Open AI service to implement text completion and summarization. The component can be configured to use both "chat" and non-chat language models.

Component configuration metadata:

Name Required Description
endpoint Yes The endpoint for the AI service, typically https://api.openai.com
key Yes The API key used to access the AI service
model Yes The name of the language model to use (e.g. text-davinci-003)
maxTokens No The maximum number of tokens to generate in the completion
summarizationInstructions No Instructions given to the AI service that describe how to summarize the text.
temperature No The sampling temperature to use, between 0 and 2. Higher values will generate more random completions while lower values will generate more deterministic completions.
topP No Consider the results with top_p probability mass, between 0 and 1.

Note: use either temperature or topP but not both.

Azure Open AI

The Azure Open AI component (bindings.azure-open-ai) uses an Azure Open AI instance to implement text completion and summarization. The component can be configured to use both "chat" and non-chat language models.

Component configuration metadata:

Name Required Description
endpoint Yes The endpoint for the AI instance, typically https://<instance>.openai.azure.com
key Yes The API key used to access the AI instance
deployment Yes The name of the deployed language model to use
maxTokens No The maximum number of tokens to generate in the completion
summarizationInstructions No Instructions given to the AI service that describe how to summarize the text.
temperature No The sampling temperature to use, between 0 and 2. Higher values will generate more random completions while lower values will generate more deterministic completions.
topP No Consider the results with top_p probability mass, between 0 and 1.

Note: use either temperature or topP but not both.

Azure AI

The Azure AI component (bindings.azure-ai) uses Azure Cognitive Services to implement text summarization.

Component configuration metadata:

Name Required Description
azure-ai-endpoint Yes The endpoint for the Cognitive Services instance, typically https://<instance>.cognitiveservices.azure.com
azure-ai-key Yes The API key used to access the instance

Note that this component does not support text completion.

About

A sample of using Dapr as an abstraction over several AI services.

Resources

License

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published