Skip to content

Navigation Menu

Sign in
Appearance settings

Search code, repositories, users, issues, pull requests...

Provide feedback

We read every piece of feedback, and take your input very seriously.

Saved searches

Use saved searches to filter your results more quickly

Appearance settings

missingstudio/gateway

Open more actions menu

Gateway

Core infrastructure stack for building production-ready AI Applications

License Discord Twitter



Introduction

🌈 A Robust cloud-native AI Gateway - core infrastructure stack for building production-ready AI Applications . A Universal API for inferencing 100+ LLMs(OpenAI, Azure, Cohere, Anthropic, HuggingFace, Replicate, Stable Diffusion).

Supported Providers

Provider Provider Name Support Supported Endpoints
OpenAI openai /chat/completions, /chat/completions:stream
Groq groq /chat/completions, /chat/completions:stream
Anyscale anyscale /chat/completions
Deepinfra deepinfra /chat/completions
Together AI togetherai /chat/completions

Not supported (yet): images, audio, files, fine-tunes, moderations

Run locally

Missing studio can be deployed in a variety of ways. It is deployable on bare metal, or in dockerized environments.

Prerequisites

  • Buf CLI is required for building proto.

To start missing studio server, simply run the following command:

make compose-dev

Your AI Gateway is now running on http://localhost:8080 💥

Send a request using curl

curl \
--header "Content-Type: application/json" \
--header "x-ms-provider: openai" \
--header "Authorization: Bearer {{OPENAI_API_KEY}}" \
--data '{"model":"gpt-3.5-turbo","messages":[{"role":"user","content":"who are you?"}]}' \
http://localhost:8080/v1/chat/completions

Send a request using grpcurl

grpcurl \
-d '{"model":"gpt-3.5-turbo","messages":[{"role":"user","content":"hi"}]}' \
-H 'x-ms-provider: openai' \
-H 'Authorization: Bearer {{OPENAI_API_KEY}}' \
-plaintext  localhost:8080  llm.v1.LLMService.ChatCompletions

🚀 Features

  • 🌐 Universal API - Call every LLM API like it's OpenAI
  • AI Gateway for Security, Reliability and Observability
    • Load balancing across multiple provider and models
    • Atomatic Retries with exponential backoff
    • Rate limiting
    • Caching
    • Fallbacks
    • Monitoring
    • Alerting
    • Analytics
  • AI Studio
  • AI Agents
  • AI Workflow builder
  • OSS AI Models Inferancing
  • Serving model api at Scale on Kubernetes 🦄️
  • Building dev tools (CLI, SDK, API Client)

🫶 Contributions

AI studio is an open-source project, and contributions are welcome. If you want to contribute, you can create new features, fix bugs, or improve the infrastructure.

It's still very early days for this so your mileage will vary here and lots of things will break. But almost any contribution will be beneficial at this point. Check the current Issues to see where you can jump in!

If you've got an improvement, just send in a pull request!

  1. Fork it
  2. Create your feature branch (git checkout -b my-new-feature)
  3. Commit your changes (git commit -am 'feat(module): add some feature')
  4. Push to the branch (git push origin my-new-feature)
  5. Create new Pull Request

If you've got feature ideas, simply open a new issues!

Please refer to the CONTRIBUTING.md file in the repository for more information on how to contribute.

License

AI Studio is Apache 2.0 licensed.

Morty Proxy This is a proxified and sanitized view of the page, visit original site.