To get deployed in minutes, use the Quickstart Wizard found at the bottom of the left-hand menu inside the Frosty AI platform. The wizard will guide you step-by-step through setting up your first workspace, connecting providers, and configuring routers.
1. Create a Workspace
Set up a collaborative environment to organize Routers and Providers. In the left-hand menu, select Workspaces and click Create Workspace.
2. Connect a Provider
Link an LLM provider like OpenAI or Anthropic to access models.
Learn how to get an API key from your provider: OpenAI, Anthropic, Mistral AI.
3. Set Up Your Router
Configure Routers to select the best LLM for your tasks. Frosty AI intelligently routes requests based on cost, performance, or custom-defined rules, allowing you to optimize workflows seamlessly. When you configure your route, you can set models for task-specific needs like cost, performance, and future-proofing by setting a failover provider.
4. Use Frosty in Your App or Workflow
Once your router is configured, you can start sending prompts using any of the following integration options:
✅ Option 1: Use the Python SDK
Install the SDK using pip:
pip install frosty-ai
Example Usage
# Import Frosty SDK
from frosty_ai import Frosty
def main():
router_id = "[YOUR_ROUTER_ID]"
router_key = "[YOUR_ROUTER_KEY]"
try:
# Create an instance of the Frosty class
frosty_sdk = Frosty(router_id, router_key)
# Make a text generation request
chat_result = frosty_sdk.chat([{
"role": "user",
"content": "Tell me a 10-word joke about the weather."
}])
print(f"Text generation result: {chat_result}")
except Exception as e:
print(f"An error occurred: {e}")
if __name__ == "__main__":
main()
✅ Option 2: Use the API Directly
You can call Frosty’s REST API using any language or platform. See API reference for request formats and authentication.
✅ Option 3: Use the Make Integration
Automate prompts and workflows visually with our Make integration. No code required—perfect for connecting Frosty to Google Sheets, Slack, Notion, and more.
✅ Option 4: Use the n8n Integration
Add Frosty to your n8n workflows using the n8n custom node for prompt routing, response logging, and more. Great for dev teams automating tasks with multiple providers.
5. Optimize & Validate
Test, compare, refine, and optimize your Router's performance in the Frosty AI Platform! Use built-in performance metrics, cost analysis, and logging to monitor and fine-tune your routing strategies.
You're all set! 🚀 Now that your router is up and running, you can explore more advanced configurations to fine-tune your setup.
Next, enhance your router by adding more providers, configuring models for specific tasks based on cost and performance, and ensuring reliability with a failover provider.
📖 To learn more, visit gofrosty.ai