Automodels

Improve your LLM calls by dynamically selecting the best model for a given task. Our APIs help you classify tasks and optimize prompts – giving you the tools you need to build robust AI applications and agents at scale.

Automodels is a suite of APIs that help you use the best LLM for the job and optimize prompts.

BEST MODEL SELECTION

Automodels detects the intent of your prompts and maps them to the best model for the given task, based on your preferences for cost and quality.

PROMPT OPTIMIZATION

Use our tools to restructure your prompts auto-magically to improve performance and quality of your LLM calls, tailored to the model you're using.

SIMPLE INTEGRATION

Our library of tools is easy to integrate into your existing workflows. Call our API directly, or start with our npm or pip packages:

from llm_automodels import AutoModels

client = AutoModels("your-api-key")
prompt = "Explain the key aspects of special relativity."

response = client.get_best_model(
    prompt,
    profile="cost"
)

print("Identified task type: ", response.task_type)
print("Best model: ", response.best_model)

reponse = client.optimize_prompt(
    prompt,
    model=reponse.best_model
)

print("Optimized prompt: ", response.optimized_prompt)

SUPPORTED MODELS

OpenAI Models
Google Gemini & Gemma Models
Anthropic Claude Models
Meta Llama Models
Mistral Models
Qwen Models
Deepseek Models
Microsoft Phi Models

PROMPT COMPRESSION

Our API can compress your prompts using advanced semantic analysis, reducing token usage by up to 70% while maintaining context and quality.

ADVANCED ANALYTICS

Get detailed insights into your token usage, compression rates, and optimization opportunities with our comprehensive analytics dashboard or via API.

EARLY ACCESS

Start optimizing your LLM calls

Use Automodels to improve the quality of your LLM calls, save on tokens, and reduce time spent on prompt engineering.

Join now
>50 users already joined

Priority Access

Shape the future of AI optimization

Dedicated Support

Direct engineering team access

Early Bird Community

Join our community of early adopters

What we're working on:

  • Adding new models to our auto-model selector
  • Prompt compression for context >2M tokens
  • Latency as a model selection parameter
  • and more in our next release, beta-0.1.2

What's new?

We recently released our public API documentation. Check it out now and to get started using our API! View docs →

Community

Discord Server Github Community Follow us on Twitter

FAQs

How can I get support? → Beta users can post a Github issue (private repo for beta users), and use our invite-only support channel in Discord.

Does Automodels have a on-prem version? → Join the waitlist

What does Automodels cost? → Our API is currently free to use.

Are you hiring? → Currently not hiring, but we're always looking for talented people to join our team. Get in touch

© Automodels 2025PrivacyTerms