§ Recipe

Production AI infrastructure.

A production AI stack is less about the flashiest model and more about control: routing, evals, fallback paths, open-model options, and cost visibility. This stack keeps you flexible while the model market keeps moving.

TeamSerious scaleOpen source
  1. Model benchmark desk

    01
    Artificial Analysis

    Artificial Analysis

    Independent AI model benchmarks for intelligence, speed, pricing, context, and modalities.

    Models & Infrastructure
    Freemium
    4.5

    Start with independent quality, latency, and pricing data before you standardize on a model. It keeps model choice tied to current evidence instead of vendor claims.

  2. Model routing

    02
    OpenRouter

    OpenRouter

    One API and routing layer for hundreds of AI models across many providers.

    LLM Providers & APIs
    Freemium
    4.4

    Put a router in front of your app so switching models is a config decision, not a rewrite. It gives you fallback paths, price comparison, and faster evaluation cycles.

  3. Open-model discovery

    03
    Hugging Face

    Hugging Face

    The central hub for AI models, datasets, Spaces, libraries, and open-source ML collaboration.

    Models & Infrastructure
    Freemium
    4.8

    Use the hub for model cards, datasets, Spaces, and community signal. It is where you find candidates before they enter your production path.

  4. Prototype model APIs

    04
    Replicate

    Replicate

    Run open and community AI models from a web playground or API.

    Models & Infrastructure
    Paid
    4.4

    Turn model demos into API calls quickly, especially for image, video, and audio experiments. It is the fastest way to learn whether a model belongs in the product.

  5. Serverless GPU jobs

    05
    Modal

    Modal

    Serverless AI infrastructure for running code, jobs, containers, and GPUs from Python.

    Models & Infrastructure
    Freemium
    4.5

    Run Python jobs, containers, scheduled work, and GPU endpoints without building cloud plumbing first. It is ideal for batch inference and AI backend work.

  6. Production inference

    06
    Baseten

    Baseten

    Production AI inference platform for deploying, optimizing, and scaling models.

    Models & Infrastructure
    Enterprise
    4.5

    When a model becomes a product dependency, Baseten handles serving, scaling, observability, and enterprise deployment patterns around that endpoint.