AI API Intel.
Agent-native.
Live health, latency, costs, failover, and benchmarks for 52 AI providers.
Free JSON API built for agents who need to make smart routing decisions.
OpenAI
GPT-4o, o3, o1 — the benchmark everyone chases
Anthropic
Claude — safety-first reasoning and long context
Google Gemini
Gemini 2.0 Flash, 2.5 Pro — native multimodal
DeepSeek
DeepSeek V3, R1 — elite reasoning at fraction of cost
Mistral AI
Mistral Large, Codestral — European frontier models
xAI Grok
Grok-3 — real-time web access, strong reasoning
Cohere
Command R+, Embed v3 — enterprise RAG specialist
Perplexity AI
Sonar — LLM-native web search API for agents
Groq
LPU inference — fastest tokens per second on the market
Cerebras
Wafer-scale chip inference — 1,000+ tokens/sec
Together AI
Open-source model inference — Llama, Mixtral, FLUX
Fireworks AI
Fast open-model inference — FireFunction, Llama, Mixtral
OpenRouter
Unified API across 200+ models — route by price or speed
Hugging Face
Serverless inference API — 100k+ open models on demand
Stability AI
Stable Diffusion 3.5, SDXL — open-source image gen
Replicate
Run any open-source model via simple API
ElevenLabs
TTS, voice cloning, STT — most realistic speech AI
Deepgram
STT, TTS, audio intelligence — production-grade ASR
Pinecone
Managed vector database for semantic search + RAG
GitHub Copilot
AI pair programmer built into GitHub + VS Code
Tavily
AI-optimised web search built for LLM agents
Runway
Gen-3 Alpha — leading text/image to video generation
NVIDIA NIM
Optimised inference microservices — GPU-native, enterprise-grade
AWS Bedrock
Managed foundation models — Claude, Llama, Titan on AWS
Azure OpenAI
OpenAI models on Microsoft Azure — enterprise SLA
Agent-Native API
Intelligence for agent routing decisions
16 endpoints covering health, latency percentiles, cost estimation, failover chains, benchmarks, and agent contract registry. Free and x402 micropayment endpoints.
GET https://topnetworks.com/api/v1/health
{
"providers": {
"openai": { "status": "operational", "response_time_ms": 112 },
"anthropic": { "status": "operational", "response_time_ms": 98 },
"groq": { "status": "degraded", "response_time_ms": 843 }
},
"summary": { "operational": 16, "degraded": 1, "outage": 0 }
}