DownForAI

Anthropic status: API, auth, latency & outage reports

Operational
Last probe 10 min agoΒ·3 surfaces
Likely your sideMEDIUM confidence
πŸ“‘ Official status page β†’πŸ“š Docs β†’πŸ’³ Pricing β†’
100.0%
24h Uptime
330ms
p50 Latency
802ms
p95 Latency
0
Incidents (30d)

Having issues with Anthropic?

Report problems quickly and help the community stay informed.

0
reports in the last 24 hours

Surface Health

Claude APIOperational
HTTP 200p50 310ms9m ago
ClaudeOperational
HTTP 200p50 434ms21m ago
Claude CodeOperational
HTTP 200p50 246ms15m ago

Uptime β€” last 24h

100.0%

Latency β€” last 24h (p50 per 30 min slot)

Loading…

Is Anthropic down for everyone?

Likely local or client-side issue
Our probes see normal responses. The issue is likely on your end or in your network path.
Moderate confidence
Probe summary (3 surfaces)
All surfaces operational as of last probe.
Signals detected
  • All monitored surfaces operational
  • No recent user reports
  • Check your network, credentials, or rate limits

Incident history (30d)

βœ“No incidents recorded in the past 30 days.

Reported symptoms

No user reports for Anthropic in the last 24 hours.

Known error signatures

Common failure patterns and how to diagnose them

Provider details

The Anthropic API provides programmatic access to Claude models via a REST interface at https://api.anthropic.com. Also available through AWS Bedrock and Google Vertex AI as separate infrastructures.

What we monitor
Claude API β€” Direct API access
Claude Console / Workbench β€” Dev dashboard
AWS Bedrock β€” Claude models β€” Alternate infrastructure
Google Vertex AI β€” Claude models β€” Alternate infrastructure
Message Batches API β€” Async
Status page segments
Claude APIClaude Consoleclaude.aiClaude Code
Model families
Claude Opus 4.6, Claude Opus 4.1Claude Sonnet 4.6, Claude Sonnet 4.5Claude Haiku 4.5Claude Sonnet 3.7 (deprecated)Model IDs differ across Claude API, Bedrock, and Vertex β€” verify per-platform
Common limits & quotas
  • Rate limits per tier; API returns real-time remaining quota via anthropic-ratelimit-* headers
  • Prompt caching reduces input token costs significantly for repeated contexts
  • Message Batches API offers ~50% cost reduction for async workloads
Ecosystem dependencies
Claude Code depends on Anthropic API + authCursor, Continue.dev route to Claude models by default for many usersEnterprise deployments often prefer Bedrock/Vertex for SLA and compliance
Operator notes
  • The anthropic-version header is required on every request β€” missing it returns a 400 that juniors often misdiagnose as an endpoint outage
  • Monitor anthropic-ratelimit-* headers proactively β€” don't wait for 429s
  • Prompt caching changes effective rate limits dramatically; measure with caching enabled
  • For high-volume prod: provision Bedrock or Vertex as a fallback route; it's a different infrastructure, often unaffected by direct API incidents
Diagnostic signals
retry-afteranthropic-ratelimit-requests-limitanthropic-ratelimit-requests-remaininganthropic-ratelimit-requests-resetanthropic-ratelimit-input-tokens-limitanthropic-ratelimit-input-tokens-remaininganthropic-ratelimit-output-tokens-limitanthropic-ratelimit-output-tokens-remainingrequest-id$ curl https://api.anthropic.com/v1/models -H "x-api-key: $ANTHROPIC_API_KEY" -H "anthropic-version: 2023-06-01" β€” reachability + model list$ curl https://api.anthropic.com/v1/messages -H "x-api-key: $ANTHROPIC_API_KEY" -H "anthropic-version: 2023-06-01" -H "content-type: application/json" -d '{"model":"claude-sonnet-4-5","max_tokens":10,"messages":[{"role":"user","content":"ping"}]}' β€” real inference test$ Check status.anthropic.com for component-level breakdown (API vs Console vs claude.ai vs Claude Code)

Fallback alternatives

What to use if this service is down

Direct Anthropic API is degraded
AWS Bedrock-hosted Claude can reduce downtime if your org is on AWS
Low cost if already provisioned
Easy switch
Anthropic and Bedrock both down
Google Vertex AI-hosted Claude is a third infrastructure path
Moderate effort
All Anthropic paths unavailable
OpenAI API or Google Gemini API can reduce downtime for general chat
Low cost with abstraction layer
Easy switch
Long-context workloads specifically
Gemini 2.5 Pro handles very long contexts
Easy switch

How we monitor

downforai.com probes each AI service every 2–5 minutes from multiple independent locations. We measure HTTP response codes, latency (p50 & p95), and endpoint availability across the surfaces listed above. Status is classified as Operational, Degraded, or Outage based on a weighted combination of probe results. Uptime is calculated over 30-minute buckets and the last 24 hours. User reports are factored into our diagnosis as a secondary signal. We are independent of all providers listed and receive no compensation to report any particular status.

Community Discussion

0/1000
No comments yet. Be the first to share your experience!