1. Install & auth
Create an account in the dashboard. You’ll receive an API key shown once — store it securely. All API calls use the X-API-Key header.
shellbash
export SC_API_KEY="sc_live_..."
export SC_BASE="https://api.seattlecompute.com/v1"2. Deploy a model
Create a deployment by specifying a model id, priority, and projected hours per day. The optimizer picks the GPU and engine for you.
deploy.shbash
curl -X POST "$SC_BASE/deployments" \
-H "X-API-Key: $SC_API_KEY" \
-H "Content-Type: application/json" \
-d '{
"model_source": "huggingface",
"model_id": "meta-llama/Llama-3.1-8B-Instruct",
"priority": "balanced",
"hours_per_day": 24,
"precision": "fp16"
}'3. Call your endpoint
Every deployment exposes an OpenAI-compatible chat completion endpoint.
inference.shbash
curl "$ENDPOINT_URL/v1/chat/completions" \
-H "Authorization: Bearer $SC_API_KEY" \
-H "Content-Type: application/json" \
-d '{
"model": "llama-3.1-8b-instruct",
"messages": [{"role": "user", "content": "Hello from Seattle."}]
}'4. List deployments
list.shbash
curl "$SC_BASE/deployments" -H "X-API-Key: $SC_API_KEY"5. Custom domains
Register your domain, point the CNAME we give you, and attach to any deployment. TLS is provisioned automatically.
domain.shbash
curl -X POST "$SC_BASE/domains" \
-H "X-API-Key: $SC_API_KEY" \
-H "Content-Type: application/json" \
-d '{"domain": "llm.yourcompany.com"}'6. Next steps
- See the full GPU tier table
- Open the dashboard to browse 100k+ models visually.
- Full API reference lives at the backend Swagger UI once deployed.