---
name: together-performance-tuning
description: |
  Together AI performance tuning for inference, fine-tuning, and model deployment.
  Use when working with Together AI's OpenAI-compatible API.
  Trigger: "together performance tuning".
allowed-tools: Read, Write, Edit, Bash(pip:*), Grep
version: 1.0.0
license: MIT
author: Jeremy Longshore <jeremy@intentsolutions.io>
tags: [saas, ai, inference, together]
compatible-with: claude-code
---

# Together AI Performance Tuning

## Overview

Guidance for performance tuning with Together AI inference and fine-tuning API.

## Instructions

### Key Points

- Together AI is OpenAI-compatible: `base_url = 'https://api.together.xyz/v1'`
- Use the `together` Python SDK or any OpenAI client library
- Supports 100+ open-source models (Llama, Mixtral, Qwen, FLUX)
- Fine-tuning available for supported models
- Batch inference at 50% cost reduction

## Error Handling

| Error | Cause | Solution |
|-------|-------|----------|
| `401 Unauthorized` | Invalid API key | Check at api.together.xyz |
| `Model not found` | Wrong model ID | Use `client.models.list()` |
| `429 Rate limit` | Too many requests | Implement backoff |
| `500 Server error` | Model overloaded | Retry with backoff |

## Resources

- [Together AI Docs](https://docs.together.ai/)
- [API Reference](https://docs.together.ai/reference/chat-completions-1)
- [Model List](https://docs.together.ai/docs/inference-models)

## Next Steps

See related Together AI skills for more patterns.
