Thanks to visit codestin.com
Credit goes to docs.together.ai

Skip to main content
from together import Together
client = Together()

completion = client.chat.completions.create(
  model="openai/gpt-oss-20b",
  messages=[{"role": "user", "content": "What are the top 3 things to do in New York?"}],
)

print(completion.choices[0].message.content)

New: Dedicated Container Inference

Deploy your own Dockerized workloads on Together’s managed GPU infrastructure. You bring the container — Together handles provisioning, autoscaling, and observability.
Our models:Together hosts many popular models via our serverless endpoints and dedicated endpoints. On serverless, you’ll be charged based on the tokens you use and size of the model. On dedicated, you’ll be charged based on GPU hours.
Build AI apps and agents with Together:
Explore our services: