cURL
curl --request POST \ --url https://external.nebius.tensormesh.ai/v1/responses \ --header 'Authorization: Bearer <token>' \ --header 'Content-Type: application/json' \ --header 'X-User-Id: <x-user-id>' \ --data ' { "model": "openai-gpt-oss-120b-gpu-type-h200x1_8nic16", "input": "Say hello." } '
{ "id": "resp_123", "object": "response", "model": "<string>", "output": [ { "id": "out_123", "type": "message", "role": "<string>", "status": "<string>", "content": [ { "type": "output_text", "text": "hello", "annotations": [ {} ] } ] } ], "created_at": 123, "status": "<string>" }
Create responses on the selected Tensormesh On-Demand host.
Authorization: Bearer <API_KEY>
X-User-Id: <uuid>
Bearer authentication using your On-Demand API key. Format: Bearer <API_KEY>
Tensormesh user id used for attribution and routing.
On-Demand served model name to use.
"openai-gpt-oss-120b-gpu-type-h200x1_8nic16"
Input passed to the responses endpoint.
Optional limit for generated output tokens.
Successful Response
"resp_123"
"response"
Show child attributes