When To Use This
Usetm infer models when you want to enumerate models from the selected inference host.
On the default Tensormesh serverless host, this route works without an inference API key. On On-Demand, it requires both Authorization and X-User-Id.
Usage
Examples
List models from the default serverless host.
List models from a routed On-Demand host.
Options
| Name | Type | Required | Default | Details | |
|---|---|---|---|---|---|
--surface | `choice[on-demand | serverless]` | no | "serverless" | Inference surface to target. |
--user-id | text | no | X-User-Id header to send. Only used for —surface on-demand. | ||
--api-key | text | no | Inference API key (Authorization: Bearer …). | ||
--base-url | text | no | Override the base URL for the selected surface. | ||
--timeout | float | no | HTTP connect timeout in seconds for the inference request. |
Inherited Global Options
| Name | Type | Required | Default | Details | ||||
|---|---|---|---|---|---|---|---|---|
--version, -V | boolean | no | false | Show the version and exit. Boolean flag. | ||||
--config | path | no | "~/.config/tensormesh/config.toml" | Path to config TOML file | ||||
--output | `choice[text | json | yaml | raw | table]` | no | "text" | Output format (text is human-readable; json is machine-friendly). |
--quiet | boolean | no | false | Suppress non-essential output. Boolean flag. | ||||
--debug | boolean | no | false | Print debug logs to stderr (secrets redacted). Boolean flag. | ||||
--ca-bundle | path | no | Path to a PEM CA bundle for TLS verification (overrides TENSORMESH_CA_BUNDLE). | |||||
--max-retries | integer | no | Max retries for idempotent HTTP requests on transient errors (overrides TENSORMESH_MAX_RETRIES; subcommands may override). | |||||
--controlplane-base | text | no | Override the Control Plane base URL. | |||||
--gateway-provider | text | no | Inference Gateway provider for built-in host selection (nebius, lambda, yotta). |
Caveats
- The returned
idvalue is the model string to pass back asmodelin serverless requests. - For On-Demand, the returned
idvalue is the served model name to reuse in routed requests.

