The name of the Inference Endpoint to inspect.
Examples
-
mixtral_8x7b
Displays detailed information about an Inference Endpoint, including its name, ID, creator, owner, configuration, runtime status, creation date, endpoint URL, and its Playground URL (a hosted Chainlit 🔗 instance that’s already hooked up to your Inference Endpoint).
flexai inference inspect <inference_endpoint_name>
The name of the Inference Endpoint to inspect.
mixtral_8x7b
Output the information in JSON format.
--json
Field | Description | Data Type |
---|---|---|
kind | The type of resource | String |
metadata | Metadata information about the resource | Object |
metadata.name | The name of the Inference Endpoint | String |
metadata.id | The ID of the Inference Endpoint | String |
metadata.creatorUserID | The ID of the user who created the Inference Endpoint | String |
metadata.ownerOrgID | The ID of the organization that owns the Inference Endpoint | String |
config | Configuration information about the Inference Endpoint | Object |
config.device | The desired architecture for the Inference Endpoint | String |
config.accelerator | The number of accelerators per server | Integer |
config.apiKeySecretName | The name of the secret containing the API key for the Inference Endpoint | String |
config.playgroundUrl | The URL of the Playground for the Inference Endpoint | String |
config.endpointUrl | The URL of the Inference Playground. Useful for testing and prototyping via a browser-based Interface | String |
config.hfTokenSecretName | The name of the secret containing the Hugging Face token | String |
config.engineArgs | Additional arguments specific to vLLM | Object |
config.engineArgs.model | The name of the model being served | String |
runtime | Runtime information about the Inference Endpoint | Object |
runtime.status | The status of the Inference Endpoint | String |
runtime.createdAt | The creation date of the Inference Endpoint | String |