Skip to content

inference inspect

Displays detailed information about an Inference Endpoint, including its name, ID, creator, owner, configuration, runtime status, creation date, endpoint URL, and its Playground URL (a hosted Chainlit 🔗 instance that’s already hooked up to your Inference Endpoint).

Terminal window
flexai inference inspect <inference_endpoint_name>
Required

The name of the Inference Endpoint to inspect.

Examples
  • mixtral_8x7b
--json
<boolean>
Optional
Flag

Output the information in JSON format.

Examples
  • --json
FieldDescriptionData Type
kindThe type of resourceString
metadataMetadata information about the resourceObject
metadata.nameThe name of the Inference EndpointString
metadata.idThe ID of the Inference EndpointString
metadata.creatorUserIDThe ID of the user who created the Inference EndpointString
metadata.ownerOrgIDThe ID of the organization that owns the Inference EndpointString
configConfiguration information about the Inference EndpointObject
config.deviceThe desired architecture for the Inference EndpointString
config.acceleratorThe number of accelerators per serverInteger
config.apiKeySecretNameThe name of the secret containing the API key for the Inference EndpointString
config.playgroundUrlThe URL of the Playground for the Inference EndpointString
config.endpointUrlThe URL of the Inference Playground. Useful for testing and prototyping via a browser-based InterfaceString
config.hfTokenSecretNameThe name of the secret containing the Hugging Face tokenString
config.engineArgsAdditional arguments specific to vLLMObject
config.engineArgs.modelThe name of the model being servedString
runtimeRuntime information about the Inference EndpointObject
runtime.statusThe status of the Inference EndpointString
runtime.createdAtThe creation date of the Inference EndpointString