Skip to content

Command: inference

The flexai inference command manages Inference Endpoints by allowing for their deployment and management.

An Inference Endpoint is a hosted model that can be used for inference tasks, such as text generation, image classification, and more.

Inference Endpoints are created from models hosted on Hugging Face, and they can be deployed to FlexAI’s infrastructure for easy access and scalability.

You can manage Inference Endpoints using the flexai inference set of subcommands.