Changelog: 2025-05-19
Changed
Section titled “Changed”-
FlexAI Inference vLLM Arguments: The
flexai inference serve
command’s optionalvLLM_arguments
have switched from a “Supported argument list” schema to a “Unsupported argument list” schema. This enables you to safely pass most ofvLLM Engine Arguments
🔗 when creating an Inference Endpoint.Currently, the only unsupported argument is
--device
, which value is handled by FlexAI.