FlexAI Checkpoints can be used not only to resume Training and Fine-tuning jobs but also to deploy Inference Endpoints. When a Checkpoint is marked as Inference Ready, it means that the Checkpoint contains all the necessary files and metadata required to deploy an Inference Endpoint directly from it.Documentation Index
Fetch the complete documentation index at: https://docs.flex.ai/llms.txt
Use this file to discover all available pages before exploring further.
Checkpoint Metadata Extraction Process
Currently, the FlexAI Checkpoint Manager automatically extracts metadata from Checkpoints created using the Hugging Face Transformers library. This metadata is used to determine if a Checkpoint contains the necessary information to be marked as Inference Ready and be deployed as an Inference Endpoint:Hugging Face Transformers Checkpoints
Currently, the FlexAI runtime supports Hugging Face Transformers checkpoints, which include thetrainer_state.json and config.json files that contain metadata about the training process and model configuration:
STEP,TRAIN LOSS&EVAL LOSS: Extracted fromtrainer_state.json’slog_historyfield (last entry).MODEL: Determined fromconfig.json’sarchitecturesfield.VERSION: Retrieved fromconfig.json’stransformers_versionfield.INFERENCE READY: Set totrueif thearchitecturesfield is present inconfig.json.
Deploying an Inference-ready Checkpoint
To deploy an Inference Endpoint from an Inference-ready Checkpoint, follow these steps:- Using the FlexAI Console
- Using the FlexAI CLI
Available soon. For now, please use the FlexAI CLI.
Want to learn more about FlexAI Inference
Check out the FlexAI Inference Endpoints documentation for more details.