Create a conda virtual environment which is used for model inference Convert data format for next step lora finetuning and stored in JSON files with 2-prefix python ...
MLX needs native macOS/Metal. Docker Desktop runs Linux containers, so inference runs on the host while the UI/API runs in Docker.