Technical Trigger

The openclaw onboard command now accepts the --auth-choice huggingface-api-key option, allowing OpenClaw agents to be moved to Hugging Face models. Additionally, the llama-server command can be used to start a local server with a built-in web UI for running models locally.

Developer / Implementation Hook

Developers can use the openclaw onboard command with the --auth-choice huggingface-api-key option to move their OpenClaw agents to Hugging Face models. They can also install llama.cpp using brew install llama.cpp or winget install llama.cpp and start a local server with the llama-server command to run models locally. The openclaw config file can be updated to use a Hugging Face model by adding the model property with the primary value set to the desired model, such as huggingface/zai-org/GLM-5:fastest.

The Structural Shift

The paradigm is shifting from relying on closed hosted models to using open models and local control, enabling greater flexibility and autonomy for OpenClaw agents.

Early Warning — Act Before Mainstream

To take advantage of this change, developers can: 1. Create a Hugging Face token and add it to their OpenClaw config using the openclaw onboard command with the --auth-choice huggingface-api-key option. 2. Install llama.cpp using brew install llama.cpp or winget install llama.cpp and start a local server with the llama-server command. 3. Update their OpenClaw config to use a Hugging Face model by adding the model property with the primary value set to the desired model, such as huggingface/zai-org/GLM-5:fastest.