Use Cases
AI Text Generation

Quantized GGUF models

7min

Here's a step-by-step guide to running quantized LLM models in multi-part GGUF format. We will use Unsloth's Deepseek-R1 Q8_0 model as an example. This model is very large and will require an 8xH200 machine configuration, but you can also follow this guide for much smaller models.

Before moving on with the guide, Setup your Vast account and add credit. Review the quickstart guide to get familar with the service if you do not have an account with credits loaded.

Llama.cpp

Llama.cpp is the recommended method for loading these models as it is able to directly load a split file of many parts without first merging them.

While it's easy to build llama.cpp inside one of our instances, we will focus on running this model in the Open WebUI template which contains a pre-compiled CUDA compatible versions of llama-server and llama-cli.

Open WebUI Template

OpenWebui + Ollama is one of our recommended templates. While its default setup uses Ollama as a backend, it can also access an OpenAI-compatible API and it has been pre-configured to find one running on http://localhost:20000

A full guide to getting started with the OpenWebUI template is available here

Ensure you have enough disk space and a suitable configuration. For Deepseek-R1 Q8_0 you'll need:

  • At least 800GB VRAM
  • 700GB storage space

The recommended configuration for this particular model is 8 x H200 with 750GB storage.

Once you have loaded up the template, you'll need to open up a terminal where we will pull and then serve the model.

Pulling the model

You will want to download the models from the Deepseek-R1 Q8_0 model hugging face repo to the /workspace/llama.cpp/models directory on your instance. We have included a script with the Ollama + Open WebUI template that you may use to easily download the models.

Bash


This download will take some time as HuggingFace limits download speed, so even on an instance with very fast download speeds it may take up to an hour to completely download.

Serving the model

Once the dowload has completed it's time to serve the model using the pre-built llama-server application.

Again, from the terminal, type the following:

JS


This command will load all of the model layers into GPU VRAM and begin serving the API at http://localhost:20000

Once the model has finished loading to the GPU, it will be availabe directly from the OpenWebui interface in the model selector. Again, this may take some time to load and if you already have OpenWebui open then you may need to refresh the page.

Building Llama.cpp

If you prefer to build llama.cpp yourself, you can simply run the following from any Vast-built template. The Recommended Nvidia CUDA template would be an ideal start.

Bash


These commands will build the llama-quantize llama-cli llama-server and llama-gguf-split tools.

For advanced build instructions you should see the official documentation on GitHub.

Further Reading

Please see the template Readme for advanced template configuration, particularly if you would like to modify the template to make the llama-server API available externally with authentication or via a SSH tunnel.