Ollama now supports HuggingFace GGUF models, making it easier for users to run AI models locally without internet. The GGUF format allows for the use of AI models on modest-sized consumer hardware.
- create a custom base image for a Cloud Workstation environment using a Dockerfile
. Uses:
Quantized models from
A deep dive into model quantization with GGUF and llama.cpp and model evaluation with LlamaIndex