Skip to content

Introduction

leafspark edited this page Aug 31, 2024 · 2 revisions

Overview of AutoGGUF

AutoGGUF is an automated graphical interface designed for GGUF model quantization. Made with PySide6 and llama.cpp, AutoGGUF simplifies the process of quantizing large language models (LLMs) for efficient local inference.

Purpose and Features

The primary purpose of AutoGGUF is to democratize access to LLMs by making quantization more accessible. Key features include:

  1. Automated download and management of llama.cpp backends (including CUDA support)
  2. Easy model selection and quantization
  3. Support for various quantization types
  4. User-friendly graphical interface
  5. Parallel quantization
  6. LoRA conversion
Clone this wiki locally