Skip to content

Add Q4_K/Q5_K/Q6_K GPU support via Q8_0 dequantization#108

Open
AdamBien wants to merge 2 commits intobeehive-lab:mainfrom
AdamBien:main
Open

Add Q4_K/Q5_K/Q6_K GPU support via Q8_0 dequantization#108
AdamBien wants to merge 2 commits intobeehive-lab:mainfrom
AdamBien:main

Conversation

@AdamBien
Copy link
Copy Markdown
Contributor

  • Add GPU support for K-quant models (Q4_K_M, Q5_K_M, Q6_K) via load-time dequantization to Q8_0
  • New FloatTensor implementations: Q4_KFloatTensor, Q5_KFloatTensor, Q6_KFloatTensor
  • Dequantization correctly handles TornadoVM's 16-byte ARRAY_HEADER memory layout
  • Centralize weight loading log message in AbstractModelLoader (shows actual model quantization, e.g. "Q4_K_M -> Q8_0")

Tested with:
./llamaTornado --gpu --verbose-init --metal --model /Users/abien/work/workspaces/llms/Devstral-Small-2-24B-Instruct-2512-Q4_K_M.gguf --prompt "who are you?" --gpu-memory 30GB

@CLAassistant
Copy link
Copy Markdown

CLA assistant check
Thank you for your submission! We really appreciate it. Like many open source projects, we ask that you sign our Contributor License Agreement before we can accept your contribution.
You have signed the CLA already but the status is still pending? Let us recheck it.

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment

Labels

None yet

Projects

None yet

Development

Successfully merging this pull request may close these issues.

2 participants