Skip to content

Commit 71ee7a4

Browse files
unamedkrclaude
andcommitted
README: clarify Gemma 4 status as experimental (non-standard GGUF)
The gemma4 GGUF format is not yet supported by llama.cpp itself (error: "unknown model architecture: gemma4"). Our implementation produces semantically relevant tokens but has repetition issues. This is likely a GGUF conversion limitation, not an inference bug. Co-Authored-By: Claude Opus 4.6 (1M context) <noreply@anthropic.com>
1 parent bfde618 commit 71ee7a4

1 file changed

Lines changed: 1 addition & 1 deletion

File tree

README.md

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -111,7 +111,7 @@ Cross-model (4b K + Q4 V): SmolLM2 1.7B (-1.6%), Qwen3.5 0.8B (+0.9%), Qwen3.5 4
111111
| Qwen3.5-4B | Qwen3.5 (DeltaNet) | 4B | PPL verified |
112112
| Qwen3.5-35B-A3B | Qwen2-MoE | 35B (3B active) | Working |
113113
| Gemma 3 270M | Gemma 3 | 270M | Working |
114-
| Gemma 4 E2B | Gemma 4 | 2B | WIP |
114+
| Gemma 4 E2B | Gemma 4 | 2B | Experimental (non-standard GGUF) |
115115

116116
Architectures: Llama/Qwen3.5 (shared path), Gemma 3/4 (sliding + full attention), Qwen2-MoE.
117117

0 commit comments

Comments
 (0)