GGUF
•
gemma
Google's smallest Gemma 3 model in new quantization format that preserves bfloat16 quality
Model info
Model
gemma 3 1B QAT
Author
Repository
Arch
gemma
Parameters
1B
Format
gguf
Size on disk
about 720.43 MB
Download the model using lms
— LM Studio's developer CLI.
lms get gemma-3-1b-it-qat
curl http://localhost:1234/v1/chat/completions \
-H "Content-Type: application/json" \
-d '{
"model": "gemma-3-1b-it-qat",
"messages": [
{ "role": "system", "content": "Always answer in rhymes." },
{ "role": "user", "content": "Introduce yourself." }
],
"temperature": 0.7,
"max_tokens": -1,
"stream": true
}'
lms log stream
to see your prompts as they are sent to the LLM.lmstudio.js
- LM Studio SDK documentation (TypeScript)lms log stream
- Stream server logslms
- LM Studio's CLI documentation