🤖 Working with AI Models in Libre WebUI
This guide explains how to use AI models in Libre WebUI. Whether you're new to AI or an experienced user, this guide will help you get the most out of your local AI assistant.
~8 minutes - Complete guide from basics to advanced model management
🎯 What You Can Do
Libre WebUI supports all the features that modern AI assistants offer:
💬 Chat & Conversations
- Have natural conversations with AI models
- Get streaming responses (words appear as they're generated)
- Use advanced settings like temperature and creativity controls
- Create custom system prompts to change the AI's personality
🖼️ Vision & Images
- Upload images and ask questions about them
- Analyze charts, diagrams, and photographs
- Get help with visual tasks like describing scenes or reading text in images
📝 Structured Responses
- Request responses in specific formats (JSON, lists, etc.)
- Get organized summaries and analysis
- Use predefined templates for common tasks
🛠️ Model Management
- Download and manage AI models locally
- Switch between different models for different tasks
- Monitor model performance and memory usage
🧠 AI Models Guide
- 🌱 Starter Models
- 🚀 Advanced Models
- 🎯 Specialized Models
Perfect for newcomers and everyday use:
Model | Size | Memory | Best For |
---|---|---|---|
gemma3:4b | ~4GB | 8GB RAM | Most users - Best balance |
llama3.2:1b | ~1GB | 4GB RAM | Fast responses - Ultra-quick |
phi4:14b | ~14GB | 16GB RAM | Power users - Microsoft's best |
gemma3:4b is currently the best single-GPU model for most users. Great balance of speed and intelligence.
For users with powerful hardware:
Model | Size | Memory | Best For |
---|---|---|---|
llama3.3:70b | ~40GB | 64GB RAM | State-of-the-art performance |
deepseek-r1:32b | ~16GB | 32GB RAM | Advanced reasoning |
devstral:24b | ~12GB | 24GB RAM | Best for coding |
qwen2.5vl:32b | ~16GB | 32GB RAM | Image understanding |
These models require significant RAM and processing power. Check your system specifications first.
For specific use cases:
Model | Size | Specialty | Use Case |
---|---|---|---|
qwen2.5vl:3b | ~3GB | Vision | Analyze images and photos |
devstral:24b | ~12GB | Coding | Programming assistance |
deepseek-r1:32b | ~16GB | Reasoning | Complex problem solving |
llama4:16x17b | ~16GB | Multimodal | Text, images, and more |
- Vision tasks → qwen2.5vl models
- Programming → devstral models
- Reasoning → deepseek-r1 models
- General use → gemma3 or llama3 models
Advanced Models
- llama3.3:70b - State-of-the-art performance (40GB+)
- deepseek-r1:32b - Advanced reasoning (16GB)
- devstral:24b - Best for coding (12GB)
- qwen2.5vl:32b - Image understanding (16GB)
Model Sizes
- 1B = ~1GB memory, ultra-fast
- 4B = ~3GB memory, great balance
- 14B = ~8GB memory, powerful
- 32B+ = 16GB+ memory, professional use
🚀 Getting Started with Models
Step 1: Download Your First Model
- Go to the Models section in the sidebar
- Click "Pull Model"
- Enter a model name like
gemma3:4b
- Wait for the download to complete
Step 2: Start Chatting
- Go back to the Chat section
- You'll see your model is now available
- Type a message and press Enter
- Watch the AI respond in real-time!
Step 3: Try Advanced Features
- Upload an image (with vision models like
qwen2.5vl:32b
) - Adjust settings like creativity and response length
- Create custom prompts to change the AI's behavior
🎨 Creative Use Cases
Writing Assistant
"Help me write a professional email to..."
"Proofread this document and suggest improvements"
"Create a story outline about..."
Learning & Research
"Explain quantum physics in simple terms"
"What are the pros and cons of..."
"Help me understand this concept by giving examples"
Programming Helper (with devstral:24b)
"Create a complete web application with authentication"
"Debug this complex codebase and suggest improvements"
"Build an autonomous coding agent for this project"
Image Analysis (with qwen2.5vl:32b)
"What's in this image and what does it mean?"
"Extract all text from this document accurately"
"Analyze this complex chart and provide insights"
Advanced Reasoning (with deepseek-r1:32b)
"Think through this complex problem step by step"
"What are the hidden implications of this decision?"
"Solve this multi-step logical puzzle"
⚙️ Advanced Features
Custom System Prompts
Change how the AI behaves by setting a system prompt:
"You are a helpful programming tutor. Always explain concepts step by step."
"You are a creative writing assistant. Help me brainstorm ideas."
"You are a professional editor. Focus on clarity and grammar."
Structured Outputs
Ask for responses in specific formats:
"List the pros and cons in JSON format"
"Give me a summary with bullet points"
"Create a table comparing these options"
Temperature & Creativity
- Low temperature (0.1-0.3): Focused, consistent responses
- Medium temperature (0.5-0.7): Balanced creativity and coherence
- High temperature (0.8-1.0): More creative and varied responses
🔍 Model Capabilities Reference
Model Type | Best For | Memory Needed | Speed |
---|---|---|---|
Compact Efficiency | Quick tasks, edge devices | 1-3GB | Ultra-Fast |
Single-GPU Champions | Balanced performance, daily use | 3-8GB | Fast |
Reasoning Specialists | Complex thinking, problem-solving | 12-16GB | Medium |
Vision-Language | Image analysis, multimodal tasks | 8-32GB | Medium |
Coding Agents | Advanced programming, debugging | 12-24GB | Medium |
State-of-the-Art | Maximum capability, research | 32GB+ | Slower |
💡 Tips for Better Results
Writing Better Prompts
- Be specific: "Write a 200-word summary" vs "Summarize this"
- Give context: "I'm a beginner" or "I'm an expert in..."
- Ask for examples: "Show me examples of..."
- Specify format: "Give me a numbered list" or "Explain step by step"
Managing Performance
- Use smaller models for simple tasks to save memory
- Switch models based on your current task
- Monitor memory usage in the Models section
- Keep frequently used models loaded for faster responses
Privacy & Security
✅ Your data never leaves your computer ✅ No internet connection required (after downloading models) ✅ Full control over your conversations ✅ No tracking or data collection
🆘 Troubleshooting
Model won't download?
- Check your internet connection
- Make sure you have enough disk space
- Try a smaller model first
Responses are slow?
- Try a smaller model (1B or 3B parameters)
- Close other applications to free up memory
- Check if multiple models are loaded
AI gives strange responses?
- Adjust the temperature setting
- Try rephrasing your question
- Use a different model for your task
Ready to explore? Head to the Quick Start Guide to get your first conversation going!