Quick start
Pull and run the default Mistral:
Need more power? Try a Mixture‑of‑Experts variant if supported by your hardware.
Fast, lightweight models for everyday tasks and coding. Run locally with Ollama on Windows.
Pull and run the default Mistral:
Need more power? Try a Mixture‑of‑Experts variant if supported by your hardware.
Mistral 7B: great speed/quality on most PCs; 8–16 GB RAM recommended. MoE variants (e.g., Mixtral) offer higher quality but need more RAM/VRAM.
Keep models on SSD. For acceleration, see GPU Acceleration.
• Coding assistance and quick snippets
• Summarization and rewriting
• Lightweight chatbots and on‑device assistants
• Rapid prototyping with low latency
• Use concise, explicit prompts (“You are a coding assistant…”) for best results.
• Prefer smaller quantizations for speed; larger ones for quality if hardware allows.
• Close GPU‑heavy apps and keep drivers up to date.
• Benchmark your setup to pick the sweet spot: Benchmarks.
Community‑driven guide. Not affiliated with the official Ollama project.