Official model cards and documentation for AI models available on Docker Hub.
π Description:
DeepCoder-14B-Preview is a code reasoning LLM fine-tuned to scale up to long context lengths.
π Model File: ai/deepcoder-preview.md
π³ Docker Hub: docker.io/ai/deepcoder-preview
Source:
π Description:
Distilled LLaMA by DeepSeek, fast and optimized for real-world tasks.
π Model File: ai/deepseek-r1-distill-llama.md
π³ Docker Hub: docker.io/ai/deepseek-r1-distill-llama
Sources:
- https://huggingface.co/deepseek-ai/DeepSeek-R1-Distill-Llama-8B
- https://huggingface.co/deepseek-ai/DeepSeek-R1-Distill-Llama-70B
π Description:
Google's latest Gemma, small yet strong for chat and generation.
π Model File: ai/gemma3.md
π³ Docker Hub: docker.io/ai/gemma3
Source:
π Description:
Google's latest Gemma, in its QAT (quantization aware trained) variant.
π Model File: ai/gemma3-qat.md
π³ Docker Hub: docker.io/ai/gemma3-qat
π Description:
Efficient multimodal AI for text, image, audio, and video on low-resource devices.
π Model File: ai/gemma3n.md
π³ Docker Hub: docker.io/ai/gemma3n
π Description:
OpenAI's open-weight models designed for powerful reasoning, agentic tasks.
π Model File: ai/gpt-oss.md
π³ Docker Hub: docker.io/ai/gpt-oss
π Description:
Granite Embedding Multilingual is a 278 million parameter, encoderβonly XLMβRoBERTaβstyle model.
π Model File: ai/granite-embedding-multilingual.md
π³ Docker Hub: docker.io/ai/granite-embedding-multilingual
π Description:
Meta's LLaMA 3.1: Chat-focused, benchmark-strong, multilingual-ready.
π Model File: ai/llama3.1.md
π³ Docker Hub: docker.io/ai/llama3.1
Source:
π Description:
Solid LLaMA 3 update, reliable for coding, chat, and Q&A tasks.
π Model File: ai/llama3.2.md
π³ Docker Hub: docker.io/ai/llama3.2
Sources:
- https://huggingface.co/meta-llama/Llama-3.2-3B-Instruct
- https://huggingface.co/meta-llama/Llama-3.2-1B-Instruct
π Description:
Newest LLaMA 3 release with improved reasoning and generation quality.
π Model File: ai/llama3.3.md
π³ Docker Hub: docker.io/ai/llama3.3
π Description:
24B multimodal instruction model by Mistral AI, tuned for accuracy, tool use & fewer repeats.
π Model File: ai/magistral-small-3.2.md
π³ Docker Hub: docker.io/ai/magistral-small-3.2
π Description:
Efficient open model with top-tier performance and fast inference.
π Model File: ai/mistral.md
π³ Docker Hub: docker.io/ai/mistral
Source:
π Description:
Mistral fine-tuned via NVIDIA NeMo for smoother enterprise use.
π Model File: ai/mistral-nemo.md
π³ Docker Hub: docker.io/ai/mistral-nemo
Source:
π Description:
mxbai-embed-large-v1 is a top English embed model by Mixedbread AI, great for RAG and more.
π Model File: ai/mxbai-embed-large.md
π³ Docker Hub: docker.io/ai/mxbai-embed-large
Source:
π Description:
Nomic Embed Text v1 is an openβsource, fully auditable text embedding model.
π Model File: ai/nomic-embed-text-v1.5.md
π³ Docker Hub: docker.io/ai/nomic-embed-text-v1.5
π Description:
Microsoft's compact model, surprisingly capable at reasoning and code.
π Model File: ai/phi4.md
π³ Docker Hub: docker.io/ai/phi4
Source:
π Description:
Versatile Qwen update with better language skills and wider support.
π Model File: ai/qwen2.5.md
π³ Docker Hub: docker.io/ai/qwen2.5
Source:
π Description:
Qwen3 is the latest Qwen LLM, built for top-tier coding, math, reasoning, and language tasks.
π Model File: ai/qwen3.md
π³ Docker Hub: docker.io/ai/qwen3
π Description:
Qwen3-Coder is Qwen's new series of coding agent models.
π Model File: ai/qwen3-coder.md
π³ Docker Hub: docker.io/ai/qwen3-coder
π Description:
Experimental Qwen variantβlean, fast, and a bit mysterious.
π Model File: ai/qwq.md
π³ Docker Hub: docker.io/ai/qwq
Source:
π Description:
Designed for reasoning, agent and general capabilities, and versatile developer-friendly features.
π Model File: ai/seed-oss.md
π³ Docker Hub: docker.io/ai/seed-oss
π Description:
Tiny LLM built for speed, edge devices, and local development.
π Model File: ai/smollm2.md
π³ Docker Hub: docker.io/ai/smollm2
Sources:
- https://huggingface.co/HuggingFaceTB/SmolLM2-360M-Instruct
- https://huggingface.co/HuggingFaceTB/SmolLM2-135M-Instruct
π Description:
SmolLM3 is a 3.1B model for efficient on-device use, with strong performance in chat.
π Model File: ai/smollm3.md
π³ Docker Hub: docker.io/ai/smollm3
π Description:
Devstral Small 1.1 is an agentic coding LLM (24B) fine-tuned from Mistral-Small-3.1 with a 128K context window.
π Model File: ai/devstral-small.md
π³ Docker Hub: docker.io/ai/devstral-small
A command-line tool for working with model cards. See tools/model-cards-cli/README.md for details.
Key features:
- Update "Available model variants" tables in model card markdown files
- Inspect model repositories to extract metadata
- Upload model overviews to Docker Hub
- Support for custom namespaces and private repositories
To add or update model cards:
- Create/edit the markdown file in the
ai/
directory - Use the Model Cards CLI to update variant tables
- Submit a pull request
See LICENSE file for details.