{"id":2500,"date":"2025-09-18T12:25:05","date_gmt":"2025-09-18T12:25:05","guid":{"rendered":"https:\/\/rssfeedtelegrambot.bnaya.co.il\/index.php\/2025\/09\/18\/docker-model-runner-general-availability\/"},"modified":"2025-09-18T12:25:05","modified_gmt":"2025-09-18T12:25:05","slug":"docker-model-runner-general-availability","status":"publish","type":"post","link":"https:\/\/rssfeedtelegrambot.bnaya.co.il\/index.php\/2025\/09\/18\/docker-model-runner-general-availability\/","title":{"rendered":"Docker Model Runner General Availability"},"content":{"rendered":"<p>We\u2019re excited to share that <a href=\"https:\/\/www.docker.com\/products\/model-runner\/\">Docker Model Runner<\/a> is now generally available (GA)! In April 2025, Docker introduced the <a href=\"https:\/\/www.docker.com\/blog\/introducing-docker-model-runner\/\">first Beta release<\/a> of Docker Model Runner, making it easy to manage, run, and distribute local AI models (specifically LLMs). Though only a short time has passed since then, the product has evolved rapidly, with continuous enhancements driving the product to a reliable level of maturity and stability.<\/p>\n<p>This blog post takes a look back at the most important and widely appreciated capabilities Docker Model Runner brings to developers, and looks ahead to share what they can expect in the near future.<\/p>\n<h2 class=\"wp-block-heading\">What is Docker Model Runner?<\/h2>\n<p>Docker Model Runner (DMR) is built for developers first, making it easy to pull, run, and distribute large language models (LLMs) directly from <a href=\"https:\/\/hub.docker.com\/u\/ai\" target=\"_blank\">Docker Hub<\/a> (in an OCI-compliant format) or <a href=\"https:\/\/huggingface.co\/\" target=\"_blank\">HuggingFace<\/a> (if models are available in the GGUF format, in which case they will be packaged as OCI Artifacts on-the-fly by the HuggingFace backend).<\/p>\n<p>Tightly integrated with Docker Desktop and Docker Engine, DMR lets you serve models through OpenAI-compatible APIs, package GGUF files as OCI artifacts, and interact with them using either the command line, a graphical interface, or developer-friendly (REST) APIs.<\/p>\n<p>Whether you\u2019re creating <a href=\"https:\/\/www.docker.com\/blog\/genai-vs-agentic-ai\/\">generative AI<\/a> applications, experimenting with machine learning workflows, or embedding AI into your software development lifecycle, Docker Model Runner delivers a consistent, secure, and efficient way to <a href=\"https:\/\/www.docker.com\/blog\/run-llms-locally\/\">work with AI models locally<\/a>.<\/p>\n<p>Check the <a href=\"https:\/\/docs.docker.com\/ai\/model-runner\/\" target=\"_blank\">official documentation<\/a> to learn more about Docker Model Runner and its capabilities.<\/p>\n<h2 class=\"wp-block-heading\">Why Docker Model Runner?<\/h2>\n<p>Docker Model Runner makes it easier for developers to experiment and build AI application, including agentic apps, using the same Docker commands and workflows they already use every day. No need to learn a new tool!<\/p>\n<p>Unlike many new AI tools that introduce complexity or require additional approvals, Docker Model Runner fits cleanly into existing enterprise infrastructure. It runs within your current security and compliance boundaries, so teams don\u2019t have to jump through hoops to adopt it.<\/p>\n<p>Model Runner supports OCI-packaged models, allowing you to store and distribute models through any OCI-compatible registry, including Docker Hub. And for teams using Docker Hub, enterprise features like Registry Access Management (RAM) provide policy-based access controls to help enforce guardrails at scale.<\/p>\n<h2 class=\"wp-block-heading\">11 Docker Model Runner Features Developers Love Most<\/h2>\n<p>Below are the features that stand out the most and have been highly valued by the community.<\/p>\n<h3 class=\"wp-block-heading\">1. Powered by llama.cpp\u00a0<\/h3>\n<p>Currently, DMR is built on top of <a href=\"https:\/\/github.com\/ggml-org\/llama.cpp\" target=\"_blank\">llama.cpp<\/a>, which we plan to continue supporting. At the same time, DMR is designed with flexibility in mind, and support for additional inference engines (such as MLX or vLLM) is under consideration for future releases.<\/p>\n<h3 class=\"wp-block-heading\">2. GPU acceleration across macOS and Windows platforms\u00a0<\/h3>\n<p>Harness the full power of your hardware with GPU support: Apple Silicon on macOS, NVIDIA GPUs on Windows, and even ARM\/Qualcomm acceleration \u2014 all seamlessly managed through Docker Desktop.<\/p>\n<h3 class=\"wp-block-heading\">3. Native Linux support\u00a0<\/h3>\n<p>Run DMR on Linux with Docker CE, making it ideal for automation, CI\/CD pipelines, and production workflows.<\/p>\n<h3 class=\"wp-block-heading\">4. CLI and UI experience\u00a0<\/h3>\n<p>Use DMR from the Docker CLI (on both Docker Desktop and Docker CE) or through Docker Desktop\u2019s UI. The UI provides guided onboarding to help even first-time AI developers start serving models smoothly, with automatic handling of available resources (RAM, GPU, etc.).<\/p>\n<div class=\"wp-block-ponyo-image\"><\/div>\n<div class=\"wp-block-ponyo-image\"><\/div>\n<p class=\"has-xs-font-size\">Figure 1: Docker Model Runner works both in Docker Desktop and the CLI, letting you run models locally with the same familiar Docker commands and workflows you already know<\/p>\n\n<h3 class=\"wp-block-heading\">5. Flexible model distribution\u00a0<\/h3>\n<p>Pull and push models from <a href=\"https:\/\/hub.docker.com\/u\/ai\" target=\"_blank\">Docker Hub<\/a> in OCI format, or pull directly from <a href=\"https:\/\/huggingface.co\/\" target=\"_blank\">HuggingFace<\/a> repositories hosting models in GGUF format for maximum flexibility in sourcing and sharing models.<\/p>\n<h3 class=\"wp-block-heading\">6. Open Source and free\u00a0<\/h3>\n<p>DMR is fully open source and free for everyone, lowering the barrier to entry for developers experimenting with or building on AI.<\/p>\n<h3 class=\"wp-block-heading\">7. Secure and controlled\u00a0<\/h3>\n<p>DMR runs in an isolated, controlled environment that doesn\u2019t interfere with the main system or user data (sandboxing). Developers and IT admins can fine-tune security and availability by enabling\/disabling DMR or configuring options like host-side TCP support and CORS.<\/p>\n<h3 class=\"wp-block-heading\">8. Configurable inference settings\u00a0<\/h3>\n<p>Developers can customize context length and llama.cpp runtime flags to fit their use cases, with more configuration options coming soon.<\/p>\n<h3 class=\"wp-block-heading\">9. Debugging support\u00a0<\/h3>\n<p>Built-in request\/response tracing and inspect capabilities make it easier to understand token usage and framework\/library behaviors, helping developers debug and optimize their applications.<\/p>\n<div class=\"wp-block-ponyo-image\"><\/div>\n<p class=\"has-xs-font-size\">Figure 2: Built-in tracing and inspect tools in Docker Desktop make debugging easier, giving developers clear visibility into token usage and framework behavior<\/p>\n<h3 class=\"wp-block-heading\">10. Integrated with the Docker ecosystem\u00a0<\/h3>\n<p>DMR works out of the box with <a href=\"https:\/\/docs.docker.com\/ai\/compose\/models-and-compose\/\" target=\"_blank\">Docker Compose<\/a> and is fully integrated with other Docker products, such as <a href=\"https:\/\/docs.docker.com\/offload\/\" target=\"_blank\">Docker Offload<\/a> (cloud offload service) and <a href=\"https:\/\/docs.docker.com\/testcontainers\/\" target=\"_blank\">Testcontainers<\/a>, extending its reach into both local and distributed workflows.<\/p>\n<h3 class=\"wp-block-heading\">11. Up-to-date model catalog\u00a0<\/h3>\n<p>Access a <a href=\"https:\/\/hub.docker.com\/u\/ai\" target=\"_blank\">curated catalog<\/a> of the most popular and powerful AI models on Docker Hub. These models can be pulled for free and used across development, pipelines, staging, or even production environments.<\/p>\n<div class=\"wp-block-ponyo-image\"><\/div>\n<p class=\"has-xs-font-size\">Figure 3: Curated model catalog on Docker Hub, packaged as OCI Artifacts and ready to run<\/p>\n<h2 class=\"wp-block-heading\">The road ahead<\/h2>\n<p>The future is bright for Docker Model Runner, and the recent GA version is only the first milestone. Below are some of the future enhancements that you should expect to be released soon.<\/p>\n<h3 class=\"wp-block-heading\">Streamlined User Experience\u00a0<\/h3>\n<p>Our goal is to make DMR simple and intuitive for developers to use and debug. This includes richer response rendering in the chat-like interface within Docker Desktop and the CLI, multimodal support in the UI (already available through the API), integration with MCP tools, and enhanced debugging features, alongside expanded configuration options for greater flexibility. Last but not least, we aim to provide smoother and more seamless integration with third-party tools and solutions across the AI ecosystem.<\/p>\n<h3 class=\"wp-block-heading\">Enhancements and better ability to execute\u00a0<\/h3>\n<p>We remain focused on continuously improving DMR\u2019s performance and flexibility for running local models. Upcoming enhancements include support for the most widely used inference libraries and engines, advanced configuration options at the engine and model level, and the ability to deploy Model Runner independently from Docker Engine for production-grade use cases, along with many more improvements on the horizon.<\/p>\n<h3 class=\"wp-block-heading\">Frictionless Onboarding\u00a0<\/h3>\n<p>We want first-time AI developers to start building their applications right away, and to do so with the right foundations. To achieve this, we plan to make onboarding into DMR even more seamless. This will include a guided, step-by-step experience to help developers get started quickly, paired with a set of sample applications built on DMR. These samples will highlight real-world use cases and best practices, providing a smooth entry point for experimenting with and adopting DMR in everyday workflows.<\/p>\n<h3 class=\"wp-block-heading\">Staying on Top of Model Launch\u00a0<\/h3>\n<p>As we continue to enhance inference capabilities, we remain committed to maintaining a first-class catalog of AI models directly in Docker Hub, the leading registry for OCI artifacts, including models. Our goal is to ensure that new, relevant models are available in Docker Hub and runnable through DMR as soon as they are publicly released.<\/p>\n<h3 class=\"wp-block-heading\">Conclusion<\/h3>\n<p>Docker Model Runner has come a long way in a short time, evolving from its Beta release into a mature and stable inference engine that\u2019s now generally available. At its core, the mission has always been clear: make it simple, consistent, and secure for developers to pull, run, and serve AI models locally,. using familiar Docker CLI commands and tools they already love!<\/p>\n<p>Now is the perfect time to get started. If you haven\u2019t already, <a href=\"https:\/\/docs.docker.com\/get-started\/\" target=\"_blank\">install Docker Desktop<\/a> and try out Docker Model Runner today. Follow the <a href=\"https:\/\/docs.docker.com\/ai\/model-runner\/\" target=\"_blank\">official documentation<\/a> to explore its capabilities and see for yourself how DMR can accelerate your journey into building AI-powered applications.<\/p>\n<h4 class=\"wp-block-heading\">Learn more<\/h4>\n<p>Read our quickstart guide to<a href=\"https:\/\/www.docker.com\/blog\/run-llms-locally\/\"> Docker Model Runner<\/a>.<\/p>\n<p>Visit our<a href=\"https:\/\/github.com\/docker\/model-runner\" target=\"_blank\"> Model Runner GitHub repo<\/a>! Docker Model Runner is open-source, and we welcome collaboration and contributions from the community!<\/p>\n<p>Learn how Compose<a href=\"https:\/\/www.docker.com\/blog\/build-ai-agents-with-docker-compose\/\"> works with Model runner, making building AI apps and agents easier<\/a><\/p>\n<p>Learn how to <a href=\"https:\/\/www.docker.com\/blog\/how-to-build-an-ai-tutor-with-model-runner\/\">build an AI tutor<\/a><\/p>\n<p>Explore <a href=\"https:\/\/www.docker.com\/blog\/hybrid-ai-and-how-it-runs-in-docker\/\">how to use both local and remote models<\/a> in hybrid AI workflows\u00a0<\/p>\n<p><a href=\"https:\/\/www.docker.com\/blog\/building-ai-agents-with-goose-and-docker\/\">Building AI agents<\/a> made easy with Goose and Docker<\/p>\n<p><a href=\"https:\/\/www.docker.com\/blog\/docker-model-runner-on-hugging-face\/\">Using Model Runner<\/a> on Hugging Face<\/p>\n<p><a href=\"https:\/\/www.docker.com\/blog\/ai-powered-mock-apis-for-testing-with-docker-and-microcks\/\">Powering AI generated testing<\/a> with Docker Model Runner<\/p>\n<p><a href=\"https:\/\/www.docker.com\/blog\/build-genai-app-with-java-spring-ai-docker-model-runner\/\">Build a GenAI App<\/a> With Java Using Spring AI and Docker Model Runner<\/p>\n<p><a href=\"https:\/\/www.docker.com\/blog\/local-llm-tool-calling-a-practical-evaluation\/\">Tool Calling with Local LLMs<\/a>: A Practical Evaluation<\/p>\n<p><a href=\"https:\/\/www.docker.com\/blog\/behind-the-scenes-how-we-designed-docker-model-runner-and-whats-next\/\">Behind the scenes: How we designed Docker Model Runner and what\u2019s next<\/a><\/p>\n<p><a href=\"https:\/\/www.docker.com\/blog\/why-docker-chose-oci-artifacts-for-ai-model-packaging\/\">Why Docker Chose OCI Artifacts for AI Model Packaging<\/a><\/p>\n<p><a href=\"https:\/\/youtu.be\/904fQI6aNos\" target=\"_blank\">What\u2019s new with Docker Model Runner\u00a0<\/a><\/p>\n<p><a href=\"https:\/\/www.docker.com\/blog\/publish-ai-models-on-docker-hub\/\">Publishing AI models to Docker Hub<\/a><\/p>\n<p><a href=\"https:\/\/docs.google.com\/document\/d\/1ObcUDFBZlb-GQf6JxaMb-9KFYgFlZqDimUc0rgnOh3g\/edit?usp=sharing\" target=\"_blank\">How to Build and Run a GenAI ChatBot from Scratch using Docker Model Runner<\/a>\u00a0<\/p>","protected":false},"excerpt":{"rendered":"<p>We\u2019re excited to share that Docker Model Runner is now generally available (GA)! In April 2025, Docker introduced the first [&hellip;]<\/p>\n","protected":false},"author":0,"featured_media":0,"comment_status":"","ping_status":"","sticky":false,"template":"","format":"standard","meta":{"site-sidebar-layout":"default","site-content-layout":"","ast-site-content-layout":"default","site-content-style":"default","site-sidebar-style":"default","ast-global-header-display":"","ast-banner-title-visibility":"","ast-main-header-display":"","ast-hfb-above-header-display":"","ast-hfb-below-header-display":"","ast-hfb-mobile-header-display":"","site-post-title":"","ast-breadcrumbs-content":"","ast-featured-img":"","footer-sml-layout":"","ast-disable-related-posts":"","theme-transparent-header-meta":"","adv-header-id-meta":"","stick-header-meta":"","header-above-stick-meta":"","header-main-stick-meta":"","header-below-stick-meta":"","astra-migrate-meta-layouts":"default","ast-page-background-enabled":"default","ast-page-background-meta":{"desktop":{"background-color":"var(--ast-global-color-4)","background-image":"","background-repeat":"repeat","background-position":"center center","background-size":"auto","background-attachment":"scroll","background-type":"","background-media":"","overlay-type":"","overlay-color":"","overlay-opacity":"","overlay-gradient":""},"tablet":{"background-color":"","background-image":"","background-repeat":"repeat","background-position":"center center","background-size":"auto","background-attachment":"scroll","background-type":"","background-media":"","overlay-type":"","overlay-color":"","overlay-opacity":"","overlay-gradient":""},"mobile":{"background-color":"","background-image":"","background-repeat":"repeat","background-position":"center center","background-size":"auto","background-attachment":"scroll","background-type":"","background-media":"","overlay-type":"","overlay-color":"","overlay-opacity":"","overlay-gradient":""}},"ast-content-background-meta":{"desktop":{"background-color":"var(--ast-global-color-5)","background-image":"","background-repeat":"repeat","background-position":"center center","background-size":"auto","background-attachment":"scroll","background-type":"","background-media":"","overlay-type":"","overlay-color":"","overlay-opacity":"","overlay-gradient":""},"tablet":{"background-color":"var(--ast-global-color-5)","background-image":"","background-repeat":"repeat","background-position":"center center","background-size":"auto","background-attachment":"scroll","background-type":"","background-media":"","overlay-type":"","overlay-color":"","overlay-opacity":"","overlay-gradient":""},"mobile":{"background-color":"var(--ast-global-color-5)","background-image":"","background-repeat":"repeat","background-position":"center center","background-size":"auto","background-attachment":"scroll","background-type":"","background-media":"","overlay-type":"","overlay-color":"","overlay-opacity":"","overlay-gradient":""}},"footnotes":""},"categories":[4],"tags":[],"class_list":["post-2500","post","type-post","status-publish","format-standard","hentry","category-docker"],"_links":{"self":[{"href":"https:\/\/rssfeedtelegrambot.bnaya.co.il\/index.php\/wp-json\/wp\/v2\/posts\/2500","targetHints":{"allow":["GET"]}}],"collection":[{"href":"https:\/\/rssfeedtelegrambot.bnaya.co.il\/index.php\/wp-json\/wp\/v2\/posts"}],"about":[{"href":"https:\/\/rssfeedtelegrambot.bnaya.co.il\/index.php\/wp-json\/wp\/v2\/types\/post"}],"replies":[{"embeddable":true,"href":"https:\/\/rssfeedtelegrambot.bnaya.co.il\/index.php\/wp-json\/wp\/v2\/comments?post=2500"}],"version-history":[{"count":0,"href":"https:\/\/rssfeedtelegrambot.bnaya.co.il\/index.php\/wp-json\/wp\/v2\/posts\/2500\/revisions"}],"wp:attachment":[{"href":"https:\/\/rssfeedtelegrambot.bnaya.co.il\/index.php\/wp-json\/wp\/v2\/media?parent=2500"}],"wp:term":[{"taxonomy":"category","embeddable":true,"href":"https:\/\/rssfeedtelegrambot.bnaya.co.il\/index.php\/wp-json\/wp\/v2\/categories?post=2500"},{"taxonomy":"post_tag","embeddable":true,"href":"https:\/\/rssfeedtelegrambot.bnaya.co.il\/index.php\/wp-json\/wp\/v2\/tags?post=2500"}],"curies":[{"name":"wp","href":"https:\/\/api.w.org\/{rel}","templated":true}]}}