Skip to content

Conversation

@BrewTestBot
Copy link
Contributor

Created by brew bump


Created with brew bump-formula-pr.

Details

release notes
# Xmas-release :santa:   **LocalAI 3.9.0**! 🚀




LocalAI 3.9.0 is focused on stability, resource efficiency, and smarter agent workflows. We've addressed critical issues with model loading, improved system resource management, and introduced a new Agent Jobs panel for scheduling and managing background agentic tasks. Whether you're running models locally or orchestrating complex agent workflows, this release makes it faster, more reliable, and easier to manage.

📌 TL;DR

Feature Summary
Agent Jobs Panel Schedule and run background tasks with cron or via API — perfect for automated workflows.
Smart Memory Reclaimer Automatically frees up GPU/VRAM by evicting least recently used models when memory is low.
LRU Model Eviction Models are automatically unloaded from memory based on usage patterns to prevent crashes.
MLX & CUDA 13 Support New model backends and enhanced GPU compatibility for modern hardware.
UI Polish & Fixes Cleaned-up navigation, fixed layout overflow, and various improvements.
Vibevoice Added support for the vibevoice backend!

🚀 New Features

🤖 Agent Jobs Panel: Schedule & Automate Tasks

LocalAI 3.9.0 introduces a new Agent Jobs panel in the web UI and API, allowing you to create, run, and schedule agentic tasks in the background that can be started programmatically via API or from the Web interface.

  • Run agent prompts on a schedule using cron syntax, or via API.
  • Agents are defined via the model settings, supporting MCP.
  • Trigger jobs via API for integration into CI/CD or external tools.
  • Optionally send results to a webhook for post-processing.
  • Templates and prompts can be dynamically populated with variables.

✅ Use cases: Daily reports, CI integration, automated data processing, scheduled model evaluations.

Screenshot 2025-12-24 at 15-26-32 LocalAI - Agent Jobs

🧠 Smart Memory Reclaimer: Auto-Optimize GPU Resources

We’ve introduced a new Memory Reclaimer that monitors system memory usage and automatically frees up GPU/VRAM when needed.

Screenshot 2025-12-24 at 15-25-30 LocalAI API - 8b3e0eb (8b3e0ebf8aab4071ef7721121f04081c32a5c9da)
  • Tracks memory consumption across all backends.
  • When usage exceeds a configured threshold, it evicts the least recently used (LRU) models.
  • Prevents out-of-memory crashes and keeps your system stable during high load.

This is a step toward adaptive resource management, future versions will expand this with more advanced policies and giving more control.


🔁 LRU Model Eviction: Intelligent Model Management

Building on the new reclaimer, LocalAI now supports LRU (Least Recently Used) eviction for loaded models.

Screenshot 2025-12-24 at 15-27-24 LocalAI - Settings
  • Set a maximum number of models to keep in memory (e.g., limit to 3).
  • When a new model is loaded and the limit is reached, the oldest unused model is automatically unloaded.
  • Fully compatible with single_active_backend mode (now defaults to LRU=1 for backward compatibility).

💡 Ideal for servers with limited VRAM or when running multiple models in parallel.


🖥️ UI & UX Polish

  • Fixed navbar ordering and login icon — clearer navigation and better visual flow.
  • Prevented tool call overflow in chat view — no more clipped or misaligned content.
  • Uniformed link paths (e.g., /browse/ instead of browse) for consistency.
  • Fixed model selection toggle — header updates correctly when switching models.
  • Consistent button styling — uniform colors, hover effects, and accessibility.

📦 Backward Compatibility & Architecture

  • Dropped x86_64 Mac support: no longer maintained in GitHub Actions; ARM64 (M1/M2/M3/M4) is now the recommended architecture.
  • Updated data storage path from /usr/share to /var/lib: follows Linux conventions for mutable data.
  • Added CUDA 13 support: now available in Docker images and L4T builds.
  • New VibeVoice TTS backend real-time text-to-speech with voice cloning support. You can install it from the model gallery!
  • StableDiffusion-GGML now supports LoRA: expand your image-generation capabilities.

🛠️ Fixes & Improvements

  • Issue: After v3.8.0, /readyz and /healthz endpoints required authentication, breaking Docker health checks and monitoring tools
  • Issue: Fixed crashes when importing models from Hugging Face URLs with subfolders (e.g., huggingface://user/model/GGUF/model.gguf).

🚀 The Complete Local Stack for Privacy-First AI

LocalAI Logo

LocalAI

The free, Open Source OpenAI alternative. Drop-in replacement REST API compatible with OpenAI specifications for local AI inferencing. No GPU required.

Link: https://github.com/mudler/LocalAI

LocalAGI Logo

LocalAGI

Local AI agent management platform. Drop-in replacement for OpenAI's Responses API, supercharged with advanced agentic capabilities and a no-code UI.

Link: https://github.com/mudler/LocalAGI

LocalRecall Logo

LocalRecall

RESTful API and knowledge base management system providing persistent memory and storage capabilities for AI agents. Works alongside LocalAI and LocalAGI.

Link: https://github.com/mudler/LocalRecall


❤️ Thank You

LocalAI is a true FOSS movement — built by contributors, powered by community.

If you believe in privacy-first AI:

  • Star the repo
  • 💬 Contribute code, docs, or feedback
  • 📣 Share with others

Your support keeps this stack alive.


✅ Full Changelog

📋 Click to expand full changelog

What's Changed

Breaking Changes 🛠

Bug fixes :bug:

Exciting New Features 🎉

🧠 Models

📖 Documentation and examples

👒 Dependencies

Other Changes

New Contributors

Full Changelog: mudler/LocalAI@v3.8.0...v3.9.0

View the full release notes at https://github.com/mudler/LocalAI/releases/tag/v3.9.0.


@github-actions github-actions bot added go Go use is a significant feature of the PR or issue bump-formula-pr PR was created using `brew bump-formula-pr` labels Dec 24, 2025
@github-actions
Copy link
Contributor

🤖 An automated task has requested bottles to be published to this PR.

Caution

Please do not push to this PR branch before the bottle commits have been pushed, as this results in a state that is difficult to recover from. If you need to resolve a merge conflict, please use a merge commit. Do not force-push to this PR branch.

@github-actions github-actions bot added the CI-published-bottle-commits The commits for the built bottles have been pushed to the PR branch. label Dec 24, 2025
@BrewTestBot BrewTestBot added this pull request to the merge queue Dec 24, 2025
Merged via the queue into main with commit 089ace0 Dec 24, 2025
22 checks passed
@BrewTestBot BrewTestBot deleted the bump-localai-3.9.0 branch December 24, 2025 17:41
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment

Labels

bump-formula-pr PR was created using `brew bump-formula-pr` CI-published-bottle-commits The commits for the built bottles have been pushed to the PR branch. go Go use is a significant feature of the PR or issue

Projects

None yet

Development

Successfully merging this pull request may close these issues.

3 participants