Ollama-Buddy 0.9.38: Unload Models, Hide AI Reasoning, and Clearly View Modified Parameters on each request

More improvements to ollama-buddy https://github.com/captainflasmr/ollama-buddy

<2025-04-29 Tue> 0.9.38

Added model unloading functionality to free system resources

Large language models consume significant RAM and GPU memory while loaded. Until now, there wasn’t an easy way to reclaim these resources without restarting the Ollama server entirely. This new functionality allows you to:

There are several ways to unload models with the new functionality:

  1. Unload All Models: Press C-c C-u to unload all running models at once (with confirmation)

  2. Model Management Interface: Access the model management interface with C-c W where you’ll find:

    • A counter showing how many models are currently running
    • An “Unload All” button to free all models at once
    • Individual “Unload” buttons next to each running model
  3. Quick Access in Management Buffer: When in the model management buffer, simply press u to unload all models

The unloading happens asynchronously in the background, with clear status indicators so you can see when the operation completes.

<2025-04-25 Fri> 0.9.37

Enhanced the token statistics section to include any modified parameters, providing a clearer insight into the active configurations. This update helps in debugging and understanding the runtime environment.

<2025-04-25 Fri> 0.9.36

Added Reasoning/Thinking section visibility toggle functionality

This feature is particularly useful when working with AI models that output their “chain of thought” or reasoning process before providing the final answer, allowing users to focus on the end results while still having the option to see the full reasoning when needed.

Comments

comments powered by Disqus