shimmy
« Back to VersTracker
Description:
Small local inference server with OpenAI-compatible GGUF endpoints
Type: Formula  |  Latest Version: 1.8.2@0  |  Tracked Since: Dec 8, 2025
Links: Homepage  |  formulae.brew.sh
Category: Ai ml
Tags: ai llm inference gguf local-ai api
Install: brew install shimmy
About:
Shimmy is a lightweight local inference server that exposes GGUF-format language models through an OpenAI-compatible API. It allows developers to drop-in replace OpenAI calls with local models without changing application code. This provides a privacy-preserving, offline-capable alternative to cloud-based AI services.
Key Features:
  • OpenAI-compatible API endpoints
  • GGUF model format support
  • Local inference without cloud dependencies
  • Lightweight and easy to deploy
Use Cases:
  • Local AI development and testing
  • Privacy-sensitive applications requiring on-premise inference
  • Cost reduction by avoiding cloud API fees
Alternatives:
  • llama.cpp – Lower-level C++ implementation; Shimmy provides easier OpenAI-compatible API layer
  • Ollama – More feature-rich with model management; Shimmy is simpler and more lightweight
Version History
Detected Version Rev Change Commit
Dec 8, 2025 11:26pm 0 VERSION_BUMP fd64710e
Oct 10, 2025 6:48pm 0 VERSION_BUMP 0b54b705
Oct 4, 2025 10:30am 0 VERSION_BUMP 0795fa15
Oct 4, 2025 8:20am 0 VERSION_BUMP 9b05c6ef
Sep 23, 2025 5:35am 0 VERSION_BUMP 1cdc6c59
Sep 23, 2025 5:12am 0 VERSION_BUMP 61c6357c
Sep 18, 2025 5:44am 0 VERSION_BUMP 16e75882
Sep 18, 2025 5:23am 0 VERSION_BUMP 64396032
Sep 14, 2025 9:24pm 0 VERSION_BUMP a83b22be
Sep 12, 2025 12:57am 0 VERSION_BUMP 9edcf1f5