Clippy Download

Local LLM chatbot featuring 1990s Microsoft Office Clippy UI, running GGUF models via Llama.cpp/node-llama-cpp with one-click installation for Gemma3/Llama 3.2/Phi-4/Qwen3, automatic hardware acceleration (Metal/CUDA/Vulkan), custom model/prompt support, offline operation, and @electron/llm reference implementation with NOASSERTION license.

⭐ 1,132 stars on GitHub
Latest Release: v0.4.3

About Software

Clippy lets you run large language models (LLMs) locally on your computer while sticking with a 1990s user interface. Through Llama.cpp, it supports models in popular GGUF format (most publicly available models). Comes with one-click installation support for Google's Gemma3, Meta's Llama 3.2, Microsoft's Phi-4, and Qwen's Qwen3. A love letter and homage to the late great Clippy, assistant from Microsoft Office 1997. Character designed by illustrator Kevan Atteberry, who created 15+ potential characters for Microsoft's Office Assistants. Not affiliated, approved, or supported by Microsoft. Consider it software art or software satire. Also meant as reference implementation of @electron/llm hoping to help other Electron app developers make use of local language models.

Features simple, familiar, classic chat interface (send messages to models, get responses), batteries included with no complicated setup (just open app and chat), thanks to llama.cpp and node-llama-cpp automatic discovery of most efficient way to run models (Metal, CUDA, Vulkan, etc.), custom models/prompts/parameters (load your own downloaded models and play with settings), offline/local/free (everything runs on computer, only network request for update checks which can be disabled). Non-features: Countless little chat apps for local LLMs exist; many likely better. This project isn't trying to be your best chat bot but rather weird mix of nostalgia for 1990s technology paired with magical 2025 technology running on computers. Supports most GGUF models via Llama.cpp; find GGUF models from TheBloke or Unsloth on HuggingFace. Thanks to Microsoft, Kevan Atteberry for Clippy, Jordan Scales for Windows 98 design, Pooya Parsa for extracting Clippy spritesheet frames, and node-llama-cpp for squeezing llama.cpp into Node.js. NOASSERTION license.

Use Cases:

  • Local LLM chatbot with 1990s Microsoft Office Clippy UI running GGUF models via Llama.cpp and node-llama-cpp
  • Nostalgic desktop app combining classic Clippy character design from 1997 with modern LLMs (Gemma3, Llama 3.2, Phi-4, Qwen3)
  • Reference implementation for @electron/llm demonstrating local language model integration in Electron apps
  • Simple familiar chat interface with no complicated setup, batteries-included one-click model installation, and offline operation
  • Custom GGUF model loading with adjustable prompts/parameters, automatic hardware acceleration (Metal/CUDA/Vulkan), and privacy-focused local execution

Downloads

v0.4.3 May 08, 2025
Clippy-0.4.3-setup-x64.exeexe
clippy_0.4.3_amd64.debdeb
v0.4.2 May 07, 2025
Clippy-0.4.2-setup-x64.exeexe
clippy_0.4.2_amd64.debdeb
v0.4.1 May 06, 2025
Clippy-0.4.1-setup-x64.exeexe
clippy_0.4.1_amd64.debdeb
v0.4.0 May 05, 2025
Clippy-0.4.0-setup-x64.exeexe
clippy_0.4.0_amd64.debdeb
v0.3.0 May 03, 2025
Clippy-0.3.0-setup-x64.exeexe
clippy_0.3.0_amd64.debdeb
v0.2.0 May 02, 2025
Clippy-0.2.0-setup-x64.exeexe
clippy_0.2.0_amd64.debdeb

Package Info

Last Updated
May 08, 2025
Latest Version
v0.4.3
License
NOASSERTION
Total Versions
6

README

Clippy

Clippy (https://felixrieseberg.github.io/clippy/) let's you run a variety of large language models (LLMs) locally on your computer while sticking with a user interface of the 1990s. Through Llama.cpp, it supports models in the popular GGUF format, which is to say most publicly available models. It comes with one-click installation support for Google's Gemma3, Meta's Llama 3.2, Microsoft's Phi-4, and Qwen's Qwen3.

It's a love letter and homage to the late, great Clippy, the assistant from Microsoft Office 1997. The character was designed by illustrator Kevan Atteberry, who created more than 15 potential characters for Microsoft's Office Assistants. This app is not affiliated, approved, or supported by Microsoft. Consider it software art. If you don't like it, consider it software satire.

It is also meant to be a reference implementation of @electron/llm (https://github.com/electron/llm), hoping to help other developers of Electron apps make use of local language models.

Features

  • Simple, familiar, and classic chat interface. Send messages to your models, get a response.
  • Batteries included: No complicated setup. Just open the app and chat away. Thanks to llama.cpp and node-llama-cpp, the app will automatically discover the most efficient way to run your models (Metal, CUDA, Vulkan, etc).
  • Custom models, prompts, and parameters: Load your own downloaded models and play with the settings.
  • Offline, local, free: Everything runs on your computers. The only network request Clippy makes is to check for updates (which you can disable).

Non-Features

Countless little chat apps for local LLMs exist out there. Many of them are likely better - and that's okay. This project isn't trying to be your best chat bot. I'd like you to enjoy a weird mix of nostalgia for 1990s technology paired with one the most magical technologies we can run on our computers in 2025.

Downloading More Models

Clippy supports (thanks to Llama.cpp) most GGUF models. You can find GGUF models in plenty of online sources - I tend to go with models quantized by TheBloke (https://huggingface.co/thebloke) or Unsloth (https://huggingface.co/unsloth).

Acknowledgements

Thanks to:

  • I am so grateful to Microsoft - not only for everything they've done for Electron, but also for giving us one of the most iconic characters and designs of computing history.
  • Kevan Atteberry (https://www.kevanatteberry.com/) for Clippy
  • Jordan Scales (@jdan) (https://github.com/jdan) for the Windows 98 design
  • Pooya Parsa (@pi0) (https://github.com/pi0) for being the (as far as I know) person to extract the length of each frame from the Clippy spritesheet.
  • node-llama-cpp (https://github.com/withcatai/node-llama-cpp) for squeezing llama.cpp into Node.js