Use your locally running AI models to assist you in your web browsing
-
Updated
Dec 20, 2025 - TypeScript
Use your locally running AI models to assist you in your web browsing
Reliable model swapping for any local OpenAI/Anthropic compatible server - llama.cpp, vllm, etc
A generalized information-seeking agent system with Large Language Models (LLMs).
[ICML 2024] SqueezeLLM: Dense-and-Sparse Quantization
[NeurIPS 2024] KVQuant: Towards 10 Million Context Length LLM Inference with KV Cache Quantization
Run Open Source/Open Weight LLMs locally with OpenAI compatible APIs
A flexible, AI powered C2 framework built with operators in mind
The PyVisionAI Official Repo
MVP of an idea using multiple local LLM models to simulate and play D&D
Run multiple resource-heavy Large Models (LM) on the same machine with limited amount of VRAM/other resources by exposing them on different ports and loading/unloading them on demand
Fenix Ai Trading Bot with LangGraph and ollama and multipe providers
Cognito: Supercharge your Chrome browser with AI. Guide, query, and control everything using natural language.
Unified management and routing for llama.cpp, MLX and vLLM models with web dashboard.
Vesta macOS Distribution - Official releases and downloads.Vesta AI Chat Assistant for macOS - Built with SwiftUI and Apple Intelligence using Apple's on device model on MacOs Tahoe (MacOS 26). Now with side-by-side Qwen3-VL for vison
A Cli, a webUI, and a MCP server for the Z-Image-Turbo text-to-image generation model (Tongyi-MAI/Z-Image-Turbo base model as well as quantized models)
Add a description, image, and links to the localllm topic page so that developers can more easily learn about it.
To associate your repository with the localllm topic, visit your repo's landing page and select "manage topics."