Koboldcpp
Easy-to-use interface for running quantized language models locally with minimal setup. Provides a browser-based frontend so you can interact with models through your web browser. Includes a local API server for integration with other applications. Works with quantized models to run on modest hardware. Supports Llama, Mistral, and other popular open source models. Great for beginners who want to run LLMs without complex configuration. Free and open source.