GitHunt
WO

woheller69/LLAMA_TK_CHAT

Simple chat interface for local AI using llama-cpp-python and llama-cpp-agent

LLAMA_TK_CHAT

Simple chat interface for local AI using llama-cpp-python and llama-cpp-agent

Quickstart

The TK GUI is based on llama-cpp-python, llama-cpp-agent, typer, and tkinter package
(install them with pip install ...).

The GUI is a self-contained Python script named LLAMA_TK_GUI.py. As long as
its package dependencies are present, you can download and run it from wherever you like.

Specify the path to the model with the -m/--model argument and the prompt template with -f/--format (default: CHATML, Options: MISTRAL, CHATML, VICUNA, LLAMA_2, SYNTHIA, NEURAL_CHAT, SOLAR, OPEN_CHAT, ALPACA, CODE_DS, B22, LLAMA_3, PHI_3, OPEN_INTERPRETER, AUTOCODER, GEMMA_2, DEEP_SEEK_CODER_2)
A optional system prompt can be set with -s/--sysprompt.
Context length defaults to 2048, it can be set with -c/--context-length. The number of threads can be set with -t/--n-threads (default: 4 threads). Temperature can be set with --temperature, default is 0.65.
-l/--mlock allows to use MLOCK instead of MMAP.
Everything runs on CPU.

Input your request in the bottom window and click Generate.

Inference can be interrupted using the Stop button.

Languages

Python100.0%

Contributors

MIT License
Created May 8, 2024
Updated January 24, 2026
woheller69/LLAMA_TK_CHAT | GitHunt