345 results for “topic:finetuning-llms”
Mastering Applied AI, One Concept at a Time
Create Custom LLMs
Deploy any AI model, agent, database, RAG, and pipeline locally or remotely in minutes
Official repository of my book "A Hands-On Guide to Fine-Tuning LLMs with PyTorch and Hugging Face"
End to End Generative AI Industry Projects on LLM Models with Deployment_Awesome LLM Projects
One-stop handbook for building, deploying, and understanding LLM agents with 60+ skeletons, tutorials, ecosystem guides, and evaluation tools.
Train Large Language Models on MLX.
Enhancing LLMs with LoRA
'afm' command cli: macOS server and single prompt mode that exposes Apple's Foundation and MLX Models and other APIs running on your Mac through a single aggregated OpenAI-compatible API endpoint. Supports Apple Vision and single command (non-server) inference with piping as well . Now with Web Browser and local AI API aggregator
IndexTTS Fine-tuning notebooks
Auto Data is a library designed for quick and effortless creation of datasets tailored for fine-tuning Large Language Models (LLMs).
🚀 Easy, open-source LLM finetuning with one-line commands, seamless cloud integration, and popular optimization frameworks. ✨
On Memorization of Large Language Models in Logical Reasoning
Vesta macOS Distribution - Official releases and downloads.Vesta AI Chat Assistant for macOS - Built with SwiftUI, Swift MLX and Apple Intelligence using Apple's on device model on MacOs Tahoe (MacOS 26). Now with side-by-side Qwen3-VL for vison
MediNotes: SOAP Note Generation through Ambient Listening, Large Language Model Fine-Tuning, and RAG
A Streamlit app for generating high-quality Q&A training datasets from text and PDFs, leveraging Gemini, Claude, and OpenAI for LLM fine-tuning.
A Gradio web UI for Large Language Models. Supports LoRA/QLoRA finetuning,RAG(Retrieval-augmented generation) and Chat
Fine-tune Mistral 7B to generate fashion style suggestions
[NeurIPS 2023] Make Your Pre-trained Model Reversible: From Parameter to Memory Efficient Fine-Tuning
the small distributed language model toolkit; fine-tune state-of-the-art LLMs anywhere, rapidly
SEIKO is a novel reinforcement learning method to efficiently fine-tune diffusion models in an online setting. Our methods outperform all baselines (PPO, classifier-based guidance, direct reward backpropagation) for fine-tuning Stable Diffusion.
Medical Language Model fine-tuned using pretraining, instruction tuning, and Direct Preference Optimization (DPO). Progresses from general medical knowledge to specific instruction following, with experiments in preference alignment for improved medical text generation and understanding.
A c++ framework on efficient training & fine-tuning LLMs
A user-friendly interface built on top of Thinking Machines Tinker API that lets you fine-tune LLMs, chat with your trained model, and deploy to Hugging Face.
A open-source framework designed to adapt pre-trained Language Models (LLMs), such as Llama, Mistral, and Mixtral, to a wide array of domains and languages.
GaLore: Memory-Efficient LLM Training by Gradient Low-Rank Projection
Fine-Tuning LLMs (Gemma, LLaMA, Mistral, etc.) A practical guide to fine-tuning various large language models using popular frameworks. Includes examples, scripts, and tips for efficient training on custom datasets.
A comprehensive GUI wrapper application for Apple's Foundation Models Adapter Training Toolkit, providing an intuitive interface for training LoRA adapters for Apple's on-device foundation models.
[EMNLP 2024] Quantize LLM to extremely low-bit, and finetune the quantized LLMs
Official GitHub repository for Phison aiDAPTIV. Storage-accelerated AI training & inference platform.