141 results for “topic:flan-t5”
Swap GPT for any LLM by changing a single line of code. Xinference lets you run open-source, speech, and multimodal models on cloud, on-prem, or your laptop — all through one unified, production-ready inference API.
Toolkit for fine-tuning, ablating and unit-testing open-source LLMs.
This repository contains code for extending the Stanford Alpaca synthetic instruction tuning to existing instruction-tuned models such as Flan-T5.
[Preprint] Learning to Filter Context for Retrieval-Augmented Generaton
Build a Large Language Model (From Scratch) book and Finetuned Models
LLMs4OL: Large Language Models for Ontology Learning
Official implementation of the paper "CoEdIT: Text Editing by Task-Specific Instruction Tuning" (EMNLP 2023)
[ACL 2024 Findings] Code implementation of Paper "Rethinking Negative Instances for Generative Named Entity Recognition"
This repository contains the code to train flan t5 with alpaca instructions and low rank adaptation.
Tools and our test data developed for the HackAPrompt 2023 competition
Empower your LLM to do more than you ever thought possible with these state-of-the-art prompt templates.
Fine-tuning of Flan-5T LLM for text classification 🤖 focuses on adapting a state-of-the-art language model to enhance its ability to classify text data.
A template Next.js app for running language models like FLAN-T5 with Replicate's API
The TABLET benchmark for evaluating instruction learning with LLMs for tabular prediction.
In this implementation, using the Flan T5 large language model, we performed the Text Classification task on the IMDB dataset and obtained a very good accuracy of 93%.
This project uses LLMs to generate music from text by understanding prompts, creating lyrics, determining genre, and composing melodies. It harnesses LLM capabilities to create songs based on text inputs through a multi-step approach.
Use AI to personify books, so that you can talk to them 🙊
This repository contains the lab work for Coursera course on "Generative AI with Large Language Models".
Training and fine-tuning flan-t5-small model based on provided text
Document Summarization App using large language model (LLM) and Langchain framework. Used a pre-trained T5 model and its tokenizer from Hugging Face Transformers library. Created a summarization pipeline to generate summary using model.
In-context learning, Fine-Tuning, RLHF on Flan-T5
The official fork of THoR Chain-of-Thought framework, enhanced and adapted for Emotion Cause Analysis (ECAC-2024)
My solutions to the lab assignments in the Generative AI with Large Language Models course offered by Amazon Web Services.
Flan-t5 model fine tune LoRA and Langchain
Revolutionizing open-world gaming, MergeX harnesses NLP advances to empower players with limitless dialogue interactions with NPCs. By imbuing each character with a unique biography, conversations authentically align with NPC personalities, transcending traditional limitations.
Research POC on the mitigation of bias in large language models (FLAN-T5 and Bloomz) through fine-tuning.
A preliminary investigation for ontology alignment (OM) with large language models (LLMs).
Struggled with creating catchy, human-like titles using ChatGPT or other LLMs? If you've ever been frustrated by AI-generated titles that lack a human touch, I’ve just published something you’ll find useful. My latest blog explores how Reinforcement Learning (RL) can push LLMs to craft more compelling, attention-grabbing headlines
Building an LLM with RLHF involves fine-tuning using human-labeled preferences. Based on Learning to Summarize from Human Feedback, it uses supervised learning, reward modeling, and PPO to improve response quality and alignment.
Master's thesis on Large Language Models for Document Visual Question Answering