90 results for “topic:captioning”
A modular framework for vision & language multimodal research from Facebook AI Research (FAIR)
streamline the fine-tuning process for multimodal models: PaliGemma 2, Florence-2, and Qwen2.5-VL
JoyCaption is an image captioning Visual Language Model (VLM) being built from the ground up as a free, open, and uncensored model for the community to use in training Diffusion models.
Code for "Aligning Linguistic Words and Visual Semantic Units for Image Captioning", ACM MM 2019
CapDec: SOTA Zero Shot Image Captioning Using CLIP and GPT2, EMNLP 2022 (findings)
Audio Captioning datasets for PyTorch.
A Tennis dataset and models for event detection & commentary generation
VisText is a benchmark dataset for semantically rich chart captioning.
Medical image captioning using OpenAI's CLIP
Fully-Convolutional Point Networks for Large-Scale Point Clouds
Python code for handling the Clotho dataset.
What and How Well You Performed? A Multitask Learning Approach to Action Quality Assessment [CVPR 2019]
A Base Tensorflow Project for Medical Report Generation
Metrics for evaluating Automated Audio Captioning systems, designed for PyTorch.
[CVPR 2023 & IJCV 2025] Positive-Augmented Contrastive Learning for Image and Video Captioning Evaluation
[CVPR21] Visual Semantic Role Labeling for Video Understanding (https://arxiv.org/abs/2104.00990)
Using LLMs and pre-trained caption models for super-human performance on image captioning.
Audio captioning baseline system for DCASE 2020 challenge.
[CVPR 2022] X-Trans2Cap: Cross-Modal Knowledge Transfer using Transformer for 3D Dense Captioning
A tool to streamline AI image captioning
CaMEL: Mean Teacher Learning for Image Captioning. ICPR 2022
Toolkit for supporting the EBU-TT Live specification
Some papers about *diverse* image (a few videos) captioning
No description provided.
A curated list of zero-shot captioning papers
My notes on some Deep Learning papers
[ICCV 2023] With a Little Help from your own Past: Prototypical Memory Networks for Image Captioning.
Tools for the evaluation of audio captioning.
S2VT (seq2seq) video captioning with bahdanau & luong attention implementation in Tensorflow
Python and command-line utility for aligning audio to a transcript.