55 results for “topic:kl-divergence”
A pytorch package for non-negative matrix factorization.
IJCAI 2021, "Comparing Kullback-Leibler Divergence and Mean Squared Error Loss in Knowledge Distillation"
A PyTorch Implementation of Generating Sentences from a Continuous Space by Bowman et al. 2015.
This repository summarizes techniques for KL divergence vanishing problem.
Repository for "Blending Data-Driven Priors in Dynamic Games" - RSS 2024
Experiments with variational autoencoders in Julia
Implementations of basic concepts dealt under the Reinforcement Learning umbrella. This project is collection of assignments in CS747: Foundations of Intelligent and Learning Agents (Autumn 2017) at IIT Bombay
This repository contains the lab work for Coursera course on "Generative AI with Large Language Models".
Code for enumerating and evaluating numerical methods for Langevin dynamics using near-equilibrium estimates of the KL-divergence. Accompanies https://doi.org/10.3390/e20050318
PyTorch implementation of α-geodesical skew divergence
PyTorch implementations of the beta divergence loss.
KL-loss
A self-distillation based training method for long context reasoning in a single LLM without reinforcement learning
Machine Learning algorithms built from scratch for AMMI Machine Learning course
Code for "EMA Policy Gradient: Taming Reinforcement Learning for LLMs with EMA Anchor and Top-k KL" (arxiv.org/abs/2602.04417)
My MSc project on applying, tuning and modifying the PPO and A2C algorithms to Pettingzoo MARL library two player poker game
No description provided.
Relative entropy, mutual information, KL divergence of 2 given Images 🖼
Change point detection using KL divergence
Basic GANs with variety of loss functions as an exercise for my Thesis with Prof. Randy Paffenroth. KL, Reverse-KL, JS and Wasserstein GAN.
Hyperspectral unmixing using Variational Autoencoders with Dirichlet latent distributions, achieving state-of-the-art performance on endembers and abundances reconstruction.
This repository includes some detailed proofs of "Bias Variance Decomposition for KL Divergence".
Implementation of KL Divergence and inverted vector model for plagiarism detection in text files
Coordinate ascent mean-field variational inference (CAVI) using the evidence lower bound (ELBO) to iteratively perform the optimal variational factor distribution parameter updates for clustering.
A collection of summarizer algorithms
The Dirichlet Mechanism for Differentially Private KL Divergence Minimization
Implementation of diffusion models with varying noise distributions (Gaussian, GMM, Gamma) and scheduling techniques (cosine, sigmoid) to assess generative performance using KL divergence and dynamic scheduling approaches.
NLP implementations like information-theoretic measures of distributional similarity, text preprocessing using shell commands, Naive Bayes text categorization model, Cocke-Younger-Kasami parsing.
Python information theory computation
Using Monte-Carlo simulated datasets, a completely transparent Boltzmann Machine trained on 1-D Ising chain data is implemented to predict model couplers in the absence of past coupler values. Methods from machine learning applied to theoretical physics are on display in this work.