3 results for “topic:non-autoregressive-transformers”
[ICLR 2025] Official Implementation of Meissonic: Revitalizing Masked Generative Transformers for Efficient High-Resolution Text-to-Image Synthesis
A fast speech-to-speech & speech-to-text translation model that supports simultaneous decoding and offers 28× speedup.
Official repository for our NeurIPS 2024 paper: DiffNorm: Self-Supervised Normalization for Non-autoregressive Speech-to-speech Translation