197 results for “topic:pre-training”
The official GitHub page for the survey paper "A Survey of Large Language Models".
Data processing for and with foundation models! 🍎 🍋 🌽 ➡️ ➡️🍸 🍹 🍷
Open Source Pre-training Model Framework in PyTorch & Pre-trained Model Zoo
Papers about pretraining and self-supervised learning on Graph Neural Networks (GNN).
Awesome resources for in-context learning and prompt engineering: Mastery of the LLMs such as ChatGPT, GPT-3, and FlanT5, with up-to-date and cutting-edge updates.
Code for TKDE paper "Self-supervised learning on graphs: Contrastive, generative, or predictive"
Unified Training of Universal Time Series Forecasting Transformers
An Open-sourced Knowledgable Large Language Model Framework.
Awesome list for research on CLIP (Contrastive Language-Image Pre-Training).
A professional list on Large (Language) Models and Foundation Models (LLM, LM, FM) for Time Series, Spatiotemporal, and Event Data.
Tencent Pre-training framework in PyTorch & Pre-trained Model Zoo
Oscar and VinVL
Pre-training of Deep Bidirectional Transformers for Language Understanding: pre-train TextCNN
Research code for ECCV 2020 paper "UNITER: UNiversal Image-TExt Representation Learning"
Code for ICLR 2020 paper "VL-BERT: Pre-training of Generic Visual-Linguistic Representations".
Large Language Model-enhanced Recommender System Papers
Official repository for "Craw4LLM: Efficient Web Crawling for LLM Pretraining"
[ICLR 2024] Sheared LLaMA: Accelerating Language Model Pre-training via Structured Pruning
[NeurIPS 2020] "Graph Contrastive Learning with Augmentations" by Yuning You, Tianlong Chen, Yongduo Sui, Ting Chen, Zhangyang Wang, Yang Shen
The repository of ET-BERT, a network traffic classification model on encrypted traffic. The work has been accepted as The Web Conference (WWW) 2022 accepted paper.
Multi-modality pre-training
Code for KDD'20 "Generative Pre-Training of Graph Neural Networks"
Code for our SIGKDD'22 paper Pre-training-Enhanced Spatial-Temporal Graph Neural Network For Multivariate Time Series Forecasting.
[NeurlPS D&B 2024] Generative AI for Math: MathPile
Conceptual 12M is a dataset containing (image-URL, caption) pairs collected for vision-and-language pre-training.
The official repo for [NeurIPS'23] "SAMRS: Scaling-up Remote Sensing Segmentation Dataset with Segment Anything Model"
One-click training of your own GPT. Training a GPT has never been easier for beginners. / 一键预训练+SFT一个属于自己的LLM,0基础训练GPT原来可以这么简单?
Paper List of Pre-trained Foundation Recommender Models
[Survey] Masked Modeling for Self-supervised Representation Learning on Vision and Beyond (https://arxiv.org/abs/2401.00897)
[CVPR 2024 Highlight] Visual Point Cloud Forecasting