The official GitHub page for the survey paper "A Survey of Large Language Models".
-
Updated
Aug 20, 2024 - Python
The official GitHub page for the survey paper "A Survey of Large Language Models".
Use PEFT or Full-parameter to finetune 350+ LLMs or 90+ MLLMs. (Qwen2, GLM4v, Internlm2.5, Yi, Llama3.1, Llava-Video, Internvl2, MiniCPM-V-2.6, Deepseek, Baichuan2, Gemma2, Phi3-Vision, ...)
Open Source Pre-training Model Framework in PyTorch & Pre-trained Model Zoo
A one-stop data processing system to make data higher-quality, juicier, and more digestible for (multimodal) LLMs! 🍎 🍋 🌽 ➡️ ➡️🍸 🍹 🍷为大模型提供更高质量、更丰富、更易”消化“的数据!
Papers about pretraining and self-supervised learning on Graph Neural Networks (GNN).
Awesome resources for in-context learning and prompt engineering: Mastery of the LLMs such as ChatGPT, GPT-3, and FlanT5, with up-to-date and cutting-edge updates.
Code for TKDE paper "Self-supervised learning on graphs: Contrastive, generative, or predictive"
An Open-sourced Knowledgable Large Language Model Framework.
Awesome list for research on CLIP (Contrastive Language-Image Pre-Training).
Oscar and VinVL
Tencent Pre-training framework in PyTorch & Pre-trained Model Zoo
Pre-training of Deep Bidirectional Transformers for Language Understanding: pre-train TextCNN
A professional list on Large (Language) Models and Foundation Models (LLM, LM, FM) for Time Series, Spatiotemporal, and Event Data.
Research code for ECCV 2020 paper "UNITER: UNiversal Image-TExt Representation Learning"
[ICML2024] Unified Training of Universal Time Series Forecasting Transformers
Code for ICLR 2020 paper "VL-BERT: Pre-training of Generic Visual-Linguistic Representations".
[NeurIPS 2020] "Graph Contrastive Learning with Augmentations" by Yuning You, Tianlong Chen, Yongduo Sui, Ting Chen, Zhangyang Wang, Yang Shen
[ICLR 2024] Sheared LLaMA: Accelerating Language Model Pre-training via Structured Pruning
Large Language Model-enhanced Recommender System Papers
Code for KDD'20 "Generative Pre-Training of Graph Neural Networks"
Add a description, image, and links to the pre-training topic page so that developers can more easily learn about it.
To associate your repository with the pre-training topic, visit your repo's landing page and select "manage topics."