Awesome-Reasoning-Foundation-Models
survey.pdf
|
A curated list of awesome large AI models, or foundation models, for reasoning.
We organize the current foundation models into three categories: language foundation models, vision foundation models, and multimodal foundation models. Further, we elaborate the foundation models in reasoning tasks, including commonsense, mathematical, logical, causal, visual, audio, multimodal, agent reasoning, etc. Reasoning techniques, including pre-training, fine-tuning, alignment training, mixture of experts, in-context learning, and autonomous agent, are also summarized.
We welcome contributions to this repository to add more resources. Please submit a pull request if you want to contribute! See CONTRIBUTING.
Table of Contents
table of contents
0 Survey
This repository is primarily based on the following paper:
A Survey of Reasoning with Foundation Models
Jiankai Sun, Chuanyang Zheng, Enze Xie, Zhengying Liu, Ruihang Chu, Jianing Qiu, Jiaqi Xu, Mingyu Ding, Hongyang Li, Mengzhe Geng, Yue Wu, Wenhai Wang, Junsong Chen, Zhangyue Yin, Xiaozhe Ren, Jie Fu, Junxian He, Wu Yuan, Qi Liu, Xihui Liu, Yu Li, Hao Dong, Yu Cheng, Ming Zhang, Pheng Ann Heng, Jifeng Dai, Ping Luo, Jingdong Wang, Ji-Rong Wen, Xipeng Qiu, Yike Guo, Hui Xiong, Qun Liu, and Zhenguo Li
If you find this repository helpful, please consider citing:
@article{sun2023survey,
title={A Survey of Reasoning with Foundation Models},
author={Sun, Jiankai and Zheng, Chuanyang and Xie, Enze and Liu, Zhengying and Chu, Ruihang and Qiu, Jianing and Xu, Jiaqi and Ding, Mingyu and Li, Hongyang and Geng, Mengzhe and others},
journal={arXiv preprint arXiv:2312.11562},
year={2023}
}
1 Relevant Surveys and Links
relevant surveys
-
Combating Misinformation in the Age of LLMs: Opportunities and Challenges - [arXiv] [Link]
-
The Rise and Potential of Large Language Model Based Agents: A Survey - [arXiv] [Link]
-
Multimodal Foundation Models: From Specialists to General-Purpose Assistants - [arXiv] [Tutorial]
-
A Survey on Multimodal Large Language Models - [arXiv] [Link]
-
Self-Supervised Multimodal Learning: A Survey - [arXiv] [Link]
-
Large AI Models in Health Informatics: Applications, Challenges, and the Future - [arXiv] [Paper] [Link]
-
Towards Reasoning in Large Language Models: A Survey - [arXiv] [Paper] [Link]
-
Reasoning with Language Model Prompting: A Survey - [arXiv] [Paper] [Link]
-
Awesome Multimodal Reasoning - [Link]
2 Foundation Models
foundation models
Table of Contents - 2
foundation models (table of contents)
2.1 Language Foundation Models
LFMs
Foundation Models (Back-to-Top)
-
2023/10
|Mistral
| Mistral 7B - [Paper] [Code] -
2023/09
|Qwen
| Qwen Technical Report - [Paper] [Code] [Project] -
2023/07
|Llama 2
| Llama 2: Open Foundation and Fine-Tuned Chat Models - [Paper] [Code] [Blog] -
2023/07
|InternLM
| InternLM: A Multilingual Language Model with Progressively Enhanced Capabilities - [Paper] [Code] [Project] -
2023/05
|PaLM 2
| PaLM 2 Technical Report - -
2023/03
|PanGu-Σ
| PanGu-Σ: Towards Trillion Parameter Language Model with Sparse Heterogeneous Computing - [Paper] -
2023/03
|Vicuna
| Vicuna: An Open-Source Chatbot Impressing GPT-4 with 90%* ChatGPT Quality - [Blog] [Code] -
2023/03
|GPT-4
| GPT-4 Technical Report - [Paper] [Blog] -
2023/02
|LLaMA
| LLaMA: Open and Efficient Foundation Language Models - [Paper] [Code] [Blog] -
2022/11
|ChatGPT
| Chatgpt: Optimizing language models for dialogue - [Blog] -
2022/04
|PaLM
| PaLM: Scaling Language Modeling with Pathways - [Paper] [Blog] -
2021/09
|FLAN
| Finetuned Language Models Are Zero-Shot Learners - -
2021/07
|Codex
| Evaluating Large Language Models Trained on Code - -
2021/05
|GPT-3
| Language Models are Few-Shot Learners - [Paper] [Code] -
2021/04
|PanGu-α
| PanGu-α: Large-scale Autoregressive Pretrained Chinese Language Models with Auto-parallel Computation - [Paper] [Code] -
2019/08
|Sentence-BERT
| Sentence-BERT: Sentence Embeddings using Siamese BERT-Networks - -
2019/07
|RoBERTa
| RoBERTa: A Robustly Optimized BERT Pretraining Approach - -
2018/10
|BERT
| BERT: Pre-training of Deep Bidirectional Transformers for Language Understanding - [Paper] [Code] [Blog]