1). Segment Anything Model - presents a set of resources to establish foundational models for image segmentation; releases the largest segmentation dataset with over 1 billion masks on 11M licensed images; the model’s zero-shot performance is competitive with or even superior to fully supervised results. (paper)
![Twitter avatar for @MetaAI](https://substackcdn.com/image/twitter_name/w_96/MetaAI.jpg)
2). Instruction Tuning with GPT-4 - presents GPT-4-LLM, a "first attempt" to use GPT-4 to generate instruction-following data for LLM fine-tuning; the dataset is released and includes 52K unique English and Chinese instruction-following data; the dataset is used to instruction-tune LLaMA models which leads to superior zero-shot performance on new tasks. (paper)
![Twitter avatar for @omarsar0](https://substackcdn.com/image/twitter_name/w_96/omarsar0.jpg)
![Image](https://substackcdn.com/image/fetch/w_600,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fpbs.substack.com%2Fmedia%2FFtEq3nMX0AEisHm.jpg)
3). 8 Things to Know about LLMs - discusses important considerations regarding the capabilities and limitations of LLMs. (paper)
![Twitter avatar for @sleepinyourhat](https://substackcdn.com/image/twitter_name/w_96/sleepinyourhat.jpg)
![A paper header for "Eight things to know about large language models" by Sam Bowman.](https://substackcdn.com/image/fetch/w_600,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fpbs.substack.com%2Fmedia%2FFsu7LOcacAAo3Uk.png)
4). A Survey of LLMs - a new 50 pages survey on large language models. (paper)
![Twitter avatar for @omarsar0](https://substackcdn.com/image/twitter_name/w_96/omarsar0.jpg)
![Image](https://substackcdn.com/image/fetch/w_600,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fpbs.substack.com%2Fmedia%2FFswIzxsX0AEjNnI.jpg)
5). Baize - an open-source chat model fine-tuned with LoRA. Leverages 100K dialogs generated from ChatGPT chatting with itself; it releases the dialogs along with 7B, 13B, and 30B parameter models. (paper)
![Twitter avatar for @XuCanwen](https://substackcdn.com/image/twitter_name/w_96/XuCanwen.jpg)
![Image](https://substackcdn.com/image/fetch/w_600,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fpbs.substack.com%2Fmedia%2FFs1eoitXsAEZzba.jpg)
6). MACHIAVELLI - a new benchmark of 134 text-based Choose-Your-Own-Adventure games to evaluate the capabilities and unethical behaviors of LLMs. (paper)
![Twitter avatar for @DanHendrycks](https://substackcdn.com/image/twitter_name/w_96/DanHendrycks.jpg)
![Image](https://substackcdn.com/image/fetch/w_600,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fpbs.substack.com%2Fmedia%2FFtH62-UacAItwf1.jpg)
7). Better Language Models of Code through Self-Improvement - generates pseudo data from knowledge gained through pre-training and fine-tuning; adds the data to the training dataset for the next step; results show that different frameworks can be improved in performance using code-related generation tasks. (paper)
![Twitter avatar for @johnjnay](https://substackcdn.com/image/twitter_name/w_96/johnjnay.jpg)
![Image](https://substackcdn.com/image/fetch/w_600,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fpbs.substack.com%2Fmedia%2FFs_lEnpX0AA9ujs.jpg)
8). Summary of ChatGPT/GPT-4 Research - an overview of applications of ChatGPT and GPT-4; the analysis is done on 194 relevant papers and discusses capabilities, limitations, concerns, and more. (paper)
![Twitter avatar for @omarsar0](https://substackcdn.com/image/twitter_name/w_96/omarsar0.jpg)
![Image](https://substackcdn.com/image/fetch/w_600,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fpbs.substack.com%2Fmedia%2FFs9MstmWwAMVAUp.png)
9). Pythia - a suite for analyzing LLMs across training and scaling; includes 16 LLMs trained on public data and ranging in size from 70M to 12B parameters. (paper)
![Twitter avatar for @BlancheMinerva](https://substackcdn.com/image/twitter_name/w_96/BlancheMinerva.jpg)
10). SegGPT - unifies segmentation tasks into a generalist model through an in-context framework that supports different kinds of data. (paper)