AI Tidbits April Round-Up
Open-source LLMs surge with Dolly 2.0, HuggingChat, WebLLM, Large Multimodal Models thrive, AutoGPT agents revolutionize industries, and Meta's Segment Anything debuts
Welcome to the April edition of AI Tidbits, your monthly dose of awe-inspiring developments in the dynamic world of artificial intelligence. This month was a veritable treasure trove of progress, with the open-source large language models space witnessing the emergence of groundbreaking models like Dolly 2.0, HuggingChat, and WebLLM.
But that's just the tip of the iceberg! Large Multimodal Models have also taken center stage, unlocking unparalleled potential in the fusion of AI capabilities. Moreover, brace yourselves for the AutoGPT revolution as Automated GPT agents, such as BabyAGI and AutoGPT, make their mark in diverse use cases, from an automated sales representative to a new frontend developer colleague.
This month's highlights also include Meta's game-changing Segment Anything, the first foundation model for image segmentation, cutting-edge prompting techniques like Self-Refine, and an insightful glimpse into the future with Sam Altman's note on the highly anticipated GPT-5.
Large Language Models (LLMs)
Researchers from UC Berkeley, CMU, Stanford, and UC San Diego unveil Vicuna-13B - an open-source alternative to GPT-4 which reportedly achieves 90% of ChatGPT’s quality
Databricks introduces Dolly 2.0 - the first open-source LLM for commercial use that was fine-tuned on human-generated instructions
Hugging Face introduces HuggingChat, an open source alternative to ChatGPT
Stability AI releases StableLM - their first-ever commercially available LLMs with 3B and 7B parameters models and more powerful ones to follow
LAION releases Open Assistant Conversational dataset, containing 600k human-generated data points covering a wide range of topics and writing styles in 35 languages
Together announces RedPajama - creating a leading, fully open-source LLMs, beginning with the release of a 1.2 trillion token dataset
CMU and OctoML release WebLLM, bringing instruction fine-tuned LLMs to the browser
Bloomberg announces BloombergGPT - an LLM trained on financial data to support financial NLP tasks
UCSD and Microsoft open source Baize - a chat model trained on100k dialogs generated by letting ChatGPT chat with itself
UC Berkeley publishes Koala, a dialogue model for research purposes that was trained by fine-tuning Meta’s LLaMA
Keep reading with a 7-day free trial
Subscribe to AI Tidbits to keep reading this post and get 7 days of free access to the full post archives.