Home Community List of Groundbreaking and Open-Source Conversational AI Models within the Language Domain

List of Groundbreaking and Open-Source Conversational AI Models within the Language Domain

0
List of Groundbreaking and Open-Source Conversational AI Models within the Language Domain

Conversational AI refers to technology like a virtual agent or a chatbot that use large amounts of knowledge and natural language processing to mimic human interactions and recognize speech and text. Lately, the landscape of conversational AI has evolved drastically, especially with the launch of ChatGPT. Listed here are another open-source large language models (LLMs) which can be revolutionizing conversational AI.

  • Release date: February 24, 2023

LLaMa is a foundational LLM developed by Meta AI. It’s designed to be more versatile and responsible than other models. The discharge of LLaMA goals to democratize access to the research community and promote responsible AI practices.

LLaMa is on the market in several sizes, with the variety of parameters starting from 7B to 65B. Permission to the model’s access will probably be granted on a case-to-case basis to industry research laboratories, academic researchers, etc.

[Sponsored] 🔥 Construct your personal brand with Taplio  🚀 The first all-in-one AI-powered tool to grow on LinkedIn. Create higher LinkedIn content 10x faster, schedule, analyze your stats & engage. Try it without spending a dime!
  • Release date: March 8, 2023

Open Assistant is a project developed by LAION-AI to supply everyone with an awesome chat-based large language model. Through extensive training in vast amounts of text and code, it has acquired the flexibility to perform various tasks, including responding to queries, generating text, translating languages, and producing creative content. 

Although OpenAssistant remains to be within the developmental stage, it has already acquired several skills, akin to interacting with external systems like Google Search to collect information. Moreover, it’s an open-source initiative, meaning that anyone can contribute to its progress.

  • Release date: March 8, 2023

Dolly is an instruction-following LLM developed by Databricks. It’s trained on the Databricks machine-learning platform licensed for industrial use. Dolly is powered by the Pythia 12B model and has been trained on a big selection of instruction/response records totaling roughly 15k in number. Although not cutting-edge, Dolly’s performance in following instructions is impressively high-quality.

  • Release date: March 13, 2023

Alpaca is a small instruction-following model developed by . It is predicated on Meta’s LLaMa (7B parameters) model. It’s designed to perform well on quite a few instruction-following tasks while being easy and low cost to breed at the identical time. 

Even though it resembles OpenAI’s text-davinci-003 model, it’s significantly cheaper (<$600) to provide. The model is open-source and has been trained on a dataset of 52,000 demonstrations of instruction-following.

Vicuna has been developed by a team of UC Berkeley, CMU, Stanford, and UC San Diego. It’s a chatbot that has been trained by fine-tuning the LLaMa model on conversations shared by users and picked up from ShareGPT. 

Based on the transformer architecture, Vicuna is an auto-regressive language model and offers natural and fascinating conversation capabilities. With 13B parameters, it produces more detailed and well-structured answers than Alpaca, and its quality is comparable to that of ChatGPT.

  • Release date: April 3, 2023

The Berkeley Artificial Intelligence Research Lab (BAIR) has developed Koala, which is a dialogue model based on the . It is meant to be safer and more easily interpretable than other LLMs. Koala has been fine-tuned on freely available interaction data, specializing in data that features interaction with highly capable closed-source models. 

Koala is helpful for studying language model safety and bias and understanding dialogue language models’ inner workings. Moreover, Koala is an open-source alternative to ChatGPT that features EasyLM, a framework for training and fine-tuning LLMs.

Eleuther AI has created a set of autoregressive language models called Pythia, that are designed to support scientific research. Pythia consists of 16 different models starting from 70M to 12B parameters. All models are trained using the identical data and architecture, allowing for comparisons and exploring how they evolve with scaling.

  • Release date: April 5, 2023

Together has developed OpenChatKit, an open-source chatbot development framework that goals to simplify and streamline the means of constructing conversational AI applications. The chatbot is designed for conversation and instruction and excels in summarizing, generating tables, classification, and dialog. 

With OpenChatKit, developers can access a sturdy, open-source foundation to create specialized and general-purpose chatbots for various applications. The framework is built on the GPT-4 architecture and is on the market in three different model sizes – 3B, 6B, and 12B parameters – to accommodate diverse computational resources and application requirements.

  • Release date: April 13, 2023

RedPajama is a project created by a team from Together, Ontocord.ai, ETH DS3Lab, Stanford CRFM, Hazy Research, and MILA Québec AI Institute. Their goal is to develop top-notch open-source models, starting with reproducing the LLaMA training dataset that comprises greater than 1.2 trillion tokens.

This project goals to create a totally open, replicable, and cutting-edge language model with three essential elements: pre-training data, base models, and instruction-tuning data and models. The dataset is currently accessible through Hugging Face, and users have the choice to duplicate the outcomes using Apache 2.0 scripts, which can be found on GitHub.

  • Release date: April 19, 2023

StableLM is an open-source language model developed by Stability AI. The model is trained on an experimental dataset 3 times larger than The Pile dataset and is effective in conversational and coding tasks despite its small size. The model is available in 3B and 7B parameters, with larger models still to return.

StableLM can generate each text and code, making it suitable for various downstream applications. Stability AI can also be making available a series of fine-tuned research models through instruction, utilizing a mixture of 5 up-to-date open-source datasets specifically designed for conversational agents. These fine-tuned models are exclusively for research and can be found under a non-commercial CC BY-NC-SA 4.0 license.


Try the Paper and GitHub link. Don’t forget to hitch our 20k+ ML SubRedditDiscord Channel, and Email Newsletter, where we share the most recent AI research news, cool AI projects, and more. If you’ve got any questions regarding the above article or if we missed anything, be at liberty to email us at Asif@marktechpost.com

🚀 Check Out 100’s AI Tools in AI Tools Club


References:

https://www.ibm.com/topics/conversational-ai
https://ai.facebook.com/blog/large-language-model-llama-meta-ai/
https://crfm.stanford.edu/2023/03/13/alpaca.html
https://vicuna.lmsys.org/
https://bair.berkeley.edu/blog/2023/04/03/koala/
https://www.together.xyz/blog/redpajama
https://arxiv.org/pdf/2304.01373.pdf
https://openchatkit.net/
https://github.com/databrickslabs/dolly


Arham Islam

” data-medium-file=”https://www.marktechpost.com/wp-content/uploads/2022/10/Screen-Shot-2022-10-03-at-10.48.33-PM-293×300.png” data-large-file=”https://www.marktechpost.com/wp-content/uploads/2022/10/Screen-Shot-2022-10-03-at-10.48.33-PM.png”>

I’m a Civil Engineering Graduate (2022) from Jamia Millia Islamia, Latest Delhi, and I even have a keen interest in Data Science, especially Neural Networks and their application in various areas.


🔥 StoryBird.ai just dropped some amazing features. Generate an illustrated story from a prompt. Test it out here. (Sponsored)

LEAVE A REPLY

Please enter your comment!
Please enter your name here