Open-source artificial intelligence
Open-source artificial intelligence is the application of open source practices to the development of artificial intelligence resources.
Many open-source artificial intelligence products are variations of other existing tools and technology which major companies have shared as open-source software.[1]
Companies often developed closed products in an attempt to keep a competitive advantage in the marketplace.[2] A journalist for Wired explored the idea that open-source AI tools have a development advantage over closed products, and could overtake them in the marketplace.[2]
Popular open-source artificial intelligence project categories include large language models, Machine translation tools, and chatbots.[3]
For software developers to produce open-source artificial intelligence resources, they must trust the various other open-source software components they use in its development.[4]
Large Language Models
LLaMA
LLaMA (Large Language Model Meta AI) is a family of large language models (LLMs), released by Meta AI starting in February 2023. [5]
Model | Developer | Parameter Count | Context Window | Licensing |
---|---|---|---|---|
LLaMA[5] | Meta AI | 7B, 13B, 33B, 65B | 2048 | —— |
LLaMA 2[6][7] | Meta AI | 7B, 13B, 70B | 4k | Custom Meta License |
Mistral 7B[8] | Mistral AI | 7 billion | 8k[9] | Apache 2.0 |
GPT-J[10] | EleutherAI | 6 billion | 2048 | Apache 2.0 |
Pythia[11] | EluetherAI | 70 million - 12 billion | —— | Apache 2.0 (Pythia-6.9B only)[12] |
References
- Heaven, Will Douglas (May 12, 2023). "The open-source AI boom is built on Big Tech's handouts. How long will it last?". MIT Technology Review.
- Solaiman, Irene (May 24, 2023). "Generative AI Systems Aren't Just Open or Closed Source". Wired.
- Castelvecchi, Davide (29 June 2023). "Open-source AI chatbots are booming — what does this mean for researchers?". Nature. 618 (7967): 891–892. doi:10.1038/d41586-023-01970-6.
- Thummadi, Babu Veeresh (2021). Artificial Intelligence (AI) Capabilities, Trust and Open Source Software Team Performance. pp. 629–640. doi:10.1007/978-3-030-85447-8_52. ISBN 978-3-030-85446-1.
{{cite book}}
:|journal=
ignored (help) - "Introducing LLaMA: A foundational, 65-billion-parameter language model". 2023-09-11. Archived from the original on 2023-09-11. Retrieved 2023-10-03.
- "meta-llama/Llama-2-70b-chat-hf · Hugging Face". huggingface.co. Retrieved 2023-10-03.
- "Llama 2 - Meta AI". ai.meta.com. Retrieved 2023-10-03.
- "mistralai/Mistral-7B-v0.1 · Hugging Face". huggingface.co. Retrieved 2023-10-03.
- AI, Mistral (2023-09-27). "Mistral 7B". mistral.ai. Retrieved 2023-10-03.
- "EleutherAI/gpt-j-6b · Hugging Face". huggingface.co. 2023-05-03. Retrieved 2023-10-03.
- Biderman, Stella; Schoelkopf, Hailey; Anthony, Quentin; Bradley, Herbie; O'Brien, Kyle; Hallahan, Eric; Mohammad Aflah Khan; Purohit, Shivanshu; USVSN Sai Prashanth; Raff, Edward; Skowron, Aviya; Sutawika, Lintang; Oskar van der Wal (2023-10-03). "[2304.01373] Pythia: A Suite for Analyzing Large Language Models Across Training and Scaling". arXiv:2304.01373 [cs.CL].
- "EleutherAI/pythia-6.9b · Hugging Face". huggingface.co. 2023-05-03. Retrieved 2023-10-03.