Movatterモバイル変換


[0]ホーム

URL:


Jump to content
WikipediaThe Free Encyclopedia
Search

BLOOM (language model)

From Wikipedia, the free encyclopedia
Multilingual open-access large language model
This article has multiple issues. Please helpimprove it or discuss these issues on thetalk page.(Learn how and when to remove these messages)
This articlerelies excessively onreferences toprimary sources. Please improve this article by addingsecondary or tertiary sources.
Find sources: "BLOOM" language model – news ·newspapers ·books ·scholar ·JSTOR
(October 2022) (Learn how and when to remove this message)
This articlecontainspromotional content. Please helpimprove it by removingpromotional language and inappropriateexternal links, and by adding encyclopedic text written from aneutral point of view.(October 2022) (Learn how and when to remove this message)
(Learn how and when to remove this message)
BLOOM
Original authorBigScience research workshop
Initial releaseJuly 12, 2022; 3 years ago (2022-07-12)
Repositoryhuggingface.co/bigscience/bloom
Written inPython
LicenseBigScience Responsible AI License (RAIL) v1.0
Websitebigscience.huggingface.co

TheBigScience Large Open-science Open-access Multilingual Language Model (BLOOM) is an open-accesslarge language model (LLM).[1] It was created by a volunteer-driven research effort to provide a transparently-created alternative to proprietary AI models.[2]

With 176 billion parameters, BLOOM is atransformer-basedautoregressive model designed to generate text in 46 natural languages and 13 programming languages. The model, source code, and the data used to train it are all distributed under free licences, allowing for public research and use.[3][4]

Development

[edit]

BLOOM is the main outcome of the BigScience initiative, a one-year-long research workshop that took place from May 2021 to May 2022.[5] The project was led byHuggingFace and involved several hundred volunteer researchers and engineers from academia and the private sector. The model was trained between March and July 2022 on the Jean Zay public supercomputer in France, managed byGENCI and IDRIS (CNRS).[6]

BLOOM's training corpus, named ROOTS, combines data extracted from the then-latest version of the web-based OSCAR corpus (38% of ROOTS) and newly collected data extracted from a manually selected and documented list of language data sources. In total, the model was trained on approximately 366 billion (1.6TB) tokens.[7][8]

External links

[edit]

References

[edit]
  1. ^"BigScience Large Open-science Open-access Multilingual Language Model". Retrieved2022-10-01.
  2. ^Heikkilä, Melissa (2022-07-12)."BLOOM: Inside the radical new project to democratize AI".MIT Technology Review. Retrieved2023-12-26.
  3. ^"The BigScience RAIL license". Retrieved2024-01-10.
  4. ^Le Scao T, Fan A, Akiki C, Pavlick E, Ilić S, Hesslow D, Castagné R, Luccioni A, Yvon F, Gallé M, Tow J, Rush AM, Biderman S, Webson A, Sasanka Ammanamanchi P, Wang T, Sagot B, Muennighoff N, Villanova del Moral A, Ruwase O, Bawden R, Bekman S, McMillan-Major A, Beltagy I, Nguyen H, Saulnier L, Tan S, Ortiz Suarez P, Sanh V, Laurençon H, Jernite Y, Launay J, Mitchell M, Raffel C, et al. (2022). "BLOOM: A 176B-Parameter Open-Access Multilingual Language Model".arXiv:2211.05100 [cs.CL].
  5. ^"BigScience". Retrieved2024-01-10.
  6. ^"Release of largest trained open-science multilingual language model ever".French National Centre for Scientific Research. 2022-07-12. Retrieved2023-12-26.
  7. ^Laurençon H, Saulnier L, Wang T, Akiki C, Villanova del Moral A, Le Scao T, Von Werra L, Mou C, González Ponferrada C, Nguyen H, Frohberg J, Šaško M, Lhoest Q, McMillan-Major A, Dupont G, Biderman S, Rogers A, Ben allal L, De Toni F, Pistilli G, Nguyen O, Nikpoor S, Masoud M, Colombo P, de la Rosa J, Villegas P, Thrush T, Longpre S, Nagel S, Weber L, Muñoz M, Zhu J, Van Strien D, Alyafeai Z, Almubarak K, Vu MC, Gonzalez-Dios I, Soroa A, Lo K, Dey M, Ortiz Suarez P, Gokaslan A, Bose S, Adelani D, Phan L, Tran H, Yu I, Pai S, Chim J, Lepercq V, Ilic S, Mitchell M, Luccioni S, Jernite Y (2022). "The BigScience ROOTS Corpus: A 1.6TB Composite Multilingual Dataset".arXiv:2303.03915 [cs.CL].
  8. ^Heikkilä, Melissa (2022-07-12)."BLOOM: Inside the radical new project to democratize AI".MIT Technology Review. Retrieved2023-12-26.
Concepts
Applications
Implementations
Audio–visual
Text
Decisional
People
Architectures
Stub icon

Thislarge language model-related article is astub. You can help Wikipedia byexpanding it.

Retrieved from "https://en.wikipedia.org/w/index.php?title=BLOOM_(language_model)&oldid=1303500099"
Categories:
Hidden categories:

[8]ページ先頭

©2009-2025 Movatter.jp