Jump to content

BLOOM (language model)

From Wikipedia, the free encyclopedia

This is an old revision of this page, as edited by 2a00:a040:19e:a325:7069:961:6d48:501e (talk) at 15:51, 12 November 2022. The present address (URL) is a permanent link to this revision, which may differ significantly from the current revision.

BigScience Large Open-science Open-access Multilingual Language Model (BLOOM[1]) is a transformer-based language model. It was created by over 1000 AI researchers to provide a free large language model for everyone who wants to try. Trained on around 176 billion parameters over March through July 2022, it is considered an alternative to OpenAI's GPT-3 trained on 176 billion parameters. BLOOM uses a decoder-only transformer model architecture modified from Megatron-LM GPT-2.

The BLOOM project[2] was started by a co-founder of Hugging Face. Six main groups of people were involved, including HuggingFace's BigScience team, the Microsoft DeepSpeed team, the NVIDIA Megatro-LM team, the IDRIS/GENCI team, the PyTorch team, and the volunteers in the BigScience Engineering workgroup.

References

  1. ^ "BigScience Large Open-science Open-access Multilingual Language Model". Retrieved 1 October 2022.
  2. ^ "The Technology Behind BLOOM Training". Retrieved 1 October 2022.