Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

BLOOM Announcement #417

Merged
merged 1 commit into from
Jul 12, 2022
Merged
Show file tree
Hide file tree
Changes from all commits
Commits
File filter

Filter by extension

Filter by extension

Conversations
Failed to load comments.
Loading
Jump to
Jump to file
Failed to load files.
Loading
Diff view
Diff view
16 changes: 13 additions & 3 deletions _blog.yml
Original file line number Diff line number Diff line change
Expand Up @@ -860,15 +860,15 @@
tags:
- graphcore
- partnerships

- local: deep-rl-dqn
title: "Deep Q-Learning with Atari"
author: ThomasSimonini
thumbnail: /blog/assets/78_deep_rl_dqn/thumbnail.gif
date: June 7, 2022
tags:
- rl

- local: annotated-diffusion
title: "The Annotated Diffusion Model"
author: nielsrogge
Expand Down Expand Up @@ -977,4 +977,14 @@
tags:
- sentiment-analysis
- nlp
- guide
- guide

- local: bloom
title: "Introducing The World's Largest Open Multilingual Language Model: BLOOM"
author: BigScience
thumbnail: /blog/assets/86_bloom/thumbnail.png
date: July 12, 2022
tags:
- open-source-collab
- community
- research
Binary file added assets/86_bloom/bloom-demo.png
Loading
Sorry, something went wrong. Reload?
Sorry, we cannot display this file.
Sorry, this file is invalid so it cannot be displayed.
Binary file added assets/86_bloom/thumbnail-2.png
Loading
Sorry, something went wrong. Reload?
Sorry, we cannot display this file.
Sorry, this file is invalid so it cannot be displayed.
Binary file added assets/86_bloom/thumbnail.png
Loading
Sorry, something went wrong. Reload?
Sorry, we cannot display this file.
Sorry, this file is invalid so it cannot be displayed.
47 changes: 47 additions & 0 deletions bloom.md
Original file line number Diff line number Diff line change
@@ -0,0 +1,47 @@
---
title: "Introducing The World's Largest Open Multilingual Language Model: BLOOM"
thumbnail: /blog/assets/86_bloom/thumbnail.png
---

<html>
<head>
<link rel=“canonical” href=“http://bigscience.huggingface.co/blog/bloom” />
<style>
.grandmahugs {
display: block;
margin-left: auto;
margin-right: auto;
width: 50%;
}
</style>
<h1>🌸 Introducing The World's Largest Open Multilingual Language Model: BLOOM 🌸</h1>

<div class="blog-metadata">
<small>Published July 12, 2022.</small>
</div>

<div class="author-card">
<a href="/bigscience">
<img class="avatar avatar-user" src="https://aeiljuispo.cloudimg.io/v7/https://s3.amazonaws.com/moonup/production/uploads/1634806038075-5df7e9e5da6d0311fd3d53f9.png?w=200&h=200&f=face" title="Gravatar">
<div class="bfc">
<code>bigscience</code>
<span class="fullname">BigScience</span>
</div>
</a>
</div>
</head>
<body>
<a href="https://huggingface.co/bigscience/bloom"><img style="middle" width="950" src="/blog/assets/86_bloom/thumbnail-2.png"></a>

Large language models (LLMs) have made a significant impact on AI research. These powerful, general models can take on a wide variety of new language tasks from a user’s instructions. However, academia, nonprofits and smaller companies' research labs find it difficult to create, study, or even use LLMs as only a few industrial labs with the necessary resources and exclusive rights can fully access them. Today, we release [BLOOM](https://huggingface.co/bigscience/bloom), the first multilingual LLM trained in complete transparency, to change this status quo — the result of the largest collaboration of AI researchers ever involved in a single research project.

With its 176 billion parameters, BLOOM is able to generate text in 46 natural languages and 13 programming languages. For almost all of them, such as Spanish, French and Arabic, BLOOM will be the first language model with over 100B parameters ever created. This is the culmination of a year of work involving over 1000 researchers from 70+ countries and 250+ institutions, leading to a final run of 117 days (March 11 - July 6) training the BLOOM model on the [Jean Zay supercomputer](http://www.idris.fr/eng/info/missions-eng.html) in the south of Paris, France thanks to a compute grant worth an estimated €3M from French research agencies CNRS and GENCI.

Researchers can [now download, run and study BLOOM](https://huggingface.co/bigscience/bloom) to investigate the performance and behavior of recently developed large language models down to their deepest internal operations. More generally, any individual or institution who agrees to the terms of the model’s [Responsible AI License](https://bigscience.huggingface.co/blog/the-bigscience-rail-license) (developed during the BigScience project itself) can use and build upon the model on a local machine or on a cloud provider. In this spirit of collaboration and continuous improvement, we’re also releasing, for the first time, the intermediary checkpoints and optimizer states of the training. Don’t have 8 A100s to play with? An inference API, currently backed by Google’s TPU cloud and a FLAX version of the model, also allows quick tests, prototyping, and lower-scale use. You can already play with it on the Hugging Face Hub.

<img class="grandmahugs" style="center" padding="5px" width="550" src="/blog/assets/86_bloom/bloom-demo.png"></a>
Copy link
Contributor

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

Maybe use this image instead, which shows multiple examples in various languages? https://assets.website-files.com/6139f3cdcbbff3a68486761d/62cd189ca5613c83309a2d58_blog_examples-p-2000.png

Copy link
Contributor Author

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

Can't get that asset to look show up at a larger size on-page
Screen Shot 2022-07-12 at 11 02 16 AM
. Tried several things but it still shows up really small.


This is only the beginning. BLOOM’s capabilities will continue to improve as the workshop continues to experiment and tinker with the model. We’ve started work to make it instructable as our earlier effort T0++ was and are slated to add more languages, compress the model into a more usable version with the same level of performance, and use it as a starting point for more complex architectures… All of the experiments researchers and practitioners have always wanted to run, starting with the power of a 100+ billion parameter model, are now possible. BLOOM is the seed of a living family of models that we intend to grow, not just a one-and-done model, and we’re ready to support community efforts to expand it.

</body>
</html>