DEV Community

Cover image for OPT: Open Pre-trained Transformer Language Models
Paperium
Paperium

Posted on • Originally published at paperium.net

OPT: Open Pre-trained Transformer Language Models

Open Pre-trained Transformers: Big language models, shared for research

Big language models usually cost a fortune and are locked behind services.
Now there is a set of models anyone studying language tech can use, from about 125 million to 175 billion pieces inside them.
They were made so researchers can look under the hood, compare things, and learn without buying a whole data center.
One of the largest, OPT-175B, matches the performance of popular closed models, while needing only about 1/7th the carbon to build.
We are also sharing our logbook of the problems we hit, and the code so people can try the models themselves.
This means more people can test ideas, find bugs, and make improvements, not just big companies.
It’s meant to be shared carefully, with safety in mind, so it helps study how these systems behave in the real world.
Expect easier experiments, new discoveries, and more open work on language tech, all without the giant price tag most models demand.

Read article comprehensive review in Paperium.net:
OPT: Open Pre-trained Transformer Language Models

🤖 This analysis and review was primarily generated and structured by an AI . The content is provided for informational and quick-review purposes.

Top comments (0)