Aditya

GPT-J - Open-source cousin of GPT-3, everyone can use it

by
GPT⁠⁠-⁠J-6B, a 6 billion parameter model trained on the Pile, is now available for use with our new codebase, Mesh Transformer JAX.

Add a comment

Replies

Best
Nassim Abd
Great, we were waiting for this
Aditya
@nassc for so long! Finally the wait is over :)
Dillon Peterson
WOAH!! Way to go guys! Thank y'all for putting this together, really amazing.
Aditya
@dillon_peterson all credit goes to the wonderful open-source community :)
Fateh BENMERZOUG, Ph.D
This just blew up the door of textual content generation, awesome!
Aditya
@fateh_benmerzoug IKR! This is literally giving super-powers to the Makers 🚀
Kelvin Zhao
Will an api of this be created?
Mustafa Al-Adhami
Awesome
Raphi Mbinga
🔥🔥🔥
Nick Dahlhoff
Looks really interesting
Pascal Weinberger
the world needs this :)
Patrick Hamelin
*GPT-J is just as good as GPT-3.* It is more efficient, but with more quirks. In our JPRED scores, it did better with simple TCS tasks, but lost with the more complex tasks. By removing the Jordan Algorithm: Our next proposed change to a probability model is removing the Jordan Algorithm. The Jordan Algorithm is a special procedure used for simple TCS tasks that allows for fast analysis of different sequence pairs, as well as being able to easily analyze simple n-gram (aka word) models.It is more efficient, but with more quirks. In our JPRED scores, it did better with simple TCS tasks, but lost with the more complex tasks. ...
Maxim Ignatev
Hey hope this is still relevant I find gpt-j quite alright in the generation but it provides silly results when it does summaries. are there any experts here that could help on how i can maybe train here to provide tl;dr's