full-sized GPT-2 model, called 1558M. Before November 5, OpenAI (An Artifical Intelligance Startup) had only released three smaller, less coherent versions of the model.
While GPT-2 was only trained to predict the next word in a text, it surprisingly learned basic competence in some tasks like...
Published 03/23/20