GPT-2 (nonfiction): Difference between revisions

From Gnomon Chronicles
Jump to navigation Jump to search
(Created page with "'''GPT-2''' is ... * [https://blog.floydhub.com/gpt2/ How to Build OpenAI's GPT-2: "The AI That's Too Dangerous to Release"]")
 
No edit summary
Line 1: Line 1:
'''GPT-2''' is ...
'''GPT-2''' ('''Generative Pretrained Transformer 2''') is a language model that was trained on 40GB of text scraped from websites that Reddit linked to and that had a Karma score of at least two.
 
The developers at OpenAI describe GPT-2 as "a large-scale unsupervised language model which generates coherent paragraphs of text, achieves state-of-the-art performance on many language modeling benchmarks, and performs rudimentary reading comprehension, machine translation, question answering, and summarization—all without task-specific training."
 
Because the model is probabilistic, it returns a different response every time you enter the same input.




* [https://blog.floydhub.com/gpt2/ How to Build OpenAI's GPT-2: "The AI That's Too Dangerous to Release"]
* [https://blog.floydhub.com/gpt2/ How to Build OpenAI's GPT-2: "The AI That's Too Dangerous to Release"]

Revision as of 08:40, 6 July 2019

GPT-2 (Generative Pretrained Transformer 2) is a language model that was trained on 40GB of text scraped from websites that Reddit linked to and that had a Karma score of at least two.

The developers at OpenAI describe GPT-2 as "a large-scale unsupervised language model which generates coherent paragraphs of text, achieves state-of-the-art performance on many language modeling benchmarks, and performs rudimentary reading comprehension, machine translation, question answering, and summarization—all without task-specific training."

Because the model is probabilistic, it returns a different response every time you enter the same input.