Search results
This is the smallest version of GPT-2, with 124M parameters. Related Models: GPT-Large, GPT-Medium and GPT-XL. Intended uses & limitations You can use the raw model for text generation or fine-tune it to a downstream task. See the model hub to look for fine-tuned versions on a task that interests you. How to use
GPT-2 Output Detector Demo. This is an online demo of the GPT-2 output detector model, based on the 🤗/Transformers implementation of RoBERTa. Enter some text in the text box; the predicted probabilities will be displayed below. The results start to get reliable after around 50 tokens.
5 lis 2019 · As the final model release of GPT-2’s staged release, we’re releasing the largest version (1.5B parameters) of GPT-2 along with code and model weights to facilitate detection of outputs of GPT-2 models.
This repository is meant to be a starting point for researchers and engineers to experiment with GPT-2. For basic information, see our model card.
ChatGPT w języku polskim już dostępny! Korzystaj z sieci neuronowej OpenAI za darmo i bez rejestracji. ChatGPT to chatbot wyposażony w sztuczną inteligencję. Potrafi generować teksty o dowolnej złożoności i tematyce, komponować eseje i raporty, napisać zabawną historię lub zasugerować pomysły na nowe projekty. Spróbuj ChatGPT ...
GPT-2 is a large transformer-based language model with 1.5 billion parameters, trained on a dataset[1] of 8 million web pages. GPT-2 is trained with a simple objective: predict the next word, given all of the previous words within some text.
This web app, built by the Hugging Face team, is the official demo of the 🤗/transformers repository's text generation capabilities. Star 134,103. Models. 🦄 GPT-2. The almighty king of text generation, GPT-2 comes in four available sizes, only three of which have been publicly made available.