Skip to content
PromptxAI Generative AI Playbook - Accelerate your generative AI value creation and impact | Product Hunt

Generative Pre-trained Transformer 3 (GPT-3)

Autoregressive language model released in 2020 that uses deep learning to produce human-like text

Publisher License Version Release
OpenAI Microsoft has exclusive licensing of GPT-3 for Microsoft's products and services GPT-3.5 May 28, 2020

Model Summary

GPT-3 is a deep learning language model released in 2020 that uses generative pre-training to produce human-like text. It is the third-generation language prediction model in the GPT series, successor to GPT-2 created by OpenAI. GPT-3 has a capacity of 175 billion parameters and is capable of performing zero-shot, few-shot and one-shot learning. It can generate text that is difficult to distinguish from human-written text, and has potential applications in both beneficial and harmful applications. Microsoft has exclusive licensing of GPT-3, while OpenAI provides a public-facing API. There are concerns about the environmental impact of training and storing the model, as well as potential issues with academic integrity.

Model Resources

📄 Research Paper | 🐱 GitHub | 🕸️ Wikipedia

Info

This model card was generated using PromptxAI API querying recent web content sources with large language model generations. As of Feb 2023 it is not possible to query models like GPT-3 (via applications like ChatGPT) on the latest web content. This is because the model is trained on a static dataset and is not updated with new web content. PromptxAI API solves this problem by chaining recent web content sources with large language model outputs. This allows you to query models like GPT-3 on latest web content.

Model Details

Size: 175 billion parameters, requiring 800GB to store

Use Cases: Text generation, summarizing texts, answering questions, coding in CSS, JSX, and Python

Training corpus: Common Crawl, WebText2, Books1, Books2, Wikipedia

Training method: Generative pre-training

Evaluation method: Human evaluators

Compute: 800GB

Features: Zero-shot, few-shot and one-shot learning, edit and insert capabilities

Limitations: Potential to perpetuate fake news, environmental impact of training and storing the models, potential for misuse

Strengths: Eerily good at writing amazingly coherent text, improved language understanding performances in natural language processing (NLP)