The latter activity is probably the least fun part of programming (and the highest barrier to entry), and it’s where OpenAI Codex excels most. Once a programmer knows what to build, the act of writing code can be thought of as (1) breaking a problem down into simpler problems, and (2) mapping those simple problems to existing code (libraries, APIs, or functions) that already exist. OpenAI Codex empowers computers to better understand people’s intent, which can empower everyone to do more with computers. OpenAI Codex has much of the natural language understanding of GPT-3, but it produces working code-meaning you can issue commands in English to any piece of software with an API. GPT-3’s main skill is generating natural language in response to a natural language prompt, meaning the only way it affects the world is through the mind of the reader. It has a memory of 14KB for Python code, compared to GPT-3 which has only 4KB-so it can take into account over 3x as much contextual information while performing any task.
OpenAI Codex is most capable in Python, but it is also proficient in over a dozen languages including JavaScript, Go, Perl, PHP, Ruby, Swift and TypeScript, and even Shell.
OpenAI Codex is a descendant of GPT-3 its training data contains both natural language and billions of lines of source code from publicly available sources, including code in public GitHub repositories.