GPT-3’s main skill is generating natural language in response to a natural language prompt, meaning the only way it affects the world is through the mind of the reader. It has a memory of 14KB for Python code, compared to GPT-3 which has only 4KB-so it can take into account over 3x as much contextual information while performing any task.
OpenAI Codex is most capable in Python, but it is also proficient in over a dozen languages including JavaScript, Go, Perl, PHP, Ruby, Swift and TypeScript, and even Shell. OpenAI Codex is a descendant of GPT-3 its training data contains both natural language and billions of lines of source code from publicly available sources, including code in public GitHub repositories.