Tokens in OpenAI

Tokens in OpenAI

May 21, 2024 | seedling, permanent

tags :

Summary #

A single English word can take up anywhere between 1 to 3 tokens. URL

According to OpenAI, 1,000 tokens roughly corresponds to 750 words of text. However, this figure can fluctuate depending on the language and text complexity.

words = 3/2(tokens)

ref 2

OCR of Images #

2024-05-16_10-35-18_screenshot.png #

Number of tokens Examples Words less than four characters long 1 Air, third, this, text Longer or less common words 1 to 3 Commotion, airline, understand Punctuation and numbers 1 ,1/39 Space 1 Emoji 1 to 3

2024-05-16_10-37-01_screenshot.png #

GPT-3 Codex You miss 100% of the shots you don't take. / Clear Show example Tokens 12 Characters 42 You miss 100% of the shots you don't take. TEXT TOKEN IDS

OCR of Images #

2024-05-16_10-35-18_screenshot.png #

Number of tokens Examples Words less than four characters long 1 Air, third, this, text Longer or less common words 1 to 3 Commotion, airline, understand Punctuation and numbers 1 ,1/39 Space 1 Emoji 1 to 3

2024-05-16_10-37-01_screenshot.png #

GPT-3 Codex You miss 100% of the shots you don't take. / Clear Show example Tokens 12 Characters 42 You miss 100% of the shots you don't take. TEXT TOKEN IDS


No notes link to this note

Go to random page

Previous Next