Skip to main content
About the Resource Center

Understand Genesys Agent Copilot AI models, LLM input and AWS region mapping

Genesys uses a combination of Genesys native and AWS Bedrock models, depending on the feature, use case, and dialect.

Note: Large language model (LLM) tokens are not the same as Genesys Cloud AI Experience tokens. For more information about Genesys Cloud AI Experience tokens and pricing, see Genesys Cloud AI Experience tokens metering and pricing.

Genesys Agent Copilot AI models and LLM input

Genesys Agent Copilot uses a combination of Genesys native and AWS Bedrock models, depending on the feature, use case, and dialect.

Bedrock region mapping

The following table lists the AWS region mappings where Genesys Cloud uses Bedrock models.

Tokens overview

Tokens represent segments of text, like parts of words, spaces, or punctuation. The API processes input and breaks that input down into tokens, which do not always align with whole words. Some key approximations include:

  • One token equals approximately four characters in English
  • One token equals approximately ¾ of a word
  • 100 tokens equal approximately 75 words

 Or:

  • 1–2 sentences equal approximately 30 tokens
  • One paragraph equals approximately 100 tokens
  • 1,500 words equal approximately 2048 tokens

For example, a popular saying like “Carpe diem” might have 10 tokens. A company’s vision statement could be about 500 tokens. A classic poem, such as “The Road Not Taken” might contain around 1,700 tokens.

Language differences

Tokenization varies by language. For instance, “Bonjour tout le monde,” which means “Hello everyone” in French, contains five tokens, despite being just 19 characters. Non-English texts may require more tokens, which impact API costs.

How tokens generate

The API converts words into tokens based on their context, processes them, and then translates the tokens back into readable text. The same word may split into different tokens, depending on its placement. For instance, the word “blue” could have a different token if it appears with a trailing space or starts with a capital letter.

Examples:

  • The token for “blue” with a trailing space might be “4321.”
  • The token for “Blue” (capitalized with a space) could be “4521.”
  • The token for “Blue” (capitalized at the start) might be “5000.”

Common tokens have lower values. Punctuation, like commas, tends to have a consistent token ID due to its frequent use. Words may have different token IDs based on their case or position in a sentence.