Skip to content

Default Tokenization for all-MiniLM-L6-v2 (L6) #207

Answered by drazvan
spehl-max asked this question in Q&A
Discussion options

You must be logged in to vote

Good questions @spehl-max!

  1. When computing the embedding, we send the text as is, so it will be truncated automatically. This could indeed be a problem, thanks for pointing this. The user and bot messages defined in a Colang config are not typically that long and flows are indexed line by line (https://github.com/NVIDIA/NeMo-Guardrails/blob/develop/nemoguardrails/actions/llm/generation.py#L179). So this is very unlikely to happen. But for the input coming from the user, this could be the case.

  2. The embeddings are computed when the configuration is initialized (https://github.com/NVIDIA/NeMo-Guardrails/blob/develop/nemoguardrails/actions/llm/generation.py#L105). In the prompt, typically …

Replies: 1 comment 1 reply

Comment options

You must be logged in to vote
1 reply
@spehl-max
Comment options

Answer selected by spehl-max
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Category
Q&A
Labels
None yet
2 participants