SentenceTransformers embedding tool has a maximum context length of 256 tokens, not characters, right?

At 4:00 in the video for you say that the maximum embedding context window length is 256 characters, but later on you clarify that it is 256 tokens, which I assume is correct. Otherwise picking a 1000-character default for splitting would make no sense.

Hi nealmcb,

Pretty sure it’s tokens.