"Some"
4366
token
"Some words are encoded as one token."
[4366, 2456, 389, 30240, 355, 530, 11241, 13]
tokenise
tokenise("Once upon a")
token(" time")
model
tokenise("Once upon a" + " time")
token(",")
model
Run for n cycles
Run until the last n tokens match one of the stop sequences
tokenise("Once upon a")
token("time")
model
tokenise("Once upon a")
(token(" time"), logprob("96.45%"))
model
tokenise("Once upon a")
[(token(" time"), logprob("96.45%")),
(token(" Time"), logprob("0.67%")),
(token(" midnight"), logprob("0.31%")),
...]
model
tokenise("Once upon a" + " midnight")
token("dreary")
model
The likelihood that a lower probability token will have its relative probability increased.
Only consider the top n most likely tokens, having a cumulative probability of p.
Generate m sequences of n tokens, then select the sequence of tokens with the highest probability.
https://blog.scottlogic.com/cprice/
https://www.youtube.com/watch?v=rURRYI66E54