Ability to set maxTokens
when calling prompt()
#36
Labels
ecosystem parity
A feature that other popular language model APIs offer
enhancement
New feature or request
I'd like to be able to constrain the output to N tokens for my use case (tab autocompletion). Otherwise the model generates strings that are too long to be useful. It also takes more time.
The text was updated successfully, but these errors were encountered: