Average time until the first token is generated by this model.
Depends on model size, server load, and prompt size.
The average prompt length for this model is -- tokens.
OutputLimittokens
OUTPUT TOKEN LIMITS
Some models have limited response lengths for each reply.
This model averages -- output tokens per reply.