Average model response time for the past few minutes.
Depends on model size, server load, and request size.
The average prompt length for this model is -- tokens.
OutputLimittokens
OUTPUT TOKEN LIMITS
Maximum Response Length for this model.
This model averages -- output tokens per reply.