Yes, cached prompt tokens are discounted compared to uncached tokens for serverless models. The default discount is 50%, but the exact discount varies by model. Check the Model Library for model-specific cached and uncached input token pricing.Documentation Index
Fetch the complete documentation index at: https://fireworks.ai/docs/llms.txt
Use this file to discover all available pages before exploring further.