Fine-tuned (LoRA) models require a dedicated deployment to serve. Here’s what you need to know: What you pay for:Documentation Index
Fetch the complete documentation index at: https://fireworks.ai/docs/llms.txt
Use this file to discover all available pages before exploring further.
- Deployment costs on a per-GPU-second basis for hosting the model
- The fine-tuning process itself, if applicable
- Live-merge deployment: Deploy your LoRA model with weights merged into the base model for optimal performance
- Multi-LoRA deployment: Deploy up to 100 LoRA models as addons on a single base model deployment