-
Notifications
You must be signed in to change notification settings - Fork 1.3k
[NA] [BE] Update model prices file #4546
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
[NA] [BE] Update model prices file #4546
Conversation
WalkthroughTwo JSON resource files containing AI model pricing and context window data are updated with numerous new models and variants from providers including Azure AI, Google Vertex AI, OpenAI, and others. Each model entry includes pricing details, supported features (function calling, vision, etc.), and context window specifications. No control flow or public API changes are introduced. Changes
Estimated code review effort🎯 2 (Simple) | ⏱️ ~12 minutes Pre-merge checks and finishing touches✅ Passed checks (3 passed)
✨ Finishing touches🧪 Generate unit tests (beta)
Comment |
|
@coderabbitai review |
✅ Actions performedReview triggered.
|
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
Actionable comments posted: 0
🧹 Nitpick comments (1)
apps/opik-frontend/src/data/model_prices_and_context_window.json (1)
1360-1373: Optional: Verify automated sync completed successfully.Since this is an automated update from LiteLLM's model pricing data, consider spot-checking a few entries against the source to confirm the sync process worked as expected.
📜 Review details
Configuration used: Organization UI
Review profile: CHILL
Plan: Pro
📒 Files selected for processing (2)
apps/opik-backend/src/main/resources/model_prices_and_context_window.jsonapps/opik-frontend/src/data/model_prices_and_context_window.json
🔇 Additional comments (19)
apps/opik-frontend/src/data/model_prices_and_context_window.json (11)
1360-1373: LGTM! New Azure AI model addition looks correct.The
azure_ai/gpt-oss-120bmodel entry is well-structured with comprehensive pricing, token limits, and feature support flags.
3724-3749: New Azure image generation models added.The
azure/gpt-image-1.5andazure/gpt-image-1.5-2025-12-16models have been added with consistent pricing structures. Both include text and image token pricing, which appears intentional for multimodal support.
15407-15434: LGTM! Gemini Veo video generation models added.Two variants of the Veo 3.1 model are properly configured with appropriate per-second output pricing and modality support for text-to-video generation.
18120-18120: Groq Llama Guard model addition.New safety model entry added to the catalog.
18140-18140: Vision support enabled for Groq Llama models.The
supports_visionflag has been added to two Groq Llama models, reflecting enhanced multimodal capabilities.Also applies to: 18153-18153
19393-19466: LGTM! Comprehensive MiniMax model catalog expansion.Multiple MiniMax models added across speech synthesis and chat modes:
- Speech models with HD and turbo variants at appropriate per-character pricing
- Chat models with extensive feature support including prompt caching, function calling, and large context windows (up to 1M tokens)
24398-24586: LGTM! Extensive Stability AI image editing models added.Significant expansion of image editing capabilities:
- 15 direct Stability AI models covering various editing modes (inpaint, outpaint, erase, sketch, style transfer, etc.)
- 13 corresponding Bedrock-hosted variants with consistent structure
- Appropriate pricing tiers based on model capabilities (fast: $0.002, creative: $0.06, etc.)
24998-24998: Response schema support enabled across Together AI models.The
supports_response_schemaflag has been consistently added to multiple Together AI models (Qwen, DeepSeek, Llama variants), indicating enhanced structured output capabilities across the provider's catalog.Also applies to: 25006-25006, 25018-25018, 25030-25030, 25053-25053, 25066-25066, 25078-25078, 25091-25091, 25111-25111, 25141-25141, 25151-25151, 25161-25161, 25217-25217, 25229-25229, 25241-25241, 25260-25260, 25297-25297, 25309-25309
25328-25363: LGTM! AWS Polly text-to-speech models added.Four AWS Polly variants are properly configured with tiered per-character pricing reflecting their capabilities (standard: $0.000004, neural: $0.000016, generative: $0.00003, long-form: $0.0001).
27908-27915: DeepSeek OCR model added to Vertex AI.The OCR model entry properly includes both per-page OCR costs ($0.0003) and per-token pricing for text processing, which is appropriate for document understanding use cases.
28084-28111: LGTM! Vertex AI Veo video generation models added.Two Veo 3.1 variants are configured for Vertex AI deployment with appropriate per-second output pricing, matching the Gemini-hosted variants added earlier in this update.
apps/opik-backend/src/main/resources/model_prices_and_context_window.json (8)
19393-19466: LGTM! Minimax model additions are comprehensive.The new Minimax models are well-structured with appropriate pricing models (character-based for speech, token-based for chat) and comprehensive capability flags including prompt caching support for the chat models.
24398-24586: LGTM! Comprehensive Stability AI image editing model coverage.The addition of Stability AI image editing models (both direct and Bedrock variants) provides comprehensive coverage with appropriate pricing tiers based on model capabilities (fast, conservative, creative, etc.).
25328-25363: LGTM! AWS Polly models are well-structured.The AWS Polly audio speech models follow appropriate pricing structures (character-based) with tiered costs reflecting quality differences. Source URLs are included for verification.
27908-27915: LGTM! Vertex AI OCR model includes appropriate pricing structure.The DeepSeek OCR model appropriately combines token-based pricing (for text I/O) with page-based pricing (for OCR processing), which aligns with typical OCR service pricing models.
28084-28111: LGTM! Vertex AI Veo models are consistent with Gemini variants.The Vertex AI Veo video generation models mirror the Gemini API variants added earlier with consistent pricing ($0.40/sec standard, $0.15/sec fast), which is expected for the same underlying models accessed through different APIs.
1360-1373: Structure is consistent within model types, but some entries lack source URLs.The JSON structure is well-organized and follows appropriate patterns for each model type. However, the azure/gpt-image-1.5 models (lines 3724-3749) do not include source URLs, unlike the azure_ai/gpt-oss-120b and gemini/veo-3.1 entries. Additionally, capability properties vary by modality: chat models include function calling flags, image generation models include endpoint and token cost specifications, and video generation models include modality support metadata.
Consider adding source URLs to the azure image generation models for consistency and verification purposes.
Likely an incorrect or invalid review comment.
18139-18140: The vision support additions are accurate. Groq's documentation confirms that bothmeta-llama/llama-4-scout-17b-16e-instructandmeta-llama/llama-4-maverick-17b-128e-instructare multimodal variants with vision capabilities, making the"supports_vision": trueflag correct for these models.
24998-25310: Structured output support across Together AI models is confirmed.Together AI's 2025 documentation confirms structured output support via JSON mode and json_schema response formats. The specific model families being updated in this PR (Qwen variants, DeepSeek-R1/V3, Llama variants, and others) are all documented as supporting structured outputs through Together AI's Chat Completions API.
|
Merged the most recent one already |
Details
Automated update of
model_prices_and_context_window.jsonfile and regeneration of supported models documentation.Files updated:
apps/opik-backend/src/main/resources/model_prices_and_context_window.json- Latest model pricing data from LiteLLMapps/opik-frontend/src/data/model_prices_and_context_window.json- Frontend copy kept in sync with backendChange checklist
Issues
N/A
Testing
Documentation
Summary by CodeRabbit
✏️ Tip: You can customize this high-level summary in your review settings.