Thanks to visit codestin.com
Credit goes to github.com

Skip to content

Conversation

@github-actions
Copy link
Contributor

@github-actions github-actions bot commented Dec 24, 2025

Details

Automated update of model_prices_and_context_window.json file and regeneration of supported models documentation.

Files updated:

  • apps/opik-backend/src/main/resources/model_prices_and_context_window.json - Latest model pricing data from LiteLLM
  • apps/opik-frontend/src/data/model_prices_and_context_window.json - Frontend copy kept in sync with backend

Change checklist

  • User facing change
  • Documentation update

Issues

N/A

Testing

  • Passed CI

Documentation

Summary by CodeRabbit

  • New Features
    • Added comprehensive support for new AI models across multiple providers including Azure, Google Vertex AI, OpenAI, Cohere, and Fireworks
    • Introduced expanded multimodal capabilities including image generation, audio transcription, video generation, embeddings, and web search
    • Enhanced model pricing configuration with detailed cost breakdowns for various operations and performance tiers
    • Improved feature support including function calling, vision capabilities, tool selection, and prompt caching

✏️ Tip: You can customize this high-level summary in your review settings.

@github-actions github-actions bot requested a review from a team as a code owner December 24, 2025 00:18
@coderabbitai
Copy link

coderabbitai bot commented Dec 24, 2025

Walkthrough

Two JSON resource files containing AI model pricing and context window data are updated with numerous new models and variants from providers including Azure AI, Google Vertex AI, OpenAI, and others. Each model entry includes pricing details, supported features (function calling, vision, etc.), and context window specifications. No control flow or public API changes are introduced.

Changes

Cohort / File(s) Summary
Model Pricing and Context Window Configuration
apps/opik-backend/src/main/resources/model_prices_and_context_window.json, apps/opik-frontend/src/data/model_prices_and_context_window.json
Added 100+ new model entries and variants across multiple providers (Azure AI, Google Vertex AI, OpenAI, Cohere, Fireworks, Deepgram, and others) with explicit pricing per token/pixel/second, multimodal support (text, image, audio, video), feature capabilities (function calling, system messages, vision), and context window limits. Updated existing entries with additional configuration variants including tiered pricing, experimental/preview labels, and deprecated model markers.

Estimated code review effort

🎯 2 (Simple) | ⏱️ ~12 minutes

Pre-merge checks and finishing touches

✅ Passed checks (3 passed)
Check name Status Explanation
Title check ✅ Passed The title '[NA] [BE] Update model prices file' clearly describes the primary change: updating the model prices configuration file across backend and frontend.
Description check ✅ Passed The description covers all required template sections with relevant details: it explains what files were updated, why, and includes testing status, though the Issues section uses 'N/A' instead of linking to specific issues.
Docstring Coverage ✅ Passed No functions found in the changed files to evaluate docstring coverage. Skipping docstring coverage check.
✨ Finishing touches
🧪 Generate unit tests (beta)
  • Create PR with unit tests
  • Post copyable unit tests in a comment
  • Commit unit tests in branch github-actions/NA-automatically-update-model-prices-file-2025-12-24-00-18-52

Comment @coderabbitai help to get the list of available commands and usage tips.

@Nimrod007
Copy link
Collaborator

@coderabbitai review

@coderabbitai
Copy link

coderabbitai bot commented Dec 24, 2025

✅ Actions performed

Review triggered.

Note: CodeRabbit is an incremental review system and does not re-review already reviewed commits. This command is applicable only when automatic reviews are paused.

Copy link

@coderabbitai coderabbitai bot left a comment

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

Actionable comments posted: 0

🧹 Nitpick comments (1)
apps/opik-frontend/src/data/model_prices_and_context_window.json (1)

1360-1373: Optional: Verify automated sync completed successfully.

Since this is an automated update from LiteLLM's model pricing data, consider spot-checking a few entries against the source to confirm the sync process worked as expected.

📜 Review details

Configuration used: Organization UI

Review profile: CHILL

Plan: Pro

📥 Commits

Reviewing files that changed from the base of the PR and between b26f7e9 and 5c97bff.

📒 Files selected for processing (2)
  • apps/opik-backend/src/main/resources/model_prices_and_context_window.json
  • apps/opik-frontend/src/data/model_prices_and_context_window.json
🔇 Additional comments (19)
apps/opik-frontend/src/data/model_prices_and_context_window.json (11)

1360-1373: LGTM! New Azure AI model addition looks correct.

The azure_ai/gpt-oss-120b model entry is well-structured with comprehensive pricing, token limits, and feature support flags.


3724-3749: New Azure image generation models added.

The azure/gpt-image-1.5 and azure/gpt-image-1.5-2025-12-16 models have been added with consistent pricing structures. Both include text and image token pricing, which appears intentional for multimodal support.


15407-15434: LGTM! Gemini Veo video generation models added.

Two variants of the Veo 3.1 model are properly configured with appropriate per-second output pricing and modality support for text-to-video generation.


18120-18120: Groq Llama Guard model addition.

New safety model entry added to the catalog.


18140-18140: Vision support enabled for Groq Llama models.

The supports_vision flag has been added to two Groq Llama models, reflecting enhanced multimodal capabilities.

Also applies to: 18153-18153


19393-19466: LGTM! Comprehensive MiniMax model catalog expansion.

Multiple MiniMax models added across speech synthesis and chat modes:

  • Speech models with HD and turbo variants at appropriate per-character pricing
  • Chat models with extensive feature support including prompt caching, function calling, and large context windows (up to 1M tokens)

24398-24586: LGTM! Extensive Stability AI image editing models added.

Significant expansion of image editing capabilities:

  • 15 direct Stability AI models covering various editing modes (inpaint, outpaint, erase, sketch, style transfer, etc.)
  • 13 corresponding Bedrock-hosted variants with consistent structure
  • Appropriate pricing tiers based on model capabilities (fast: $0.002, creative: $0.06, etc.)

24998-24998: Response schema support enabled across Together AI models.

The supports_response_schema flag has been consistently added to multiple Together AI models (Qwen, DeepSeek, Llama variants), indicating enhanced structured output capabilities across the provider's catalog.

Also applies to: 25006-25006, 25018-25018, 25030-25030, 25053-25053, 25066-25066, 25078-25078, 25091-25091, 25111-25111, 25141-25141, 25151-25151, 25161-25161, 25217-25217, 25229-25229, 25241-25241, 25260-25260, 25297-25297, 25309-25309


25328-25363: LGTM! AWS Polly text-to-speech models added.

Four AWS Polly variants are properly configured with tiered per-character pricing reflecting their capabilities (standard: $0.000004, neural: $0.000016, generative: $0.00003, long-form: $0.0001).


27908-27915: DeepSeek OCR model added to Vertex AI.

The OCR model entry properly includes both per-page OCR costs ($0.0003) and per-token pricing for text processing, which is appropriate for document understanding use cases.


28084-28111: LGTM! Vertex AI Veo video generation models added.

Two Veo 3.1 variants are configured for Vertex AI deployment with appropriate per-second output pricing, matching the Gemini-hosted variants added earlier in this update.

apps/opik-backend/src/main/resources/model_prices_and_context_window.json (8)

19393-19466: LGTM! Minimax model additions are comprehensive.

The new Minimax models are well-structured with appropriate pricing models (character-based for speech, token-based for chat) and comprehensive capability flags including prompt caching support for the chat models.


24398-24586: LGTM! Comprehensive Stability AI image editing model coverage.

The addition of Stability AI image editing models (both direct and Bedrock variants) provides comprehensive coverage with appropriate pricing tiers based on model capabilities (fast, conservative, creative, etc.).


25328-25363: LGTM! AWS Polly models are well-structured.

The AWS Polly audio speech models follow appropriate pricing structures (character-based) with tiered costs reflecting quality differences. Source URLs are included for verification.


27908-27915: LGTM! Vertex AI OCR model includes appropriate pricing structure.

The DeepSeek OCR model appropriately combines token-based pricing (for text I/O) with page-based pricing (for OCR processing), which aligns with typical OCR service pricing models.


28084-28111: LGTM! Vertex AI Veo models are consistent with Gemini variants.

The Vertex AI Veo video generation models mirror the Gemini API variants added earlier with consistent pricing ($0.40/sec standard, $0.15/sec fast), which is expected for the same underlying models accessed through different APIs.


1360-1373: Structure is consistent within model types, but some entries lack source URLs.

The JSON structure is well-organized and follows appropriate patterns for each model type. However, the azure/gpt-image-1.5 models (lines 3724-3749) do not include source URLs, unlike the azure_ai/gpt-oss-120b and gemini/veo-3.1 entries. Additionally, capability properties vary by modality: chat models include function calling flags, image generation models include endpoint and token cost specifications, and video generation models include modality support metadata.

Consider adding source URLs to the azure image generation models for consistency and verification purposes.

Likely an incorrect or invalid review comment.


18139-18140: The vision support additions are accurate. Groq's documentation confirms that both meta-llama/llama-4-scout-17b-16e-instruct and meta-llama/llama-4-maverick-17b-128e-instruct are multimodal variants with vision capabilities, making the "supports_vision": true flag correct for these models.


24998-25310: Structured output support across Together AI models is confirmed.

Together AI's 2025 documentation confirms structured output support via JSON mode and json_schema response formats. The specific model families being updated in this PR (Qwen variants, DeepSeek-R1/V3, Llama variants, and others) are all documented as supporting structured outputs through Together AI's Chat Completions API.

@thiagohora
Copy link
Contributor

Merged the most recent one already

@thiagohora thiagohora closed this Jan 5, 2026
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment

Labels

None yet

Projects

None yet

Development

Successfully merging this pull request may close these issues.

4 participants