|
8 | 8 | "input_cache_read_cost_per_mil_tokens": 0.025, |
9 | 9 | "displayName": "GPT-5 mini", |
10 | 10 | "reasoning": true, |
11 | | - "max_input_tokens": 400000, |
| 11 | + "max_input_tokens": 272000, |
12 | 12 | "max_output_tokens": 128000 |
13 | 13 | }, |
14 | 14 | "gpt-5-mini-2025-08-07": { |
|
21 | 21 | "displayName": "GPT-5 mini (2025-08-07)", |
22 | 22 | "reasoning": true, |
23 | 23 | "parent": "gpt-5-mini", |
24 | | - "max_input_tokens": 400000, |
| 24 | + "max_input_tokens": 272000, |
25 | 25 | "max_output_tokens": 128000 |
26 | 26 | }, |
27 | 27 | "gpt-5": { |
|
33 | 33 | "input_cache_read_cost_per_mil_tokens": 0.125, |
34 | 34 | "displayName": "GPT-5", |
35 | 35 | "reasoning": true, |
36 | | - "max_input_tokens": 400000, |
| 36 | + "max_input_tokens": 272000, |
37 | 37 | "max_output_tokens": 128000 |
38 | 38 | }, |
39 | 39 | "gpt-5-2025-08-07": { |
|
46 | 46 | "displayName": "GPT-5 (2025-08-07)", |
47 | 47 | "reasoning": true, |
48 | 48 | "parent": "gpt-5", |
49 | | - "max_input_tokens": 400000, |
| 49 | + "max_input_tokens": 272000, |
50 | 50 | "max_output_tokens": 128000 |
51 | 51 | }, |
52 | 52 | "gpt-5-pro": { |
|
57 | 57 | "output_cost_per_mil_tokens": 120, |
58 | 58 | "displayName": "GPT-5 Pro", |
59 | 59 | "reasoning": true, |
60 | | - "max_input_tokens": 400000, |
| 60 | + "max_input_tokens": 128000, |
61 | 61 | "max_output_tokens": 272000 |
62 | 62 | }, |
63 | 63 | "gpt-5-pro-2025-10-06": { |
|
68 | 68 | "output_cost_per_mil_tokens": 120, |
69 | 69 | "reasoning": true, |
70 | 70 | "parent": "gpt-5-pro", |
71 | | - "max_input_tokens": 400000, |
| 71 | + "max_input_tokens": 128000, |
72 | 72 | "max_output_tokens": 272000 |
73 | 73 | }, |
74 | 74 | "gpt-5-codex": { |
|
80 | 80 | "input_cache_read_cost_per_mil_tokens": 0.125, |
81 | 81 | "displayName": "GPT-5 Codex", |
82 | 82 | "reasoning": true, |
83 | | - "max_input_tokens": 400000, |
| 83 | + "max_input_tokens": 272000, |
84 | 84 | "max_output_tokens": 128000 |
85 | 85 | }, |
86 | 86 | "gpt-5-nano": { |
|
92 | 92 | "input_cache_read_cost_per_mil_tokens": 0.005, |
93 | 93 | "displayName": "GPT-5 nano", |
94 | 94 | "reasoning": true, |
95 | | - "max_input_tokens": 400000, |
| 95 | + "max_input_tokens": 272000, |
96 | 96 | "max_output_tokens": 128000 |
97 | 97 | }, |
98 | 98 | "gpt-5-nano-2025-08-07": { |
|
105 | 105 | "displayName": "GPT-5 nano (2025-08-07)", |
106 | 106 | "reasoning": true, |
107 | 107 | "parent": "gpt-5-nano", |
108 | | - "max_input_tokens": 400000, |
| 108 | + "max_input_tokens": 272000, |
109 | 109 | "max_output_tokens": 128000 |
110 | 110 | }, |
111 | 111 | "gpt-5-chat-latest": { |
|
117 | 117 | "input_cache_read_cost_per_mil_tokens": 0.125, |
118 | 118 | "displayName": "GPT-5 chat", |
119 | 119 | "reasoning": true, |
120 | | - "max_input_tokens": 400000, |
| 120 | + "max_input_tokens": 128000, |
121 | 121 | "max_output_tokens": 16384 |
122 | 122 | }, |
123 | 123 | "gpt-4o": { |
|
850 | 850 | "reasoning_budget": true, |
851 | 851 | "deprecation_date": "2025-06-01", |
852 | 852 | "max_input_tokens": 200000, |
853 | | - "max_output_tokens": 128000 |
| 853 | + "max_output_tokens": 64000 |
854 | 854 | }, |
855 | 855 | "claude-3-7-sonnet-20250219": { |
856 | 856 | "format": "anthropic", |
|
865 | 865 | "deprecation_date": "2026-02-19", |
866 | 866 | "parent": "claude-3-7-sonnet-latest", |
867 | 867 | "max_input_tokens": 200000, |
868 | | - "max_output_tokens": 128000 |
| 868 | + "max_output_tokens": 64000 |
869 | 869 | }, |
870 | 870 | "claude-haiku-4-5": { |
871 | 871 | "format": "anthropic", |
|
974 | 974 | "displayName": "Claude 4.6 Opus", |
975 | 975 | "reasoning": true, |
976 | 976 | "reasoning_budget": true, |
977 | | - "max_input_tokens": 200000, |
| 977 | + "max_input_tokens": 1000000, |
978 | 978 | "max_output_tokens": 128000 |
979 | 979 | }, |
980 | 980 | "claude-opus-4-5": { |
|
2389 | 2389 | "displayName": "Gemini 3 Pro (Preview)", |
2390 | 2390 | "reasoning": true, |
2391 | 2391 | "reasoning_budget": true, |
| 2392 | + "deprecation_date": "2026-03-26", |
2392 | 2393 | "max_input_tokens": 1048576, |
2393 | 2394 | "max_output_tokens": 65535 |
2394 | 2395 | }, |
|
2587 | 2588 | "output_cost_per_mil_tokens": 0.4, |
2588 | 2589 | "input_cache_read_cost_per_mil_tokens": 0.025, |
2589 | 2590 | "displayName": "Gemini 2.0 Flash Latest", |
2590 | | - "deprecation_date": "2026-03-31", |
| 2591 | + "deprecation_date": "2026-06-01", |
2591 | 2592 | "max_input_tokens": 1048576, |
2592 | 2593 | "max_output_tokens": 8192 |
2593 | 2594 | }, |
|
2598 | 2599 | "input_cost_per_mil_tokens": 0.15, |
2599 | 2600 | "output_cost_per_mil_tokens": 0.6, |
2600 | 2601 | "input_cache_read_cost_per_mil_tokens": 0.0375, |
2601 | | - "deprecation_date": "2026-03-31", |
| 2602 | + "deprecation_date": "2026-06-01", |
2602 | 2603 | "parent": "gemini-2.0-flash", |
2603 | 2604 | "max_input_tokens": 1048576, |
2604 | 2605 | "max_output_tokens": 8192 |
|
2611 | 2612 | "output_cost_per_mil_tokens": 0.3, |
2612 | 2613 | "input_cache_read_cost_per_mil_tokens": 0.01875, |
2613 | 2614 | "displayName": "Gemini 2.0 Flash-Lite", |
2614 | | - "deprecation_date": "2026-03-31", |
| 2615 | + "deprecation_date": "2026-06-01", |
2615 | 2616 | "max_input_tokens": 1048576, |
2616 | 2617 | "max_output_tokens": 8192 |
2617 | 2618 | }, |
|
2622 | 2623 | "input_cost_per_mil_tokens": 0.075, |
2623 | 2624 | "output_cost_per_mil_tokens": 0.3, |
2624 | 2625 | "input_cache_read_cost_per_mil_tokens": 0.01875, |
2625 | | - "deprecation_date": "2026-03-31", |
| 2626 | + "deprecation_date": "2026-06-01", |
2626 | 2627 | "parent": "gemini-2.0-flash-lite", |
2627 | 2628 | "max_input_tokens": 1048576, |
2628 | 2629 | "max_output_tokens": 8192 |
|
3001 | 3002 | "output_cost_per_mil_tokens": 0.5, |
3002 | 3003 | "input_cache_read_cost_per_mil_tokens": 0.075, |
3003 | 3004 | "reasoning": true, |
| 3005 | + "deprecation_date": "2026-02-28", |
3004 | 3006 | "max_input_tokens": 131072, |
3005 | 3007 | "max_output_tokens": 131072 |
3006 | 3008 | }, |
|
3021 | 3023 | "output_cost_per_mil_tokens": 0.5, |
3022 | 3024 | "input_cache_read_cost_per_mil_tokens": 0.075, |
3023 | 3025 | "reasoning": true, |
| 3026 | + "deprecation_date": "2026-02-28", |
3024 | 3027 | "max_input_tokens": 131072, |
3025 | 3028 | "max_output_tokens": 131072 |
3026 | 3029 | }, |
|
3098 | 3101 | "multimodal": true, |
3099 | 3102 | "input_cost_per_mil_tokens": 2, |
3100 | 3103 | "output_cost_per_mil_tokens": 10, |
| 3104 | + "deprecation_date": "2026-02-28", |
3101 | 3105 | "parent": "grok-2-vision", |
3102 | 3106 | "max_input_tokens": 32768, |
3103 | 3107 | "max_output_tokens": 32768 |
|
3417 | 3421 | "input_cache_read_cost_per_mil_tokens": 0.3, |
3418 | 3422 | "input_cache_write_cost_per_mil_tokens": 3.75, |
3419 | 3423 | "displayName": "Claude 3.5 Sonnet v2", |
3420 | | - "max_input_tokens": 200000, |
| 3424 | + "max_input_tokens": 1000000, |
3421 | 3425 | "max_output_tokens": 8192 |
3422 | 3426 | }, |
3423 | 3427 | "us.anthropic.claude-3-5-sonnet-20241022-v2:0": { |
|
3452 | 3456 | "multimodal": true, |
3453 | 3457 | "input_cost_per_mil_tokens": 3, |
3454 | 3458 | "output_cost_per_mil_tokens": 15, |
| 3459 | + "input_cache_read_cost_per_mil_tokens": 0.3, |
| 3460 | + "input_cache_write_cost_per_mil_tokens": 3.75, |
3455 | 3461 | "displayName": "Claude 3.5 Sonnet", |
3456 | | - "max_input_tokens": 200000, |
| 3462 | + "max_input_tokens": 1000000, |
3457 | 3463 | "max_output_tokens": 4096 |
3458 | 3464 | }, |
3459 | 3465 | "us.anthropic.claude-3-5-sonnet-20240620-v1:0": { |
|
3462 | 3468 | "multimodal": true, |
3463 | 3469 | "input_cost_per_mil_tokens": 3, |
3464 | 3470 | "output_cost_per_mil_tokens": 15, |
| 3471 | + "input_cache_read_cost_per_mil_tokens": 0.3, |
| 3472 | + "input_cache_write_cost_per_mil_tokens": 3.75, |
3465 | 3473 | "displayName": "US Claude 3.5 Sonnet", |
3466 | 3474 | "parent": "anthropic.claude-3-5-sonnet-20240620-v1:0", |
3467 | 3475 | "max_input_tokens": 200000, |
|
3473 | 3481 | "multimodal": true, |
3474 | 3482 | "input_cost_per_mil_tokens": 3, |
3475 | 3483 | "output_cost_per_mil_tokens": 15, |
| 3484 | + "input_cache_read_cost_per_mil_tokens": 0.3, |
| 3485 | + "input_cache_write_cost_per_mil_tokens": 3.75, |
3476 | 3486 | "displayName": "APAC Claude 3.5 Sonnet", |
3477 | 3487 | "parent": "anthropic.claude-3-5-sonnet-20240620-v1:0", |
3478 | 3488 | "max_input_tokens": 200000, |
|
3484 | 3494 | "multimodal": true, |
3485 | 3495 | "input_cost_per_mil_tokens": 3, |
3486 | 3496 | "output_cost_per_mil_tokens": 15, |
| 3497 | + "input_cache_read_cost_per_mil_tokens": 0.3, |
| 3498 | + "input_cache_write_cost_per_mil_tokens": 3.75, |
3487 | 3499 | "displayName": "EU Claude 3.5 Sonnet", |
3488 | 3500 | "parent": "anthropic.claude-3-5-sonnet-20240620-v1:0", |
3489 | 3501 | "max_input_tokens": 200000, |
|
3607 | 3619 | "multimodal": true, |
3608 | 3620 | "input_cost_per_mil_tokens": 15, |
3609 | 3621 | "output_cost_per_mil_tokens": 75, |
| 3622 | + "input_cache_read_cost_per_mil_tokens": 1.5, |
| 3623 | + "input_cache_write_cost_per_mil_tokens": 18.75, |
3610 | 3624 | "displayName": "Claude 3 Opus", |
3611 | 3625 | "max_input_tokens": 200000, |
3612 | 3626 | "max_output_tokens": 4096 |
|
3617 | 3631 | "multimodal": true, |
3618 | 3632 | "input_cost_per_mil_tokens": 15, |
3619 | 3633 | "output_cost_per_mil_tokens": 75, |
| 3634 | + "input_cache_read_cost_per_mil_tokens": 1.5, |
| 3635 | + "input_cache_write_cost_per_mil_tokens": 18.75, |
3620 | 3636 | "displayName": "US Claude 3 Opus", |
3621 | 3637 | "parent": "anthropic.claude-3-opus-20240229-v1:0", |
3622 | 3638 | "max_input_tokens": 200000, |
|
3628 | 3644 | "multimodal": true, |
3629 | 3645 | "input_cost_per_mil_tokens": 3, |
3630 | 3646 | "output_cost_per_mil_tokens": 15, |
| 3647 | + "input_cache_read_cost_per_mil_tokens": 0.3, |
| 3648 | + "input_cache_write_cost_per_mil_tokens": 3.75, |
3631 | 3649 | "displayName": "Claude 3 Sonnet", |
3632 | 3650 | "max_input_tokens": 200000, |
3633 | 3651 | "max_output_tokens": 4096 |
|
3638 | 3656 | "multimodal": true, |
3639 | 3657 | "input_cost_per_mil_tokens": 3, |
3640 | 3658 | "output_cost_per_mil_tokens": 15, |
| 3659 | + "input_cache_read_cost_per_mil_tokens": 0.3, |
| 3660 | + "input_cache_write_cost_per_mil_tokens": 3.75, |
3641 | 3661 | "displayName": "US Claude 3 Sonnet", |
3642 | 3662 | "parent": "anthropic.claude-3-sonnet-20240229-v1:0", |
3643 | 3663 | "max_input_tokens": 200000, |
|
3649 | 3669 | "multimodal": true, |
3650 | 3670 | "input_cost_per_mil_tokens": 3, |
3651 | 3671 | "output_cost_per_mil_tokens": 15, |
| 3672 | + "input_cache_read_cost_per_mil_tokens": 0.3, |
| 3673 | + "input_cache_write_cost_per_mil_tokens": 3.75, |
3652 | 3674 | "displayName": "APAC Claude 3 Sonnet", |
3653 | 3675 | "parent": "anthropic.claude-3-sonnet-20240229-v1:0", |
3654 | 3676 | "max_input_tokens": 200000, |
|
3660 | 3682 | "multimodal": true, |
3661 | 3683 | "input_cost_per_mil_tokens": 3, |
3662 | 3684 | "output_cost_per_mil_tokens": 15, |
| 3685 | + "input_cache_read_cost_per_mil_tokens": 0.3, |
| 3686 | + "input_cache_write_cost_per_mil_tokens": 3.75, |
3663 | 3687 | "displayName": "EU Claude 3 Sonnet", |
3664 | 3688 | "parent": "anthropic.claude-3-sonnet-20240229-v1:0", |
3665 | 3689 | "max_input_tokens": 200000, |
|
3671 | 3695 | "multimodal": true, |
3672 | 3696 | "input_cost_per_mil_tokens": 0.25, |
3673 | 3697 | "output_cost_per_mil_tokens": 1.25, |
| 3698 | + "input_cache_read_cost_per_mil_tokens": 0.025, |
| 3699 | + "input_cache_write_cost_per_mil_tokens": 0.3125, |
3674 | 3700 | "displayName": "Claude 3 Haiku", |
3675 | 3701 | "max_input_tokens": 200000, |
3676 | 3702 | "max_output_tokens": 4096 |
|
3681 | 3707 | "multimodal": true, |
3682 | 3708 | "input_cost_per_mil_tokens": 0.25, |
3683 | 3709 | "output_cost_per_mil_tokens": 1.25, |
| 3710 | + "input_cache_read_cost_per_mil_tokens": 0.025, |
| 3711 | + "input_cache_write_cost_per_mil_tokens": 0.3125, |
3684 | 3712 | "displayName": "US Claude 3 Haiku", |
3685 | 3713 | "parent": "anthropic.claude-3-haiku-20240307-v1:0", |
3686 | 3714 | "max_input_tokens": 200000, |
|
3692 | 3720 | "multimodal": true, |
3693 | 3721 | "input_cost_per_mil_tokens": 0.25, |
3694 | 3722 | "output_cost_per_mil_tokens": 1.25, |
| 3723 | + "input_cache_read_cost_per_mil_tokens": 0.025, |
| 3724 | + "input_cache_write_cost_per_mil_tokens": 0.3125, |
3695 | 3725 | "displayName": "APAC Claude 3 Haiku", |
3696 | 3726 | "parent": "anthropic.claude-3-haiku-20240307-v1:0", |
3697 | 3727 | "max_input_tokens": 200000, |
|
3703 | 3733 | "multimodal": true, |
3704 | 3734 | "input_cost_per_mil_tokens": 0.25, |
3705 | 3735 | "output_cost_per_mil_tokens": 1.25, |
| 3736 | + "input_cache_read_cost_per_mil_tokens": 0.025, |
| 3737 | + "input_cache_write_cost_per_mil_tokens": 0.3125, |
3706 | 3738 | "displayName": "EU Claude 3 Haiku", |
3707 | 3739 | "parent": "anthropic.claude-3-haiku-20240307-v1:0", |
3708 | 3740 | "max_input_tokens": 200000, |
|
4012 | 4044 | "displayName": "Gemini 3 Pro Preview", |
4013 | 4045 | "reasoning": true, |
4014 | 4046 | "reasoning_budget": true, |
4015 | | - "locations": ["global"], |
| 4047 | + "locations": [ |
| 4048 | + "global" |
| 4049 | + ], |
4016 | 4050 | "max_input_tokens": 1048576, |
4017 | 4051 | "max_output_tokens": 65535 |
4018 | 4052 | }, |
|
4026 | 4060 | "displayName": "Gemini 3 Flash Preview", |
4027 | 4061 | "reasoning": true, |
4028 | 4062 | "reasoning_budget": true, |
4029 | | - "locations": ["global"], |
| 4063 | + "locations": [ |
| 4064 | + "global" |
| 4065 | + ], |
4030 | 4066 | "max_input_tokens": 1048576, |
4031 | 4067 | "max_output_tokens": 65535 |
4032 | 4068 | }, |
|
4765 | 4801 | "input_cost_per_mil_tokens": 1.2, |
4766 | 4802 | "output_cost_per_mil_tokens": 1.2, |
4767 | 4803 | "input_cache_read_cost_per_mil_tokens": 0.1, |
4768 | | - "reasoning": true, |
4769 | 4804 | "displayName": "Kimi K2.5", |
| 4805 | + "reasoning": true, |
4770 | 4806 | "max_input_tokens": 131072, |
4771 | 4807 | "max_output_tokens": 131072 |
4772 | 4808 | }, |
|
5070 | 5106 | "input_cost_per_mil_tokens": 0.22, |
5071 | 5107 | "output_cost_per_mil_tokens": 0.88, |
5072 | 5108 | "displayName": "Qwen3 235B A22B Instruct 2507", |
| 5109 | + "locations": [ |
| 5110 | + "global", |
| 5111 | + "us-south1" |
| 5112 | + ], |
5073 | 5113 | "max_input_tokens": 262144, |
5074 | | - "max_output_tokens": 16384, |
5075 | | - "locations": ["global", "us-south1"] |
| 5114 | + "max_output_tokens": 16384 |
5076 | 5115 | }, |
5077 | 5116 | "accounts/fireworks/models/deepseek-v3p2": { |
5078 | 5117 | "format": "openai", |
|
0 commit comments