|
3663 | 3663 | "max_audio_length_hours": 8.4,
|
3664 | 3664 | "max_audio_per_prompt": 1,
|
3665 | 3665 | "max_pdf_size_mb": 30,
|
3666 |
| - "input_cost_per_audio_token": 0.001, |
3667 |
| - "input_cost_per_token": 0.00015, |
3668 |
| - "output_cost_per_token": 0.0006, |
| 3666 | + "input_cost_per_audio_token": 0.0000007, |
| 3667 | + "input_cost_per_token": 0.0000001, |
| 3668 | + "output_cost_per_token": 0.0000004, |
3669 | 3669 | "litellm_provider": "gemini",
|
3670 | 3670 | "mode": "chat",
|
| 3671 | + "rpm": 10000, |
| 3672 | + "tpm": 10000000, |
3671 | 3673 | "supports_system_messages": true,
|
3672 | 3674 | "supports_function_calling": true,
|
3673 | 3675 | "supports_vision": true,
|
3674 | 3676 | "supports_response_schema": true,
|
3675 | 3677 | "supports_audio_output": true,
|
3676 | 3678 | "supports_tool_choice": true,
|
3677 |
| - "source": "https://cloud.google.com/vertex-ai/generative-ai/docs/learn/models#gemini-2.0-flash" |
| 3679 | + "source": "https://ai.google.dev/pricing#2_0flash" |
3678 | 3680 | },
|
3679 | 3681 | "gemini-2.0-flash-001": {
|
3680 | 3682 | "max_tokens": 8192,
|
|
3767 | 3769 | "source": "https://cloud.google.com/vertex-ai/generative-ai/docs/learn/models#gemini-2.0-flash",
|
3768 | 3770 | "supports_tool_choice": true
|
3769 | 3771 | },
|
| 3772 | + "gemini/gemini-2.0-flash-001": { |
| 3773 | + "max_tokens": 8192, |
| 3774 | + "max_input_tokens": 1048576, |
| 3775 | + "max_output_tokens": 8192, |
| 3776 | + "max_images_per_prompt": 3000, |
| 3777 | + "max_videos_per_prompt": 10, |
| 3778 | + "max_video_length": 1, |
| 3779 | + "max_audio_length_hours": 8.4, |
| 3780 | + "max_audio_per_prompt": 1, |
| 3781 | + "max_pdf_size_mb": 30, |
| 3782 | + "input_cost_per_audio_token": 0.0000007, |
| 3783 | + "input_cost_per_token": 0.0000001, |
| 3784 | + "output_cost_per_token": 0.0000004, |
| 3785 | + "litellm_provider": "gemini", |
| 3786 | + "mode": "chat", |
| 3787 | + "rpm": 10000, |
| 3788 | + "tpm": 10000000, |
| 3789 | + "supports_system_messages": true, |
| 3790 | + "supports_function_calling": true, |
| 3791 | + "supports_vision": true, |
| 3792 | + "supports_response_schema": true, |
| 3793 | + "supports_audio_output": false, |
| 3794 | + "supports_tool_choice": true, |
| 3795 | + "source": "https://ai.google.dev/pricing#2_0flash" |
| 3796 | + }, |
3770 | 3797 | "gemini/gemini-2.0-flash-exp": {
|
3771 | 3798 | "max_tokens": 8192,
|
3772 | 3799 | "max_input_tokens": 1048576,
|
|
3803 | 3830 | "source": "https://cloud.google.com/vertex-ai/generative-ai/docs/learn/models#gemini-2.0-flash",
|
3804 | 3831 | "supports_tool_choice": true
|
3805 | 3832 | },
|
| 3833 | + "gemini/gemini-2.0-flash-lite-preview-02-05": { |
| 3834 | + "max_tokens": 8192, |
| 3835 | + "max_input_tokens": 1048576, |
| 3836 | + "max_output_tokens": 8192, |
| 3837 | + "max_images_per_prompt": 3000, |
| 3838 | + "max_videos_per_prompt": 10, |
| 3839 | + "max_video_length": 1, |
| 3840 | + "max_audio_length_hours": 8.4, |
| 3841 | + "max_audio_per_prompt": 1, |
| 3842 | + "max_pdf_size_mb": 30, |
| 3843 | + "input_cost_per_audio_token": 0.000000075, |
| 3844 | + "input_cost_per_token": 0.000000075, |
| 3845 | + "output_cost_per_token": 0.0000003, |
| 3846 | + "litellm_provider": "gemini", |
| 3847 | + "mode": "chat", |
| 3848 | + "rpm": 60000, |
| 3849 | + "tpm": 10000000, |
| 3850 | + "supports_system_messages": true, |
| 3851 | + "supports_function_calling": true, |
| 3852 | + "supports_vision": true, |
| 3853 | + "supports_response_schema": true, |
| 3854 | + "supports_audio_output": false, |
| 3855 | + "supports_tool_choice": true, |
| 3856 | + "source": "https://cloud.google.com/vertex-ai/generative-ai/docs/learn/models#gemini-2.0-flash-lite" |
| 3857 | + }, |
3806 | 3858 | "gemini/gemini-2.0-flash-thinking-exp": {
|
3807 | 3859 | "max_tokens": 8192,
|
3808 | 3860 | "max_input_tokens": 1048576,
|
|
8758 | 8810 | "source": "https://www.databricks.com/product/pricing/foundation-model-serving",
|
8759 | 8811 | "metadata": {"notes": "Input/output cost per token is dbu cost * $0.070, based on databricks Llama 3.1 70B conversion. Number provided for reference, '*_dbu_cost_per_token' used in actual calculation."},
|
8760 | 8812 | "supports_tool_choice": true
|
8761 |
| - |
8762 |
| - }, |
8763 |
| - "databricks/databricks-meta-llama-3-3-70b-instruct": { |
8764 |
| - "max_tokens": 128000, |
8765 |
| - "max_input_tokens": 128000, |
8766 |
| - "max_output_tokens": 128000, |
8767 |
| - "input_cost_per_token": 0.00000100002, |
8768 |
| - "input_dbu_cost_per_token": 0.000014286, |
8769 |
| - "output_cost_per_token": 0.00000299999, |
8770 |
| - "output_dbu_cost_per_token": 0.000042857, |
8771 |
| - "litellm_provider": "databricks", |
8772 |
| - "mode": "chat", |
8773 |
| - "source": "https://www.databricks.com/product/pricing/foundation-model-serving", |
8774 |
| - "metadata": {"notes": "Input/output cost per token is dbu cost * $0.070, based on databricks Llama 3.1 70B conversion. Number provided for reference, '*_dbu_cost_per_token' used in actual calculation."}, |
8775 |
| - "supports_tool_choice": true |
8776 | 8813 | },
|
8777 | 8814 | "databricks/databricks-dbrx-instruct": {
|
8778 | 8815 | "max_tokens": 32768,
|
|
0 commit comments