DeepInfraChat.py 3.7 KB

1234567891011121314151617181920212223242526272829303132333435363738394041424344454647484950515253545556575859606162636465666768697071727374757677787980818283
  1. from __future__ import annotations
  2. from .template import OpenaiTemplate
  3. class DeepInfraChat(OpenaiTemplate):
  4. url = "https://deepinfra.com/chat"
  5. api_base = "https://api.deepinfra.com/v1/openai"
  6. working = True
  7. default_model = 'deepseek-ai/DeepSeek-V3'
  8. default_vision_model = 'microsoft/Phi-4-multimodal-instruct'
  9. vision_models = [default_vision_model, 'meta-llama/Llama-3.2-90B-Vision-Instruct']
  10. models = [
  11. 'deepseek-ai/DeepSeek-Prover-V2-671B',
  12. 'Qwen/Qwen3-235B-A22B',
  13. 'Qwen/Qwen3-30B-A3B',
  14. 'Qwen/Qwen3-32B',
  15. 'Qwen/Qwen3-14B',
  16. 'meta-llama/Llama-4-Maverick-17B-128E-Instruct-FP8',
  17. 'meta-llama/Llama-4-Scout-17B-16E-Instruct',
  18. 'microsoft/phi-4-reasoning-plus',
  19. 'microsoft/meta-llama/Llama-Guard-4-12B',
  20. 'Qwen/QwQ-32B',
  21. 'deepseek-ai/DeepSeek-V3-0324',
  22. 'google/gemma-3-27b-it',
  23. 'google/gemma-3-12b-it',
  24. 'meta-llama/Meta-Llama-3.1-8B-Instruct',
  25. 'meta-llama/Llama-3.3-70B-Instruct-Turbo',
  26. default_model,
  27. 'mistralai/Mistral-Small-24B-Instruct-2501',
  28. 'deepseek-ai/DeepSeek-R1',
  29. 'deepseek-ai/DeepSeek-R1-Turbo',
  30. 'deepseek-ai/DeepSeek-R1-Distill-Llama-70B',
  31. 'deepseek-ai/DeepSeek-R1-Distill-Qwen-32B',
  32. 'microsoft/phi-4',
  33. 'microsoft/WizardLM-2-8x22B',
  34. 'Qwen/Qwen2.5-72B-Instruct',
  35. 'Qwen/Qwen2-72B-Instruct',
  36. 'cognitivecomputations/dolphin-2.6-mixtral-8x7b',
  37. 'cognitivecomputations/dolphin-2.9.1-llama-3-70b',
  38. 'deepinfra/airoboros-70b',
  39. 'lizpreciatior/lzlv_70b_fp16_hf',
  40. 'microsoft/WizardLM-2-7B',
  41. 'mistralai/Mixtral-8x22B-Instruct-v0.1',
  42. ] + vision_models
  43. model_aliases = {
  44. "deepseek-prover-v2-671b": "deepseek-ai/DeepSeek-Prover-V2-671B",
  45. "deepseek-prover-v2": "deepseek-ai/DeepSeek-Prover-V2-671B",
  46. "qwen-3-235b": "Qwen/Qwen3-235B-A22B",
  47. "qwen-3-30b": "Qwen/Qwen3-30B-A3B",
  48. "qwen-3-32b": "Qwen/Qwen3-32B",
  49. "qwen-3-14b": "Qwen/Qwen3-14B",
  50. "llama-4-maverick": "meta-llama/Llama-4-Maverick-17B-128E-Instruct-FP8",
  51. "llama-4-scout": "meta-llama/Llama-4-Scout-17B-16E-Instruct",
  52. "phi-4-reasoning-plus": "microsoft/phi-4-reasoning-plus",
  53. #"": "meta-llama/Llama-Guard-4-12B",
  54. "qwq-32b": "Qwen/QwQ-32B",
  55. "deepseek-v3": "deepseek-ai/DeepSeek-V3-0324",
  56. "deepseek-v3-0324": "deepseek-ai/DeepSeek-V3-0324",
  57. "gemma-3-27b": "google/gemma-3-27b-it",
  58. "gemma-3-12b": "google/gemma-3-12b-it",
  59. "phi-4-multimodal": "microsoft/Phi-4-multimodal-instruct",
  60. "llama-3.1-8b": "meta-llama/Meta-Llama-3.1-8B-Instruct",
  61. "llama-3.2-90b": "meta-llama/Llama-3.2-90B-Vision-Instruct",
  62. "llama-3.3-70b": "meta-llama/Llama-3.3-70B-Instruct",
  63. "deepseek-v3": default_model,
  64. "mistral-small": "mistralai/Mistral-Small-24B-Instruct-2501",
  65. "mixtral-small-24b": "mistralai/Mistral-Small-24B-Instruct-2501",
  66. "deepseek-r1-turbo": "deepseek-ai/DeepSeek-R1-Turbo",
  67. "deepseek-r1": "deepseek-ai/DeepSeek-R1",
  68. "deepseek-r1-distill-llama-70b": "deepseek-ai/DeepSeek-R1-Distill-Llama-70B",
  69. "deepseek-r1-distill-qwen-32b": "deepseek-ai/DeepSeek-R1-Distill-Qwen-32B",
  70. "phi-4": "microsoft/phi-4",
  71. "wizardlm-2-8x22b": "microsoft/WizardLM-2-8x22B",
  72. "qwen-2-72b": "Qwen/Qwen2-72B-Instruct",
  73. "dolphin-2.6": "cognitivecomputations/dolphin-2.6-mixtral-8x7b",
  74. "dolphin-2.9": "cognitivecomputations/dolphin-2.9.1-llama-3-70b",
  75. "airoboros-70b": "deepinfra/airoboros-70b",
  76. "lzlv-70b": "lizpreciatior/lzlv_70b_fp16_hf",
  77. "wizardlm-2-7b": "microsoft/WizardLM-2-7B",
  78. "mixtral-8x22b": "mistralai/Mixtral-8x22B-Instruct-v0.1"
  79. }