119 lines
3.5 KiB
Python
119 lines
3.5 KiB
Python
from llama_index.core.llms.types import (
|
|
ChatMessage,
|
|
ChatResponse,
|
|
ChatResponseAsyncGen,
|
|
ChatResponseGen,
|
|
CompletionResponse,
|
|
CompletionResponseAsyncGen,
|
|
CompletionResponseGen,
|
|
LLMMetadata,
|
|
MessageRole,
|
|
)
|
|
from llama_index.llms.ai21 import AI21
|
|
from llama_index.llms.anthropic import Anthropic
|
|
from llama_index.llms.anyscale import Anyscale
|
|
from llama_index.llms.azure_openai import AzureOpenAI
|
|
from llama_index.llms.bedrock import Bedrock
|
|
from llama_index.llms.clarifai import Clarifai
|
|
from llama_index.llms.cohere import Cohere
|
|
from llama_index.llms.custom import CustomLLM
|
|
from llama_index.llms.dashscope import DashScope, DashScopeGenerationModels
|
|
from llama_index.llms.everlyai import EverlyAI
|
|
from llama_index.llms.gemini import Gemini
|
|
from llama_index.llms.gradient import GradientBaseModelLLM, GradientModelAdapterLLM
|
|
from llama_index.llms.huggingface import HuggingFaceInferenceAPI, HuggingFaceLLM
|
|
from llama_index.llms.konko import Konko
|
|
from llama_index.llms.langchain import LangChainLLM
|
|
from llama_index.llms.litellm import LiteLLM
|
|
from llama_index.llms.llama_cpp import LlamaCPP
|
|
from llama_index.llms.llm import LLM
|
|
from llama_index.llms.localai import LOCALAI_DEFAULTS, LocalAI
|
|
from llama_index.llms.mistral import MistralAI
|
|
from llama_index.llms.mock import MockLLM
|
|
from llama_index.llms.monsterapi import MonsterLLM
|
|
from llama_index.llms.neutrino import Neutrino
|
|
from llama_index.llms.nvidia_tensorrt import LocalTensorRTLLM
|
|
from llama_index.llms.nvidia_triton import NvidiaTriton
|
|
from llama_index.llms.ollama import Ollama
|
|
from llama_index.llms.openai import OpenAI
|
|
from llama_index.llms.openai_like import OpenAILike
|
|
from llama_index.llms.openllm import OpenLLM, OpenLLMAPI
|
|
from llama_index.llms.openrouter import OpenRouter
|
|
from llama_index.llms.palm import PaLM
|
|
from llama_index.llms.perplexity import Perplexity
|
|
from llama_index.llms.portkey import Portkey
|
|
from llama_index.llms.predibase import PredibaseLLM
|
|
from llama_index.llms.replicate import Replicate
|
|
from llama_index.llms.sagemaker_llm_endpoint import SageMakerLLM, SageMakerLLMEndPoint
|
|
from llama_index.llms.together import TogetherLLM
|
|
from llama_index.llms.vertex import Vertex
|
|
from llama_index.llms.vllm import Vllm, VllmServer
|
|
from llama_index.llms.watsonx import WatsonX
|
|
from llama_index.llms.xinference import Xinference
|
|
from llama_index.multi_modal_llms.dashscope import (
|
|
DashScopeMultiModal,
|
|
DashScopeMultiModalModels,
|
|
)
|
|
|
|
__all__ = [
|
|
"AI21",
|
|
"Anthropic",
|
|
"Anyscale",
|
|
"AzureOpenAI",
|
|
"Bedrock",
|
|
"ChatMessage",
|
|
"ChatResponse",
|
|
"ChatResponseAsyncGen",
|
|
"LLM",
|
|
"ChatResponseGen",
|
|
"Clarifai",
|
|
"Cohere",
|
|
"CompletionResponse",
|
|
"CompletionResponseAsyncGen",
|
|
"CompletionResponseGen",
|
|
"CustomLLM",
|
|
"EverlyAI",
|
|
"Gemini",
|
|
"GradientBaseModelLLM",
|
|
"GradientModelAdapterLLM",
|
|
"HuggingFaceInferenceAPI",
|
|
"HuggingFaceLLM",
|
|
"Konko",
|
|
"LLMMetadata",
|
|
"LangChainLLM",
|
|
"LiteLLM",
|
|
"LlamaCPP",
|
|
"LocalAI",
|
|
"LOCALAI_DEFAULTS",
|
|
"LocalTensorRTLLM",
|
|
"MessageRole",
|
|
"MockLLM",
|
|
"MonsterLLM",
|
|
"Neutrino",
|
|
"NvidiaTriton",
|
|
"MistralAI",
|
|
"Ollama",
|
|
"OpenAI",
|
|
"OpenAILike",
|
|
"OpenLLM",
|
|
"OpenLLMAPI",
|
|
"OpenRouter",
|
|
"PaLM",
|
|
"Perplexity",
|
|
"Portkey",
|
|
"PredibaseLLM",
|
|
"Replicate",
|
|
"SageMakerLLM",
|
|
"SageMakerLLMEndPoint", # deprecated
|
|
"TogetherLLM",
|
|
"WatsonX",
|
|
"Xinference",
|
|
"Vllm",
|
|
"VllmServer",
|
|
"Vertex",
|
|
"DashScope",
|
|
"DashScopeGenerationModels",
|
|
"DashScopeMultiModalModels",
|
|
"DashScopeMultiModal",
|
|
]
|