faiss_rag_enterprise/llama_index/llms/__init__.py

119 lines
3.5 KiB
Python

from llama_index.core.llms.types import (
ChatMessage,
ChatResponse,
ChatResponseAsyncGen,
ChatResponseGen,
CompletionResponse,
CompletionResponseAsyncGen,
CompletionResponseGen,
LLMMetadata,
MessageRole,
)
from llama_index.llms.ai21 import AI21
from llama_index.llms.anthropic import Anthropic
from llama_index.llms.anyscale import Anyscale
from llama_index.llms.azure_openai import AzureOpenAI
from llama_index.llms.bedrock import Bedrock
from llama_index.llms.clarifai import Clarifai
from llama_index.llms.cohere import Cohere
from llama_index.llms.custom import CustomLLM
from llama_index.llms.dashscope import DashScope, DashScopeGenerationModels
from llama_index.llms.everlyai import EverlyAI
from llama_index.llms.gemini import Gemini
from llama_index.llms.gradient import GradientBaseModelLLM, GradientModelAdapterLLM
from llama_index.llms.huggingface import HuggingFaceInferenceAPI, HuggingFaceLLM
from llama_index.llms.konko import Konko
from llama_index.llms.langchain import LangChainLLM
from llama_index.llms.litellm import LiteLLM
from llama_index.llms.llama_cpp import LlamaCPP
from llama_index.llms.llm import LLM
from llama_index.llms.localai import LOCALAI_DEFAULTS, LocalAI
from llama_index.llms.mistral import MistralAI
from llama_index.llms.mock import MockLLM
from llama_index.llms.monsterapi import MonsterLLM
from llama_index.llms.neutrino import Neutrino
from llama_index.llms.nvidia_tensorrt import LocalTensorRTLLM
from llama_index.llms.nvidia_triton import NvidiaTriton
from llama_index.llms.ollama import Ollama
from llama_index.llms.openai import OpenAI
from llama_index.llms.openai_like import OpenAILike
from llama_index.llms.openllm import OpenLLM, OpenLLMAPI
from llama_index.llms.openrouter import OpenRouter
from llama_index.llms.palm import PaLM
from llama_index.llms.perplexity import Perplexity
from llama_index.llms.portkey import Portkey
from llama_index.llms.predibase import PredibaseLLM
from llama_index.llms.replicate import Replicate
from llama_index.llms.sagemaker_llm_endpoint import SageMakerLLM, SageMakerLLMEndPoint
from llama_index.llms.together import TogetherLLM
from llama_index.llms.vertex import Vertex
from llama_index.llms.vllm import Vllm, VllmServer
from llama_index.llms.watsonx import WatsonX
from llama_index.llms.xinference import Xinference
from llama_index.multi_modal_llms.dashscope import (
DashScopeMultiModal,
DashScopeMultiModalModels,
)
__all__ = [
"AI21",
"Anthropic",
"Anyscale",
"AzureOpenAI",
"Bedrock",
"ChatMessage",
"ChatResponse",
"ChatResponseAsyncGen",
"LLM",
"ChatResponseGen",
"Clarifai",
"Cohere",
"CompletionResponse",
"CompletionResponseAsyncGen",
"CompletionResponseGen",
"CustomLLM",
"EverlyAI",
"Gemini",
"GradientBaseModelLLM",
"GradientModelAdapterLLM",
"HuggingFaceInferenceAPI",
"HuggingFaceLLM",
"Konko",
"LLMMetadata",
"LangChainLLM",
"LiteLLM",
"LlamaCPP",
"LocalAI",
"LOCALAI_DEFAULTS",
"LocalTensorRTLLM",
"MessageRole",
"MockLLM",
"MonsterLLM",
"Neutrino",
"NvidiaTriton",
"MistralAI",
"Ollama",
"OpenAI",
"OpenAILike",
"OpenLLM",
"OpenLLMAPI",
"OpenRouter",
"PaLM",
"Perplexity",
"Portkey",
"PredibaseLLM",
"Replicate",
"SageMakerLLM",
"SageMakerLLMEndPoint", # deprecated
"TogetherLLM",
"WatsonX",
"Xinference",
"Vllm",
"VllmServer",
"Vertex",
"DashScope",
"DashScopeGenerationModels",
"DashScopeMultiModalModels",
"DashScopeMultiModal",
]