LLMs
ποΈ AI21
AI21 Studio provides API access to Jurassic-2 large language models.
ποΈ Aleph Alpha
The Luminous series is a family of large language models.
ποΈ Amazon API Gateway
Amazon API Gateway is a fully managed service that makes it easy for developers to create, publish, maintain, monitor, and secure APIs at any scale. APIs act as the "front door" for applications to access data, business logic, or functionality from your backend services. Using API Gateway, you can create RESTful APIs and WebSocket APIs that enable real-time two-way communication applications. API Gateway supports containerized and serverless workloads, as well as web applications.
ποΈ Anyscale
Anyscale is a fully-managed Ray platform, on which you can build, deploy, and manage scalable AI and Python applications
ποΈ Azure OpenAI
This notebook goes over how to use Langchain with Azure OpenAI.
ποΈ AzureML Online Endpoint
AzureML is a platform used to build, train, and deploy machine learning models. Users can explore the types of models to deploy in the Model Catalog, which provides Azure Foundation Models and OpenAI Models. Azure Foundation Models include various open-source models and popular Hugging Face models. Users can also import models of their liking into AzureML.
ποΈ Banana
Banana is focused on building the machine learning infrastructure.
ποΈ Baseten
Baseten provides all the infrastructure you need to deploy and serve ML models performantly, scalably, and cost-efficiently.
ποΈ Beam
Calls the Beam API wrapper to deploy and make subsequent calls to an instance of the gpt2 LLM in a cloud deployment. Requires installation of the Beam library and registration of Beam Client ID and Client Secret. By calling the wrapper an instance of the model is created and run, with returned text relating to the prompt. Additional calls can then be made by directly calling the Beam API.
ποΈ Bedrock
Amazon Bedrock is a fully managed service that makes FMs from leading AI startups and Amazon available via an API, so you can choose from a wide range of FMs to find the model that is best suited for your use case
ποΈ CerebriumAI
Cerebrium is an AWS Sagemaker alternative. It also provides API access to several LLM models.
ποΈ ChatGLM
ChatGLM-6B is an open bilingual language model based on General Language Model (GLM) framework, with 6.2 billion parameters. With the quantization technique, users can deploy locally on consumer-grade graphics cards (only 6GB of GPU memory is required at the INT4 quantization level).
ποΈ Clarifai
Clarifai is an AI Platform that provides the full AI lifecycle ranging from data exploration, data labeling, model training, evaluation, and inference.
ποΈ Cohere
Cohere is a Canadian startup that provides natural language processing models that help companies improve human-machine interactions.
ποΈ C Transformers
The C Transformers library provides Python bindings for GGML models.
ποΈ Databricks
The Databricks Lakehouse Platform unifies data, analytics, and AI on one platform.
ποΈ DeepInfra
DeepInfra provides several LLMs.
ποΈ ForefrontAI
The Forefront platform gives you the ability to fine-tune and use open source large language models.
ποΈ Google Cloud Platform Vertex AI PaLM
Note: This is seperate from the Google PaLM integration. Google has chosen to offer an enterprise version of PaLM through GCP, and this supports the models made available through there.
ποΈ GooseAI
GooseAI is a fully managed NLP-as-a-Service, delivered via API. GooseAI provides access to these models.
ποΈ GPT4All
GitHub:nomic-ai/gpt4all an ecosystem of open-source chatbots trained on a massive collections of clean assistant data including code, stories and dialogue.
ποΈ Hugging Face Hub
The Hugging Face Hub is a platform with over 120k models, 20k datasets, and 50k demo apps (Spaces), all open source and publicly available, in an online platform where people can easily collaborate and build ML together.
ποΈ Hugging Face Local Pipelines
Hugging Face models can be run locally through the HuggingFacePipeline class.
ποΈ Huggingface TextGen Inference
Text Generation Inference is a Rust, Python and gRPC server for text generation inference. Used in production at HuggingFace to power LLMs api-inference widgets.
ποΈ JSONFormer
JSONFormer is a library that wraps local HuggingFace pipeline models for structured decoding of a subset of the JSON Schema.
ποΈ KoboldAI API
KoboldAI is a "a browser-based front-end for AI-assisted writing with multiple local & remote AI models...". It has a public and local API that is able to be used in langchain.
ποΈ Llama-cpp
llama-cpp is a Python binding for llama.cpp.
ποΈ Caching integrations
This notebook covers how to cache results of individual LLM calls.
ποΈ Manifest
This notebook goes over how to use Manifest and LangChain.
ποΈ Modal
The Modal cloud platform provides convenient, on-demand access to serverless cloud compute from Python scripts on your local computer.
ποΈ MosaicML
MosaicML offers a managed inference service. You can either use a variety of open source models, or deploy your own.
ποΈ NLP Cloud
The NLP Cloud serves high performance pre-trained or custom models for NER, sentiment-analysis, classification, summarization, paraphrasing, grammar and spelling correction, keywords and keyphrases extraction, chatbot, product description and ad generation, intent classification, text generation, image generation, blog post generation, code generation, question answering, automatic speech recognition, machine translation, language detection, semantic search, semantic similarity, tokenization, POS tagging, embeddings, and dependency parsing. It is ready for production, served through a REST API.
ποΈ OctoAI Compute Service
This example goes over how to use LangChain to interact with OctoAI LLM endpoints
ποΈ OpenAI
OpenAI offers a spectrum of models with different levels of power suitable for different tasks.
ποΈ OpenLLM
π¦Ύ OpenLLM is an open platform for operating large language models (LLMs) in production. It enables developers to easily run inference with any open-source LLMs, deploy to the cloud or on-premises, and build powerful AI apps.
ποΈ OpenLM
OpenLM is a zero-dependency OpenAI-compatible LLM provider that can call different inference endpoints directly via HTTP.
ποΈ Petals
Petals runs 100B+ language models at home, BitTorrent-style.
ποΈ PipelineAI
PipelineAI allows you to run your ML models at scale in the cloud. It also provides API access to several LLM models.
ποΈ Predibase
Predibase allows you to train, finetune, and deploy any ML modelβfrom linear regression to large language model.
ποΈ Prediction Guard
Basic LLM usage
ποΈ PromptLayer OpenAI
PromptLayer is the first platform that allows you to track, manage, and share your GPT prompt engineering. PromptLayer acts a middleware between your code and OpenAIβs python library.
ποΈ RELLM
RELLM is a library that wraps local Hugging Face pipeline models for structured decoding.
ποΈ Replicate
Replicate runs machine learning models in the cloud. We have a library of open-source models that you can run with a few lines of code. If you're building your own machine learning models, Replicate makes it easy to deploy them at scale.
ποΈ Runhouse
The Runhouse allows remote compute and data across environments and users. See the Runhouse docs.
ποΈ SageMakerEndpoint
Amazon SageMaker is a system that can build, train, and deploy machine learning (ML) models for any use case with fully managed infrastructure, tools, and workflows.
ποΈ StochasticAI
Stochastic Acceleration Platform aims to simplify the life cycle of a Deep Learning model. From uploading and versioning the model, through training, compression and acceleration to putting it into production.
ποΈ TextGen
GitHub:oobabooga/text-generation-webui A gradio web UI for running Large Language Models like LLaMA, llama.cpp, GPT-J, Pythia, OPT, and GALACTICA.
ποΈ Tongyi Qwen
Tongyi Qwen is a large-scale language model developed by Alibaba's Damo Academy. It is capable of understanding user intent through natural language understanding and semantic analysis, based on user input in natural language. It provides services and assistance to users in different domains and tasks. By providing clear and detailed instructions, you can obtain results that better align with your expectations.
ποΈ Writer
Writer is a platform to generate different language content.