You are here: Home

Modified items

All recently modified items, latest first.
RPMPackage python3-langchain-xai-0.2.3-1.lbn36.noarch
langchain-xai This package contains the LangChain integrations for xAI through their APIs. Installation and Setup Install the LangChain partner package pip install -U langchain-xai Get your xAI api key from the xAI Dashboard and set it as an environment variable (XAI_API_KEY) Chat Completions This package contains the ChatXAI class, which is the recommended way to interface with xAI chat models.
RPMPackage python3-langchain-unstructured-0.1.5-1.lbn36.noarch
langchain-unstructured This package contains the LangChain integration with Unstructured Installation pip install -U langchain-unstructured And you should configure credentials by setting the following environment variables: export UNSTRUCTURED_API_KEY="your-api-key" Loaders Partition and load files using either the unstructured-client sdk and the Unstructured API or locally using the unstructured library. API: To partition via the Unstructured API pip install unstructured-client and set partition_via_api=True and define api_key. If you are running the unstructured API locally, you can change the API rule by defining url when you initialize the loader. The hosted Unstructured API requires an API key. See the links below to learn more about our API offerings and get an API key. Local: By default the file loader uses the Unstructured partition function and will automatically detect the file type. In addition to document specific partition parameters, Unstructured has a rich set of "chu
RPMPackage python3-langchain-together-0.3.0-1.lbn36.noarch
langchain-together This package contains the LangChain integrations for Together AI through their APIs. Installation and Setup Install the LangChain partner package pip install -U langchain-together Get your Together AI api key from the Together Dashboard and set it as an environment variable (TOGETHER_API_KEY) Chat Completions This package contains the ChatTogether class, which is the recommended way to interface with Together AI chat models. ADD USAGE EXAMPLE HERE. Can we add this in the langchain docs? NEED to add image endpoint + completions endpoint as well Embeddings See a usage example Use togethercomputer/m2-bert-80M-8k-retrieval as the default model for embeddings.
RPMPackage python3-langchain-text-splitters-0.3.8-1.lbn36.noarch
šŸ¦œāœ‚ļø LangChain Text Splitters Quick Install pip install langchain-text-splitters What is it? LangChain Text Splitters contains utilities for splitting into chunks a wide variety of text documents. For full documentation see the API reference and the Text Splitters module in the main docs. šŸ“• Releases & Versioning langchain-text-splitters is currently on version 0.0.x. Minor version increases will occur for: Breaking changes for any public interfaces NOT marked beta Patch version increases will occur for: Bug fixes New features Any changes to private interfaces Any changes to beta features šŸ’ Contributing As an open-source project in a rapidly developing field, we are extremely open to contributions, whether it be in the form of a new feature, improved infrastructure, or better documentation. For detailed information on how to contribute, see the Contributing Guide.
RPMPackage python3-langchain-tests-0.3.19-1.lbn36.noarch
langchain-tests This is a testing library for LangChain integrations. It contains the base classes for a standard set of tests. Installation We encourage pinning your version to a specific version in order to avoid breaking your CI when we publish new tests. We recommend upgrading to the latest version periodically to make sure you have the latest tests. Not pinning your version will ensure you always have the latest tests, but it may also break your CI if we introduce tests that your integration doesn't pass. Pip: ```bash pip install -U langchain-tests ``` Poetry: ```bash poetry add langchain-tests ``` Usage To add standard tests to an integration package's e.g. ChatModel, you need to create A unit test class that inherits from ChatModelUnitTests An integration test class that inherits from ChatModelIntegrationTests tests/unit_tests/test_standard.py: """Standard LangChain interface tests""" from typing import Type import pytest from langchain_core.language_models import BaseChat
RPMPackage python3-langchain-sambanova-0.1.0-1.lbn36.noarch
langchain-sambanova
RPMPackage python3-langchain-pinecone-0.2.2-1.lbn36.noarch
langchain-pinecone This package contains the LangChain integration with Pinecone. Installation pip install -U langchain-pinecone And you should configure credentials by setting the following environment variables: PINECONE_API_KEY PINECONE_INDEX_NAME Usage The PineconeVectorStore class exposes the connection to the Pinecone vector store. from langchain_pinecone import PineconeVectorStore embeddings = ... # use a LangChain Embeddings class vectorstore = PineconeVectorStore(embeddings=embeddings)
RPMPackage python3-langchain-perplexity-0.1.1-1.lbn36.noarch
langchain-perplexity
RPMPackage python3-langchain-openai-0.3.16-1.lbn36.noarch
langchain-openai This package contains the LangChain integrations for OpenAI through their openai SDK. Installation and Setup Install the LangChain partner package pip install langchain-openai Get an OpenAI api key and set it as an environment variable (OPENAI_API_KEY) LLM See a usage example. from langchain_openai import OpenAI If you are using a model hosted on Azure, you should use different wrapper for that: from langchain_openai import AzureOpenAI For a more detailed walkthrough of the Azure wrapper, see here Chat model from langchain_openai import ChatOpenAI from langchain_openai import AzureChatOpenAI Text Embedding Model See a usage example from langchain_openai import OpenAIEmbeddings from langchain_openai import AzureOpenAIEmbeddings
RPMPackage python3-langchain-ollama-0.3.2-1.lbn36.noarch
langchain-ollama This package contains the LangChain integration with Ollama Installation pip install -U langchain-ollama You will also need to run the Ollama server locally. You can download it here. Chat Models ChatOllama class exposes chat models from Ollama. from langchain_ollama import ChatOllama llm = ChatOllama(model="llama3-groq-tool-use") llm.invoke("Sing a ballad of LangChain.") Embeddings OllamaEmbeddings class exposes embeddings from Ollama. from langchain_ollama import OllamaEmbeddings embeddings = OllamaEmbeddings(model="llama3") embeddings.embed_query("What is the meaning of life?") LLMs OllamaLLM class exposes LLMs from Ollama. from langchain_ollama import OllamaLLM llm = OllamaLLM(model="llama3") llm.invoke("The meaning of life is")
RPMPackage python3-langchain-nvidia-ai-endpoints-0.3.8-1.lbn36.noarch
NVIDIA NIM Microservices The langchain-nvidia-ai-endpoints package contains LangChain integrations for chat models and embeddings powered by NVIDIA AI Foundation Models, and hosted on NVIDIA API Catalog. NVIDIA AI Foundation modelsĀ are community and NVIDIA-built models and are NVIDIA-optimized to deliver the best performance on NVIDIA accelerated infrastructure.Ā  Using the API, you can query live endpoints available on theĀ NVIDIA API CatalogĀ to get quick results from a DGX-hosted cloud compute environment. All models are source-accessible and can be deployed on your own compute cluster using NVIDIA NIMā„¢ microservices which is part of NVIDIA AI Enterprise. Models can be exported from NVIDIA’s API catalog with NVIDIA NIM, which is included with the NVIDIA AI Enterprise license, and run them on-premises, giving Enterprises ownership of their customizations and full control of their IP and AI application. NIM microservices are packaged as container images on a per model/model family basis
RPMPackage python3-langchain-mongodb-0.6.1-1.lbn36.noarch
langchain-mongodb Installation pip install -U langchain-mongodb Usage See Getting Started with the LangChain Integration for a walkthrough on using your first LangChain implementation with MongoDB Atlas. Using MongoDBAtlasVectorSearch from langchain_mongodb import MongoDBAtlasVectorSearch MONGODB_ATLAS_CLUSTER_URI = os.environ.get("MONGODB_ATLAS_CLUSTER_URI") DB_NAME = "langchain_db" COLLECTION_NAME = "test" ATLAS_VECTOR_SEARCH_INDEX_NAME = "index_name" MONGODB_COLLECTION = client[DB_NAME][COLLECTION_NAME] vector_search = MongoDBAtlasVectorSearch.from_connection_string( MONGODB_ATLAS_CLUSTER_URI, DB_NAME + "." + COLLECTION_NAME, OpenAIEmbeddings(disallowed_special=()), index_name=ATLAS_VECTOR_SEARCH_INDEX_NAME, ) client = MongoClient(MONGODB_ATLAS_CLUSTER_URI) vector_search_2 =
RPMPackage python3-langchain-mistralai-0.2.10-1.lbn36.noarch
langchain-mistralai This package contains the LangChain integrations for MistralAI through their mistralai SDK. Installation pip install -U langchain-mistralai Chat Models This package contains the ChatMistralAI class, which is the recommended way to interface with MistralAI models. To use, install the requirements, and configure your environment. export MISTRAL_API_KEY=your-api-key Then initialize from langchain_core.messages import HumanMessage from langchain_mistralai.chat_models import ChatMistralAI chat = ChatMistralAI(model="mistral-small") messages = [HumanMessage(content="say a brief hello")] chat.invoke(messages) ChatMistralAI also supports async and streaming functionality: await chat.ainvoke(messages) for chunk in chat.stream(messages): print(chunk.content, end="", flush=True) Embeddings With MistralAIEmbeddings, you can directly use the default model 'mistral-embed', or set a different one if available. Choose model embedding.mode
RPMPackage python3-langchain-milvus-0.1.7-1.lbn36.noarch
langchain-milvus This is a library integration with Milvus and Zilliz Cloud. Installation pip install -U langchain-milvus Milvus vector database See a usage example from langchain_milvus import Milvus Milvus hybrid search See a usage example. from langchain_milvus import MilvusCollectionHybridSearchRetriever Zilliz Cloud vector database from langchain_milvus import Zilliz Zilliz Cloud Pipeline Retriever from langchain_milvus import ZillizCloudPipelineRetriever
RPMPackage python3-langchain-ibm-0.3.8-1.lbn36.noarch
langchain-ibm This package provides the integration between LangChain and IBM watsonx.ai through the ibm-watsonx-ai SDK. Installation To use the langchain-ibm package, follow these installation steps: pip install langchain-ibm Usage Setting up To use IBM's models, you must have an IBM Cloud user API key. Here's how to obtain and set up your API key: Obtain an API Key: For more details on how to create and manage an API key, refer to IBM's documentation. Set the API Key as an Environment Variable: For security reasons, it's recommended to not hard-code your API key directly in your scripts. Instead, set it up as an environment variable. You can use the following code to prompt for the API key and set it as an environment variable: import os from getpass import getpass watsonx_api_key = getpass() os.environ["WATSONX_APIKEY"] = watsonx_api_key In alternative, you can set the environment variable in your terminal. Linux/macOS: Open your terminal and execute the following command: ex
RPMPackage python3-langchain-groq-0.3.2-1.lbn36.noarch
langchain-groq Welcome to Groq! šŸš€ At Groq, we've developed the world's first Language Processing Unitā„¢, or LPU. The Groq LPU has a deterministic, single core streaming architecture that sets the standard for GenAI inference speed with predictable and repeatable performance for any given workload. Beyond the architecture, our software is designed to empower developers like you with the tools you need to create innovative, powerful AI applications. With Groq as your engine, you can: Achieve uncompromised low latency and performance for real-time AI and HPC inferences šŸ”„ Know the exact performance and compute time for any given workload šŸ”® Take advantage of our cutting-edge technology to stay ahead of the competition šŸ’Ŗ Want more Groq? Check out our website for more resources and join our Discord community to connect with our developers!
RPMPackage python3-langchain-graph-retriever-0.6.1-1.lbn36.noarch
LangChain Graph Retriever LangChain Graph Retriever is a Python library that supports traversing a document graph on top of vector-based similarity search. It works seamlessly with LangChain's retriever framework and supports various graph traversal strategies for efficient document discovery. Features Vector Search: Perform similarity searches using vector embeddings. Graph Traversal: Apply traversal strategies such as breadth-first (Eager) or Maximal Marginal Relevance (MMR) to explore document relationships. Customizable Strategies: Easily extend and configure traversal strategies to meet your specific use case. Multiple Adapters: Support for various vector stores, including AstraDB, Cassandra, Chroma, OpenSearch, and in-memory storage. Synchronous and Asynchronous Retrieval: Supports both sync and async workflows for flexibility in different applications. Installation Install the library via pip: pip install langchain-graph-retriever Getting Started Here is an example of how to
RPMPackage python3-langchain-google-vertexai-2.0.7-1.lbn36.noarch
langchain-google-vertexai This package contains the LangChain integrations for Google Cloud generative models. Installation pip install -U langchain-google-vertexai Chat Models ChatVertexAI class exposes models such as gemini-pro and chat-bison. To use, you should have Google Cloud project with APIs enabled, and configured credentials. Initialize the model as: from langchain_google_vertexai import ChatVertexAI llm = ChatVertexAI(model_name="gemini-pro") llm.invoke("Sing a ballad of LangChain.") You can use other models, e.g. chat-bison: from langchain_google_vertexai import ChatVertexAI llm = ChatVertexAI(model_name="chat-bison", temperature=0.3) llm.invoke("Sing a ballad of LangChain.") Multimodal inputs Gemini vision model supports image inputs when providing a single chat message. Example: from langchain_core.messages import HumanMessage from langchain_google_vertexai import ChatVertexAI llm = ChatVertexAI(model_name="gemini-pro-vision") message = HumanMessage( c
RPMPackage python3-langchain-google-community-2.0.3-1.lbn36.noarch
langchain-google-community This package contains the LangChain integrations for Google products that are not part of langchain-google-vertexai or langchain-google-genai packages. Installation pip install -U langchain-google-community
RPMPackage python3-langchain-google-calendar-tools-0.0.1-1.lbn36.noarch
Langchain Google Calendar Tools This repo walks through connecting to the Google Calendar API. Installation pip install langchain-google-calendar-tools For local development: pip install -e . How to use Create a Google Cloud project and enable Google Calendar API. To get Oauth credentials for the Desktop app, please refer https:/developers.google.com/calendar/api/guides/overview for detail. Download the credentials file to ./notebooks and rename it to credentials.json. If you want to keep its original file name, please replace the value of client_secrets_file in demo.ipynb with the valid path which points to the credentials file. Run this notebook to perform the listed functions Limitations Due to the short development time, some of the following parts have not been completed and will be improved in the future: Timezone: Currently being fixed to Vietnam's timezone, it will be taken from the user's Calendar or the system in the future Update recurring events: has not been implemen