Personal tools
Skip to content. | Skip to navigation
libibverbs is a library that allows userspace processes to use RDMA "verbs" as described in the InfiniBand Architecture Specification and the RDMA Protocol Verbs Specification. This includes direct hardware access from userspace to InfiniBand/iWARP adapters (kernel bypass) for fast path operations. Device-specific plug-in ibverbs userspace drivers are included: - libcxgb4: Chelsio T4 iWARP HCA - libefa: Amazon Elastic Fabric Adapter - libhfi1: Intel Omni-Path HFI - libhns: HiSilicon Hip06 SoC - libipathverbs: QLogic InfiniPath HCA - libirdma: Intel Ethernet Connection RDMA - libmana: Microsoft Azure Network Adapter - libmlx4: Mellanox ConnectX-3 InfiniBand HCA - libmlx5: Mellanox Connect-IB/X-4+ InfiniBand HCA - libmthca: Mellanox InfiniBand HCA - libocrdma: Emulex OneConnect RDMA/RoCE Device - libqedr: QLogic QL4xxx RoCE HCA - librxe: A software implementation of the RoCE protocol - libsiw: A software implementation of the iWarp protocol - libvmw_pvrdma: VMware paravirtual RDMA device
This extension index Plone content into ElasticSearch. This doesn’t replace the Plone catalog with ElasticSearch, nor interact with the Plone catalog at all, it merely index content inside ElasticSearch when it is modified or published. In addition to this, it provides a simple search page called search.html that queries ElasticSearch using Javascript (so Plone is not involved in searching) and propose the same features than the default Plone search page. A search portlet let you redirect people to this new search page as well. This extension have been built for Plone 4, but might work with Plone 3. Usage After adding this extension to your buildout (including the zcml), you can install the extension in Plone. A configuration screen is available inside site setup. It will let you configure the URLs of the ElasticSearch servers to use in order to index, and search. To proceed: Fill in the ElasticSearch settings, Click on Save, Click on Create Index in order to create the ElasticSearch index, Click on Import site content in order to index already existing content in ElasticSearch. You can use the same ElasticSearch server (and probably index) for multiple Plone sites, creating a federated search that way. Security disclaimer By default is no authentication or access validation while searching or indexing content. The original purpose of this search is to be public. If you have private content that you don’t want to be searchable or viewable by unauthorized people, please be sure to check the checkbox index only published content in the configuration screen. In addition to this ElasticSearch is not secured by default, meaning there is no authentication to provide in order to index or look-up content. Be sure to hide it behind a firewall and use a proxy or Apache in order to restrict the requests made to it: you only need to allow access via POST to the sub-URL _search after the index name configured in the configuration screen. For instance, if the index name is plone, you shall allow only requests to http://your-public-es-url/plone/_search. After you configured your proxy, be sure to configure its public URL, like http://your-public-es-url in the configuration screen so the search page knows how to contact it. However if you want to allow users to search though restricted and not yet published content, you can check index security and uncheck index only published content in the configuration screen. After reindexing your content, if you check proxy search requests though Plone and apply security filter, search will work on restricted and not yet published content, but will be slower as the queries will be proxied though Plone.
Boto is a Python package that provides interfaces to Amazon Web Services. It supports over thirty services, such as S3 (Simple Storage Service), SQS (Simple Queue Service), and EC2 (Elastic Compute Cloud) via their REST and Query APIs. The goal of boto is to support the full breadth and depth of Amazon Web Services. In addition, boto provides support for other public services such as Google Storage in addition to private cloud systems like Eucalyptus, OpenStack and Open Nebula.
langchain-elasticsearch This package contains the LangChain integration with Elasticsearch. Installation pip install -U langchain-elasticsearch Elasticsearch setup Elastic Cloud You need a running Elasticsearch deployment. The easiest way to start one is through Elastic Cloud. You can sign up for a free trial. Create a deployment Get your Cloud ID: In the Elastic Cloud console, click "Manage" next to your deployment Copy the Cloud ID and paste it into the es_cloud_id parameter below Create an API key: In the Elastic Cloud console, click "Open" next to your deployment In the left-hand side menu, go to "Stack Management", then to "API Keys" Click "Create API key" Enter a name for the API key and click "Create" Copy the API key and paste it into the es_api_key parameter below Alternatively, you can run Elasticsearch via Docker as described in the docs. Usage ElasticsearchStore The ElasticsearchStore class exposes Elasticsearch as a vector store. from langchain_elasticsearch impor
Beats - Lightweight shippers for Elasticsearch & Logstash The Beats are lightweight processes, written in Go, that you install on your servers to capture all sorts of operational data like logs, operating system metrics or network packet data, and to send it to Elasticsearch, either directly or via Logstash, so it can be visualized with Kibana.
English | [简体中文](README-CN.md)  to install pip.bash Install the alibabacloud_ecs20140526 pip install alibabacloud_ecs20140526 Issues[Opening an Issue]( Issues not...
The eslint config used by the kibana team
elasticsearch datemath parser, used in kibana
Metricbeat fetches a set of metrics on a predefined interval from the operating system and services such as Apache web server, Redis, and more.
Dockbeat is the new Dockerbeat name. We had to rename the project due to the Docker trademarking policy. Dockbeat is a Beat used for docker daemon monitoring. It is a lightweight agent that installed on your servers, reads periodically docker container statistics and indexes them in Elasticsearch. Exported document types There are five types of documents exported: type: container: container attributes type: cpu: container CPU usage statistics. One document per container is generated. type: net: container network statistics. One document per network container is generated. type: memory: container memory statistics. One document per container is generated. type: blkio: container io access statistics. One document per container is generated. type: log: dockbeat status information. One document per tick is generated if an error occurred.