You are here: Home

Modified items

All recently modified items, latest first.
RPMPackage invokebinder-1.2-5.fc25.noarch
This library hopes to provide a more friendly DSL for binding method handles. Unlike the normal MethodHandle API, handles are bound forward from a source MethodType and eventually adapted to a final target MethodHandle. Along the way the transformations are pushed onto a stack and eventually applied in reverse order, as the standard API demands.
RPMPackage hyperledger-fabric-node-sdk-0.5-1.0.0.git1116c74.lbn19.x86_64
Node SDK
RPMPackage hyperledger-fabric-0.5-1.0.0.git1116c74.lbn19.x86_64
The fabric is an implementation of blockchain technology, leveraging familiar and proven technologies. It is a modular architecture allowing pluggable implementations of various function. It features powerful container technology to host any mainstream language for smart contracts development.
RPMPackage highlight-3.13-1.fc19.x86_64
A utility that converts sourcecode to HTML, XHTML, RTF, LaTeX, TeX, XSL-FO, XML or ANSI escape sequences with syntax highlighting. It supports several programming and markup languages. Language descriptions are configurable and support regular expressions. The utility offers indentation and reformatting capabilities. It is easily possible to create new language definitions and colour themes.
RPMPackage hiera-3.0.1-2.lbn19.noarch
A simple hierarchical database supporting plugin data sources.
RPMPackage hfsplus-tools-540.1.linux3-4.fc19.x86_64
HFS+, HFS Plus, or Mac OS Extended are names for a file system developed by Apple Computer to replace their Hierarchical File System (HFS). In addition to being the default file system on modern Apple computers, HFS+ is one of two formats, FAT being the other, that are supported by the iPod hard-disk based music player. Unlike FAT, HFS+ supports UNIX style file permissions, which makes it useful, for serving and sharing files in a secured manner. As Apple Computer's devices and systems become increasingly ubiquitous, it becomes important that Linux fully support this format. This package provides tools to create and check HFS+ filesystems under Linux. The Linux kernel does not support writing to HFS+ journals, writing to a hfsplus partition is recommended only after disabling journaling; however, the kernel, as of version 2.6.16, supports case-sensitivity (also known as HFSX) commit.
RPMPackage haveged-1.9.1-3.lbn19.x86_64
A Linux entropy source using the HAVEGE algorithm Haveged is a user space entropy daemon which is not dependent upon the standard mechanisms for harvesting randomness for the system entropy pool. This is important in systems with high entropy needs or limited user interaction (e.g. headless servers). Haveged uses HAVEGE (HArdware Volatile Entropy Gathering and Expansion) to maintain a 1M pool of random bytes used to fill /dev/random whenever the supply of random bits in /dev/random falls below the low water mark of the device. The principle inputs to haveged are the sizes of the processor instruction and data caches used to setup the HAVEGE collector. The haveged default is a 4kb data cache and a 16kb instruction cache. On machines with a cpuid instruction, haveged will attempt to select appropriate values from internal tables.
RPMPackage hashicorp-vault-0.7.2-0.1.gitd28dd5a.lbn19.x86_64
Vault is a tool for securely accessing secrets. A secret is anything that you want to tightly control access to, such as API keys, passwords, certificates, and more. Vault provides a unified interface to any secret, while providing tight access control and recording a detailed audit log. A modern system requires access to a multitude of secrets: database credentials, API keys for external services, credentials for service-oriented architecture communication, etc. Understanding who is accessing what secrets is already very difficult and platform-specific. Adding on key rolling, secure storage, and detailed audit logs is almost impossible without a custom solution. This is where Vault steps in. The key features of Vault are: Secure Secret Storage: Arbitrary key/value secrets can be stored in Vault. Vault encrypts these secrets prior to writing them to persistent storage, so gaining access to the raw storage isnt enough to access your secrets. Vault can write to disk, Consul, and more. Dynamic Secrets: Vault can generate secrets on-demand for some systems, such as AWS or SQL databases. For example, when an application needs to access an S3 bucket, it asks Vault for credentials, and Vault will generate an AWS keypair with valid permissions on demand. After creating these dynamic secrets, Vault will also automatically revoke them after the lease is up. Data Encryption: Vault can encrypt and decrypt data without storing it. This allows security teams to define encryption parameters and developers to store encrypted data in a location such as SQL without having to design their own encryption methods. Leasing and Renewal: All secrets in Vault have a lease associated with it. At the end of the lease, Vault will automatically revoke that secret. Clients are able to renew leases via built-in renew APIs. Revocation: Vault has built-in support for secret revocation. Vault can revoke not only single secrets, but a tree of secrets, for example all secrets read by a specific user, or all secrets of a particular type. Revocation assists in key rolling as well as locking down systems in the case of an intrusion.
RPMPackage hashicorp-terraform-0.7.3-0.1.git0dd7c65.lbn19.x86_64
Terraform is a tool for building, changing, and versioning infrastructure safely and efficiently. Terraform can manage existing and popular service providers as well as custom in-house solutions. The key features of Terraform are: Infrastructure as Code: Infrastructure is described using a high-level configuration syntax. This allows a blueprint of your datacenter to be versioned and treated as you would any other code. Additionally, infrastructure can be shared and re-used. Execution Plans: Terraform has a "planning" step where it generates an execution plan. The execution plan shows what Terraform will do when you call apply. This lets you avoid any surprises when Terraform manipulates infrastructure. Resource Graph: Terraform builds a graph of all your resources, and parallelizes the creation and modification of any non-dependent resources. Because of this, Terraform builds infrastructure as efficiently as possible, and operators get insight into dependencies in their infrastructure. Change Automation: Complex changesets can be applied to your infrastructure with minimal human interaction. With the previously mentioned execution plan and resource graph, you know exactly what Terraform will change and in what order, avoiding many possible human errors. For more information, see the introduction section of the Terraform website.
RPMPackage hashicorp-serf-0.8.0-0.1.gitb9642a4.lbn19.x86_64
Serf is a decentralized solution for service discovery and orchestration that is lightweight, highly available, and fault tolerant. Serf runs on Linux, Mac OS X, and Windows. An efficient and lightweight gossip protocol is used to communicate with other nodes. Serf can detect node failures and notify the rest of the cluster. An event system is built on top of Serf, letting you use Serf's gossip protocol to propagate events such as deploys, configuration changes, etc. Serf is completely masterless with no single point of failure. Here are some example use cases of Serf, though there are many others: Discovering web servers and automatically adding them to a load balancer Organizing many memcached or redis nodes into a cluster, perhaps with something like twemproxy or maybe just configuring an application with the address of all the nodes Triggering web deploys using the event system built on top of Serf Propagating changes to configuration to relevant nodes. Updating DNS records to reflect cluster changes as they occur. Much, much more.
RPMPackage hashicorp-packer-0.10.2-2.0.1.lbn19.x86_64
Packer is a tool for building identical machine images for multiple platforms from a single source configuration. Packer is lightweight, runs on every major operating system, and is highly performant, creating machine images for multiple platforms in parallel. Packer comes out of the box with support for the following platforms: Amazon EC2 (AMI). Both EBS-backed and instance-store AMIs DigitalOcean Docker Google Compute Engine OpenStack Parallels QEMU. Both KVM and Xen images. VirtualBox VMware Support for other platforms can be added via plugins. After Packer is installed, create your first template, which tells Packer what platforms to build images for and how you want to build them. In our case, we'll create a simple AMI that has Redis pre-installed. Save this file as quick-start.json. Be sure to replace any credentials with your own. { "builders": [{ "type": "amazon-ebs", "access_key": "YOUR KEY HERE", "secret_key": "YOUR SECRET KEY HERE", "region": "us-east-1", "source_ami": "ami-de0d9eb7", "instance_type": "t1.micro", "ssh_username": "ubuntu", "ami_name": "packer-example {{timestamp}}" }] } Next, tell Packer to build the image: $ packer build quick-start.json ... Packer will build an AMI according to the "quick-start" template. The AMI will be available in your AWS account. To delete the AMI, you must manually delete it using the AWS console. Packer builds your images, it does not manage their lifecycle. Where they go, how they're run, etc. is up to you.
RPMPackage hashicorp-nomad-server-0.1.0-0.1.gitb9a3d1e.lbn19.noarch
Server agent for Nomad
RPMPackage hashicorp-nomad-client-0.1.0-0.1.gitb9a3d1e.lbn19.noarch
Client agent for Nomad
RPMPackage hashicorp-nomad-0.5.2-0.1.gitfccf115.lbn19.x86_64
Nomad is a cluster manager, designed for both long lived services and short lived batch processing workloads. Developers use a declarative job specification to submit work, and Nomad ensures constraints are satisfied and resource utilization is optimized by efficient task packing. Nomad supports all major operating systems and virtualized, containerized, or standalone applications. The key features of Nomad are: Docker Support: Jobs can specify tasks which are Docker containers. Nomad will automatically run the containers on clients which have Docker installed, scale up and down based on the number of instances request, and automatically recover from failures. Multi-Datacenter and Multi-Region Aware: Nomad is designed to be a global-scale scheduler. Multiple datacenters can be managed as part of a larger region, and jobs can be scheduled across datacenters if requested. Multiple regions join together and federate jobs making it easy to run jobs anywhere. Operationally Simple: Nomad runs as a single binary that can be either a client or server, and is completely self contained. Nomad does not require any external services for storage or coordination. This means Nomad combines the features of a resource manager and scheduler in a single system. Distributed and Highly-Available: Nomad servers cluster together and perform leader election and state replication to provide high availability in the face of failure. The Nomad scheduling engine is optimized for optimistic concurrency allowing all servers to make scheduling decisions to maximize throughput. HashiCorp Ecosystem: Nomad integrates with the entire HashiCorp ecosystem of tools. Along with all HashiCorp tools, Nomad is designed in the unix philosophy of doing something specific and doing it well. Nomad integrates with tools like Packer, Consul, and Terraform to support building artifacts, service discovery, monitoring and capacity management.
RPMPackage hashicorp-consul-webui-0.7.3-0.1.gita90bb8f.lbn19.noarch
Consul comes with support for a beautiful, functional web UI out of the box. This UI can be used for viewing all services and nodes, viewing all health checks and their current status, and for reading and setting key/value data. The UI automatically supports multi-datacenter. For ease of deployment, the UI is distributed as static HTML and JavaScript. You do not need a separate web server to run the web UI. The Consul agent itself can be configured to serve the UI. The UI is available at the /ui path on the same port as the HTTP API. By default this is http://localhost:8500/ui.
RPMPackage hashicorp-consul-0.6.4-1.lbn19.x86_64
Consul is a tool for service discovery and configuration. Consul is distributed, highly available, and extremely scalable. Consul provides several key features: Service Discovery - Consul makes it simple for services to register themselves and to discover other services via a DNS or HTTP interface. External services such as SaaS providers can be registered as well. Health Checking - Health Checking enables Consul to quickly alert operators about any issues in a cluster. The integration with service discovery prevents routing traffic to unhealthy hosts and enables service level circuit breakers. Key/Value Storage - A flexible key/value store enables storing dynamic configuration, feature flagging, coordination, leader election and more. The simple HTTP API makes it easy to use anywhere. Multi-Datacenter - Consul is built to be datacenter aware, and can support any number of regions without complex configuration.
RPMPackage hashicorp-atlas-0-0.5.git8e45a6c.lbn19.x86_64
Authenticating with username and password Atlas Go can automatically generate an API authentication token given a username and password. For example: client := atlas.DefaultClient() token, err := client.Login("username", "password") if err != nil { panic(err) } The Login function returns an API token that can be used to sign requests. This function also sets the Token parameter on the Atlas Client, so future requests are signed with this access token. If you have two-factor authentication enabled, you must manually generate an access token on the Atlas website.
RPMPackage hamcrest-1.3-1.fc19.noarch
Provides a library of matcher objects (also known as constraints or predicates) allowing 'match' rules to be defined declaratively, to be used in other frameworks. Typical scenarios include testing frameworks, mocking libraries and UI validation rules.
RPMPackage graylog-2.1.1-2.lbn19.noarch
A syslog processing system that stores received messages in an Elasticsearch database. When coupled with the graylog-web-interface, which provides a front-end web interface, will allow for powerful message analytics for a server network. Other information, including but not limited to user credentials, stream configurations, etc, are stored in MongoDB
RPMPackage grafana-4.1.1-1.lbn19.x86_64
Grafana is an open source, feature rich metrics dashboard and graph editor for Graphite, InfluxDB & OpenTSDB. Features Graphite Target Editor Graphite target expression parser Feature rich query composer Quickly add and edit functions & parameters Templated queries See it in action Graphing Fast rendering, even over large timespans Click and drag to zoom Multiple Y-axis, logarithmic scales Bars, Lines, Points Smart Y-axis formating Series toggles & color selector Legend values, and formatting options Grid thresholds, axis labels Annotations Any panel can be rendered to PNG (server side using phantomjs) Dashboards Create, edit, save & search dashboards Change column spans and row heights Drag and drop panels to rearrange Templating Scripted dashboards Dashboard playlists Time range controls Share snapshots publicly InfluxDB Use InfluxDB as a metric data source, annotation source Query editor with series and column typeahead, easy group by and function selection OpenTSDB Use as metric data source Query editor with metric name typeahead and tag filtering