Displaying 1 to 20 from 45 results

AutoSpotting - Saves up to 90% of AWS EC2 costs by automating the use of spot instances on existing AutoScaling groups

  •    Go

AutoSpotting is the leading open source spot market automation tool, optimized towards quick/easy/frictionless adoption of the EC2 spot market at any scale. It is usually set up to monitor existing long-running AutoScaling groups with minimal configuration changes(often just tagging them, but even that can be avoided by using their existing tags), yielding the usual 70%-90% Spot cost savings but in a better integrated, more cost effective and easier to adopt way than the alternative tools and solutions.

autospotting - Automatically replace AWS EC2 instances in AutoScaling groups with identically configured but up to 90% cheaper spot instances

  •    Go

A simple and easy to use tool designed to significantly lower your Amazon AWS costs by automating the use of spot instances. When installed and enabled on an existing on-demand AutoScaling group, AutoSpotting clones one of your on-demand instances from the group with a spot instance that is cheaper, at least as large (automatically considering memory, CPU cores and disk volumes) and configured identically to it. Once the new spot instance is ready, it is attached to the group and an on-demand instance is detached and terminated to keep the group at constant capacity.

pipeline - Pipeline enables developers to go from commit to scale in minutes by turning Kubernetes into a feature rich application platform integrating CI/CD, centralized logging, monitoring, enterprise-grade security and autoscaling

  •    Go

Banzai Pipeline, or simply Pipeline is a tabletop reef break located in Hawaii, Oahu's North Shore. The most famous and infamous reef in the universe is the benchmark by which all other waves are measured. Pipeline enables developers to go from commit to scale in minutes by turning Kubernetes into a feature rich application platform integrating CI/CD, centralized logging, monitoring, enterprise-grade security, cost management and autoscaling.

KEDA - Kubernetes-based Event Driven Autoscaling component. It provides event driven scale for any container running in Kubernetes

  •    Go

KEDA is a Kubernetes-based Event Driven Autoscaler. With KEDA, you can drive the scaling of any container in Kubernetes based on the number of events needing to be processed. KEDA works alongside standard Kubernetes components like the Horizontal Pod Autoscaler and can extend functionality without overwriting or duplication.



CloudInit.NET is a service like the linux CloudInit application but designed for windows. Useful for Amazon EC2 but could be used for other environments as well. The service will download a file when the server starts and execute the code as a powershell script. Cloud Init ...

escalator - Escalator is a batch or job optimized horizontal autoscaler for Kubernetes

  •    Go

It is designed for large batch or job based workloads that cannot be force-drained and moved when the cluster needs to scale down - Escalator will ensure pods have been completed on nodes before terminating them. It is also optimised for scaling up the cluster as fast as possible to ensure pods are not left in a pending state. The need for this autoscaler is derived from our own experiences with very large batch workloads being scheduled and the default autoscaler not scaling up the cluster fast enough. These workloads can't be force-drained by the default autoscaler and must complete before the node can be terminated.

kube-aws-autoscaler - Simple, elastic Kubernetes cluster autoscaler for AWS Auto Scaling Groups

  •    Python

Simple cluster autoscaler for AWS Auto Scaling Groups which sets the DesiredCapacity of one or more ASGs to the calculated number of nodes. Use at your own risk! This autoscaler was only tested with Kubernetes versions 1.5.2 to 1.7.7. There is no guarantee that it works in previous Kubernetes versions.

aws-sdk-perl - A community AWS SDK for Perl Programmers

  •    Perl

This project is attempting to build an entire AWS SDK from the information that is stored in other AWS SDKs. Other AWS SDKs have a "data-driven" approach, meaning that the definitions for the method calls are stored in a data structure describing input and output parameters. If you want to develop a feature, or contribute code in some way, you need a development setup. This is done by cloning the repo into a local directory.

terraform-ecs-autoscale-alb - ECS cluster with instance and service autoscaling configured and running behind an ALB with path based routing set up

  •    HCL

Amazon EC2 Container Service (ECS) is a highly scalable, fast, container management service that makes it easy to run, stop, and manage Docker containers on a cluster of EC2 instances (called container instances). The idea behind ECS is to create a cluster (a group of container instances managed by ECS), define what Docker containers we want to run (with configuration for each container) and ECS will take care of deploying those containers across the cluster, rolling out new versions and integrating with other AWS infrastructure/services.

ecs-ansible-packer-terraform-wordpress - Proof of concept: Install wordpress environment using ansible, packer, docker, terraform and AWS

  •    HCL

The goal is to setup a wordpress container on an ECS cluster using tools like terraform, packer and ansible. This wordpress will use RDS as a database. First up I built the docker image with Packer docker builder using local shell provisioner, ansible-local provisioner and docker post-processor to upload the image to private docker registry aka AWS ECR. This image contains wordpress files with apache2 as a webserver. In addition, it has an entrypoint that copies files to the EFS folder at the first time if necessary, then populate DB config and wpSalt through environment variables. Resulting from the build I get a docker image that can run apache2 with wordpress configuration. Also, I used vagrant to do some ansible tests.

waiter - Runs, manages, and autoscales web services on Mesos and Kubernetes

  •    Clojure

Waiter is a web service platform that runs, manages, and automatically scales services without requiring human intervention. Waiter Design is a good place to start to learn more.

aws-auto-scaling-custom-resource - Libraries, samples, and tools to help AWS customers onboard with custom resource auto scaling


In this aws-auto-scaling-custom-resource repository, we demonstrate how to set up automatic scaling for custom resources using AWS services. In this context, a custom resource is an object that allows you to introduce your own application or service to the automatic scaling features of AWS. The included AWS CloudFormation template launches a collection of AWS resources, including a new Amazon API Gateway endpoint. The API Gateway endpoint allows secure access to scalable resources in the application or service that you want automatic scaling to work with.

terraform-aws-autoscaling - Terraform module which creates Auto Scaling resources on AWS

  •    HCL

Terraform module which creates Auto Scaling resources on AWS. There are two ways to specify tags for auto-scaling group in this module - tags and tags_as_map. See examples/asg_ec2/main.tf for example.

aws-autoscaling-gitlab-runner - CloudFormation template to deploy a GitLab Runner with auto-scaling on AWS


This repository consists of an AWS CloudFormation template that may be used to deploy a GitLab runner with Docker executor and auto-scaling based on number of builds on AWS. The runners have a shared cache to speed up builds. Objects in the bucket are automatically expired after a configurable number of days — 0 means that cache objects will never expire.

kube-metrics-adapter - General purpose metrics adapter for Kubernetes HPA metrics

  •    Go

Kube Metrics Adapter is a general purpose metrics adapter for Kubernetes that can collect and serve custom and external metrics for Horizontal Pod Autoscaling. It discovers Horizontal Pod Autoscaling resources and starts to collect the requested metrics and stores them in memory. It's implemented using the custom-metrics-apiserver library.

terraform-aws-ec2-autoscale-group - Terraform module to provision Auto Scaling Group and Launch Template on AWS

  •    HCL

Terraform module to provision Auto Scaling Group and Launch Template on AWS. The module also creates AutoScaling Policies and CloudWatch Metric Alarms to monitor CPU utilization on the EC2 instances and scale the number of instance in the AutoScaling Group up or down. If you don't want to use the provided functionality, or want to provide your own policies, disable it by setting the variable autoscaling_policies_enabled to "false".

magento-cluster - Highly Available and Auto-scalable Magento Cluster

  •    VCL

More details can be found at the article Running Auto-Scalable Magento Platform in Containers. Click the Deploy button below, specify your email address within the widget, choose one of the Jelastic Public Cloud providers and press Install.

replicator - Automated Cluster and Job Scaling For HashiCorp Nomad

  •    Go

Replicator is a fast and highly concurrent Go daemon that provides dynamic scaling of Nomad jobs and worker nodes. Replicator job scaling policies are configured as meta parameters within the job specification. A job scaling policy allows scaling constraints to be defined per task-group. Currently supported scaling metrics are CPU and Memory; there are plans for additional metrics as well as different metric backends in the future. Details of configuring job scaling and other important information can be found on the Replicator Job Scaling wiki page.

We have large collection of open source products. Follow the tags from Tag Cloud >>

Open source products are scattered around the web. Please provide information about the open source projects you own / you use. Add Projects.