We also need to download a recent version of Kubernetes project (version v1.3.0 or later).įirst, we set up a cluster with Cluster Autoscaler turned on. For more information on project creation, please read our Getting Started Guide.
#Volcano box cannot connect to server manual#
For GKE please check the autoscaling section in cluster operations manual available here.īefore we begin, we need to have an active GCE project with Google Cloud Monitoring, Google Cloud Logging and Stackdriver enabled. The new Cluster Autoscaling feature together with Horizontal Pod Autoscaler can handle this for you automatically. Ideally, we would want the number of nodes in the cluster and the number of pods in deployment to dynamically adjust to the load to meet end user demand.
Imagine you have a 24/7 production service with a load that is variable in time, where it is very busy during the day in the US, and relatively low at night. To understand better where autoscaling would provide the most value, let’s start with an example. On Google Compute Engine (GCE) and Google Container Engine (GKE) (and coming soon on AWS), Kubernetes will automatically scale up your cluster as soon as you need it, and scale it back down to save you money when you don’t. But what happens when you build a service that is even more popular than you planned for, and run out of compute? In Kubernetes 1.3, we are proud to announce that we have a solution: autoscaling. Editor’s note: this post is part of a series of in-depth articles on what's new in Kubernetes 1.3Ĭustomers using Kubernetes respond to end user requests quickly and ship software faster than ever before.