Skip to content

Commit

Permalink
Merge pull request #1478 from jeruane/jruane-patch-4
Browse files Browse the repository at this point in the history
ec2_instance_types CA cherry picks to 1.2
  • Loading branch information
k8s-ci-robot authored Dec 6, 2018
2 parents 0747bca + ce9da03 commit dc71b12
Show file tree
Hide file tree
Showing 2 changed files with 104 additions and 1 deletion.
3 changes: 2 additions & 1 deletion cluster-autoscaler/cloudprovider/aws/aws_manager_test.go
Original file line number Diff line number Diff line change
Expand Up @@ -18,7 +18,6 @@ package aws

import (
"fmt"
"strings"
"testing"

"github.com/stretchr/testify/mock"
Expand Down Expand Up @@ -171,6 +170,7 @@ func TestFetchExplicitAsgs(t *testing.T) {
validateAsg(t, asgs[0].config, groupname, min, max)
}

/* Disabled due to flakiness. See https://github.com/kubernetes/autoscaler/issues/608
func TestFetchAutoAsgs(t *testing.T) {
min, max := 1, 10
groupname, tags := "coolasg", []string{"tag", "anothertag"}
Expand Down Expand Up @@ -244,3 +244,4 @@ func TestFetchAutoAsgs(t *testing.T) {
assert.NoError(t, err)
assert.Empty(t, m.asgCache.get())
}
*/
102 changes: 102 additions & 0 deletions cluster-autoscaler/cloudprovider/aws/ec2_instance_types.go
Original file line number Diff line number Diff line change
Expand Up @@ -255,6 +255,12 @@ var InstanceTypes = map[string]*instanceType{
MemoryMb: 124928,
GPU: 0,
},
"f1.4xlarge": {
InstanceType: "f1.4xlarge",
VCPU: 16,
MemoryMb: 249856,
GPU: 0,
},
"g2": {
InstanceType: "g2",
VCPU: 32,
Expand Down Expand Up @@ -297,6 +303,12 @@ var InstanceTypes = map[string]*instanceType{
MemoryMb: 249856,
GPU: 2,
},
"g3s.xlarge": {
InstanceType: "g3s.xlarge",
VCPU: 4,
MemoryMb: 31232,
GPU: 0,
},
"h1": {
InstanceType: "h1",
VCPU: 64,
Expand Down Expand Up @@ -567,6 +579,42 @@ var InstanceTypes = map[string]*instanceType{
MemoryMb: 16384,
GPU: 0,
},
"m5a.12xlarge": {
InstanceType: "m5a.12xlarge",
VCPU: 48,
MemoryMb: 196608,
GPU: 0,
},
"m5a.24xlarge": {
InstanceType: "m5a.24xlarge",
VCPU: 96,
MemoryMb: 393216,
GPU: 0,
},
"m5a.2xlarge": {
InstanceType: "m5a.2xlarge",
VCPU: 8,
MemoryMb: 32768,
GPU: 0,
},
"m5a.4xlarge": {
InstanceType: "m5a.4xlarge",
VCPU: 16,
MemoryMb: 65536,
GPU: 0,
},
"m5a.large": {
InstanceType: "m5a.large",
VCPU: 2,
MemoryMb: 8192,
GPU: 0,
},
"m5a.xlarge": {
InstanceType: "m5a.xlarge",
VCPU: 4,
MemoryMb: 16384,
GPU: 0,
},
"m5d": {
InstanceType: "m5d",
VCPU: 96,
Expand Down Expand Up @@ -777,6 +825,42 @@ var InstanceTypes = map[string]*instanceType{
MemoryMb: 32768,
GPU: 0,
},
"r5a.12xlarge": {
InstanceType: "r5a.12xlarge",
VCPU: 48,
MemoryMb: 393216,
GPU: 0,
},
"r5a.24xlarge": {
InstanceType: "r5a.24xlarge",
VCPU: 96,
MemoryMb: 786432,
GPU: 0,
},
"r5a.2xlarge": {
InstanceType: "r5a.2xlarge",
VCPU: 8,
MemoryMb: 65536,
GPU: 0,
},
"r5a.4xlarge": {
InstanceType: "r5a.4xlarge",
VCPU: 16,
MemoryMb: 131072,
GPU: 0,
},
"r5a.large": {
InstanceType: "r5a.large",
VCPU: 2,
MemoryMb: 16384,
GPU: 0,
},
"r5a.xlarge": {
InstanceType: "r5a.xlarge",
VCPU: 4,
MemoryMb: 32768,
GPU: 0,
},
"r5d": {
InstanceType: "r5d",
VCPU: 96,
Expand Down Expand Up @@ -909,6 +993,24 @@ var InstanceTypes = map[string]*instanceType{
MemoryMb: 16384,
GPU: 0,
},
"u-12tb1": {
InstanceType: "u-12tb1",
VCPU: 448,
MemoryMb: 12582912,
GPU: 0,
},
"u-6tb1": {
InstanceType: "u-6tb1",
VCPU: 448,
MemoryMb: 6291456,
GPU: 0,
},
"u-9tb1": {
InstanceType: "u-9tb1",
VCPU: 448,
MemoryMb: 9437184,
GPU: 0,
},
"x1": {
InstanceType: "x1",
VCPU: 128,
Expand Down

0 comments on commit dc71b12

Please sign in to comment.