Update godeps
This commit is contained in:
parent
423433bc5f
commit
701c5a0e30
482 changed files with 86915 additions and 19741 deletions
75
vendor/k8s.io/kubernetes/pkg/kubelet/qos/memory_policy.go
generated
vendored
75
vendor/k8s.io/kubernetes/pkg/kubelet/qos/memory_policy.go
generated
vendored
|
|
@ -1,75 +0,0 @@
|
|||
/*
|
||||
Copyright 2015 The Kubernetes Authors All rights reserved.
|
||||
|
||||
Licensed under the Apache License, Version 2.0 (the "License");
|
||||
you may not use this file except in compliance with the License.
|
||||
You may obtain a copy of the License at
|
||||
|
||||
http://www.apache.org/licenses/LICENSE-2.0
|
||||
|
||||
Unless required by applicable law or agreed to in writing, software
|
||||
distributed under the License is distributed on an "AS IS" BASIS,
|
||||
WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
|
||||
See the License for the specific language governing permissions and
|
||||
limitations under the License.
|
||||
*/
|
||||
|
||||
package qos
|
||||
|
||||
import (
|
||||
"k8s.io/kubernetes/pkg/api"
|
||||
)
|
||||
|
||||
const (
|
||||
PodInfraOOMAdj int = -999
|
||||
KubeletOOMScoreAdj int = -999
|
||||
KubeProxyOOMScoreAdj int = -999
|
||||
)
|
||||
|
||||
// isMemoryBestEffort returns true if the container's memory requirements are best-effort.
|
||||
func isMemoryBestEffort(container *api.Container) bool {
|
||||
// A container is memory best-effort if its memory request is unspecified or 0.
|
||||
// If a request is specified, then the user expects some kind of resource guarantee.
|
||||
return container.Resources.Requests.Memory().Value() == 0
|
||||
}
|
||||
|
||||
// isMemoryGuaranteed returns true if the container's memory requirements are Guaranteed.
|
||||
func isMemoryGuaranteed(container *api.Container) bool {
|
||||
// A container is memory guaranteed if its memory request == memory limit.
|
||||
// If memory request == memory limit, the user is very confident of resource consumption.
|
||||
memoryRequest := container.Resources.Requests.Memory()
|
||||
memoryLimit := container.Resources.Limits.Memory()
|
||||
return (*memoryRequest).Cmp(*memoryLimit) == 0 && memoryRequest.Value() != 0
|
||||
}
|
||||
|
||||
// GetContainerOOMAdjust returns the amount by which the OOM score of all processes in the
|
||||
// container should be adjusted. The OOM score of a process is the percentage of memory it consumes
|
||||
// multiplied by 10 (barring exceptional cases) + a configurable quantity which is between -1000
|
||||
// and 1000. Containers with higher OOM scores are killed if the system runs out of memory.
|
||||
// See https://lwn.net/Articles/391222/ for more information.
|
||||
func GetContainerOOMScoreAdjust(container *api.Container, memoryCapacity int64) int {
|
||||
if isMemoryGuaranteed(container) {
|
||||
// Memory guaranteed containers should be the last to get killed.
|
||||
return -999
|
||||
} else if isMemoryBestEffort(container) {
|
||||
// Memory best-effort containers should be the first to be killed.
|
||||
return 1000
|
||||
} else {
|
||||
// Burstable containers are a middle tier, between Guaranteed and Best-Effort. Ideally,
|
||||
// we want to protect Burstable containers that consume less memory than requested.
|
||||
// The formula below is a heuristic. A container requesting for 10% of a system's
|
||||
// memory will have an oom score adjust of 900. If a process in container Y
|
||||
// uses over 10% of memory, its OOM score will be 1000. The idea is that containers
|
||||
// which use more than their request will have an OOM score of 1000 and will be prime
|
||||
// targets for OOM kills.
|
||||
// Note that this is a heuristic, it won't work if a container has many small processes.
|
||||
memoryRequest := container.Resources.Requests.Memory().Value()
|
||||
oomScoreAdjust := 1000 - (1000*memoryRequest)/memoryCapacity
|
||||
// A memory guaranteed container using 100% of memory can have an OOM score of 1. Ensure
|
||||
// that memory burstable containers have a higher OOM score.
|
||||
if oomScoreAdjust < 2 {
|
||||
return 2
|
||||
}
|
||||
return int(oomScoreAdjust)
|
||||
}
|
||||
}
|
||||
67
vendor/k8s.io/kubernetes/pkg/kubelet/qos/policy.go
generated
vendored
Normal file
67
vendor/k8s.io/kubernetes/pkg/kubelet/qos/policy.go
generated
vendored
Normal file
|
|
@ -0,0 +1,67 @@
|
|||
/*
|
||||
Copyright 2015 The Kubernetes Authors All rights reserved.
|
||||
|
||||
Licensed under the Apache License, Version 2.0 (the "License");
|
||||
you may not use this file except in compliance with the License.
|
||||
You may obtain a copy of the License at
|
||||
|
||||
http://www.apache.org/licenses/LICENSE-2.0
|
||||
|
||||
Unless required by applicable law or agreed to in writing, software
|
||||
distributed under the License is distributed on an "AS IS" BASIS,
|
||||
WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
|
||||
See the License for the specific language governing permissions and
|
||||
limitations under the License.
|
||||
*/
|
||||
|
||||
package qos
|
||||
|
||||
import (
|
||||
"k8s.io/kubernetes/pkg/api"
|
||||
"k8s.io/kubernetes/pkg/kubelet/qos/util"
|
||||
)
|
||||
|
||||
const (
|
||||
PodInfraOOMAdj int = -999
|
||||
KubeletOOMScoreAdj int = -999
|
||||
KubeProxyOOMScoreAdj int = -999
|
||||
guaranteedOOMScoreAdj int = -998
|
||||
besteffortOOMScoreAdj int = 1000
|
||||
)
|
||||
|
||||
// GetContainerOOMAdjust returns the amount by which the OOM score of all processes in the
|
||||
// container should be adjusted.
|
||||
// The OOM score of a process is the percentage of memory it consumes
|
||||
// multiplied by 10 (barring exceptional cases) + a configurable quantity which is between -1000
|
||||
// and 1000. Containers with higher OOM scores are killed if the system runs out of memory.
|
||||
// See https://lwn.net/Articles/391222/ for more information.
|
||||
func GetContainerOOMScoreAdjust(pod *api.Pod, container *api.Container, memoryCapacity int64) int {
|
||||
switch util.GetPodQos(pod) {
|
||||
case util.Guaranteed:
|
||||
// Guaranteed containers should be the last to get killed.
|
||||
return guaranteedOOMScoreAdj
|
||||
case util.BestEffort:
|
||||
return besteffortOOMScoreAdj
|
||||
}
|
||||
|
||||
// Burstable containers are a middle tier, between Guaranteed and Best-Effort. Ideally,
|
||||
// we want to protect Burstable containers that consume less memory than requested.
|
||||
// The formula below is a heuristic. A container requesting for 10% of a system's
|
||||
// memory will have an OOM score adjust of 900. If a process in container Y
|
||||
// uses over 10% of memory, its OOM score will be 1000. The idea is that containers
|
||||
// which use more than their request will have an OOM score of 1000 and will be prime
|
||||
// targets for OOM kills.
|
||||
// Note that this is a heuristic, it won't work if a container has many small processes.
|
||||
memoryRequest := container.Resources.Requests.Memory().Value()
|
||||
oomScoreAdjust := 1000 - (1000*memoryRequest)/memoryCapacity
|
||||
// A guaranteed pod using 100% of memory can have an OOM score of 1. Ensure
|
||||
// that burstable pods have a higher OOM score adjustment.
|
||||
if oomScoreAdjust < 2 {
|
||||
return 2
|
||||
}
|
||||
// Give burstable pods a higher chance of survival over besteffort pods.
|
||||
if int(oomScoreAdjust) == besteffortOOMScoreAdj {
|
||||
return int(oomScoreAdjust - 1)
|
||||
}
|
||||
return int(oomScoreAdjust)
|
||||
}
|
||||
68
vendor/k8s.io/kubernetes/pkg/kubelet/qos/util/qos.go
generated
vendored
68
vendor/k8s.io/kubernetes/pkg/kubelet/qos/util/qos.go
generated
vendored
|
|
@ -18,7 +18,7 @@ package util
|
|||
|
||||
import (
|
||||
"k8s.io/kubernetes/pkg/api"
|
||||
"k8s.io/kubernetes/pkg/util/sets"
|
||||
"k8s.io/kubernetes/pkg/api/resource"
|
||||
)
|
||||
|
||||
const (
|
||||
|
|
@ -48,27 +48,69 @@ func isResourceBestEffort(container *api.Container, resource api.ResourceName) b
|
|||
}
|
||||
|
||||
// GetPodQos returns the QoS class of a pod.
|
||||
// The QoS class of a pod is the lowest QoS class for each resource in each container.
|
||||
// A pod is besteffort if none of its containers have specified any requests or limits.
|
||||
// A pod is guaranteed only when requests and limits are specified for all the containers and they are equal.
|
||||
// A pod is burstable if limits and requests do not match across all containers.
|
||||
func GetPodQos(pod *api.Pod) string {
|
||||
qosValues := sets.NewString()
|
||||
requests := api.ResourceList{}
|
||||
limits := api.ResourceList{}
|
||||
zeroQuantity := resource.MustParse("0")
|
||||
isGuaranteed := true
|
||||
for _, container := range pod.Spec.Containers {
|
||||
qosPerResource := GetQoS(&container)
|
||||
for _, qosValue := range qosPerResource {
|
||||
qosValues.Insert(qosValue)
|
||||
// process requests
|
||||
for name, quantity := range container.Resources.Requests {
|
||||
if quantity.Cmp(zeroQuantity) == 1 {
|
||||
delta := quantity.Copy()
|
||||
if _, exists := requests[name]; !exists {
|
||||
requests[name] = *delta
|
||||
} else {
|
||||
delta.Add(requests[name])
|
||||
requests[name] = *delta
|
||||
}
|
||||
}
|
||||
}
|
||||
// process limits
|
||||
for name, quantity := range container.Resources.Limits {
|
||||
if quantity.Cmp(zeroQuantity) == 1 {
|
||||
delta := quantity.Copy()
|
||||
if _, exists := limits[name]; !exists {
|
||||
limits[name] = *delta
|
||||
} else {
|
||||
delta.Add(limits[name])
|
||||
limits[name] = *delta
|
||||
}
|
||||
}
|
||||
}
|
||||
if len(container.Resources.Limits) != len(supportedComputeResources) {
|
||||
isGuaranteed = false
|
||||
}
|
||||
}
|
||||
if qosValues.Has(BestEffort) {
|
||||
if len(requests) == 0 && len(limits) == 0 {
|
||||
return BestEffort
|
||||
}
|
||||
if qosValues.Has(Burstable) {
|
||||
return Burstable
|
||||
// Check is requests match limits for all resources.
|
||||
if isGuaranteed {
|
||||
for name, req := range requests {
|
||||
if lim, exists := limits[name]; !exists || lim.Cmp(req) != 0 {
|
||||
isGuaranteed = false
|
||||
break
|
||||
}
|
||||
}
|
||||
}
|
||||
return Guaranteed
|
||||
if isGuaranteed &&
|
||||
len(requests) == len(limits) &&
|
||||
len(limits) == len(supportedComputeResources) {
|
||||
return Guaranteed
|
||||
}
|
||||
return Burstable
|
||||
}
|
||||
|
||||
// GetQos returns a mapping of resource name to QoS class of a container
|
||||
func GetQoS(container *api.Container) map[api.ResourceName]string {
|
||||
resourceToQoS := map[api.ResourceName]string{}
|
||||
// QoSList is a set of (resource name, QoS class) pairs.
|
||||
type QoSList map[api.ResourceName]string
|
||||
|
||||
// GetQoS returns a mapping of resource name to QoS class of a container
|
||||
func GetQoS(container *api.Container) QoSList {
|
||||
resourceToQoS := QoSList{}
|
||||
for resource := range allResources(container) {
|
||||
switch {
|
||||
case isResourceGuaranteed(container, resource):
|
||||
|
|
|
|||
Loading…
Add table
Add a link
Reference in a new issue