mirror of
https://github.com/kubernetes-sigs/prometheus-adapter.git
synced 2026-04-06 01:38:10 +00:00
Add tests for the resource metrics provider
This adds basic tests for the resource metrics provider, and fixes up the default config slightly.
This commit is contained in:
parent
49287fecc9
commit
6c8f44623e
5 changed files with 336 additions and 10 deletions
|
|
@ -95,7 +95,11 @@ func DefaultConfig(rateInterval time.Duration, labelPrefix string) *MetricsDisco
|
||||||
ContainerQuery: fmt.Sprintf("sum(rate(container_cpu_usage_seconds_total{<<.LabelMatchers>>}[%s])) by (<<.GroupBy>>)", pmodel.Duration(rateInterval).String()),
|
ContainerQuery: fmt.Sprintf("sum(rate(container_cpu_usage_seconds_total{<<.LabelMatchers>>}[%s])) by (<<.GroupBy>>)", pmodel.Duration(rateInterval).String()),
|
||||||
NodeQuery: fmt.Sprintf("sum(rate(container_cpu_usage_seconds_total{<<.LabelMatchers>>, id='/'}[%s])) by (<<.GroupBy>>)", pmodel.Duration(rateInterval).String()),
|
NodeQuery: fmt.Sprintf("sum(rate(container_cpu_usage_seconds_total{<<.LabelMatchers>>, id='/'}[%s])) by (<<.GroupBy>>)", pmodel.Duration(rateInterval).String()),
|
||||||
Resources: ResourceMapping{
|
Resources: ResourceMapping{
|
||||||
Template: fmt.Sprintf("%s<<.Resource>>", labelPrefix),
|
Overrides: map[string]GroupResource{
|
||||||
|
"namespace": {Resource: "namespace"},
|
||||||
|
"pod_name": {Resource: "pod"},
|
||||||
|
"instance": {Resource: "node"},
|
||||||
|
},
|
||||||
},
|
},
|
||||||
ContainerLabel: fmt.Sprintf("%scontainer_name", labelPrefix),
|
ContainerLabel: fmt.Sprintf("%scontainer_name", labelPrefix),
|
||||||
},
|
},
|
||||||
|
|
@ -103,7 +107,11 @@ func DefaultConfig(rateInterval time.Duration, labelPrefix string) *MetricsDisco
|
||||||
ContainerQuery: "sum(container_memory_working_set_bytes{<<.LabelMatchers>>}) by (<<.GroupBy>>)",
|
ContainerQuery: "sum(container_memory_working_set_bytes{<<.LabelMatchers>>}) by (<<.GroupBy>>)",
|
||||||
NodeQuery: "sum(container_memory_working_set_bytes{<<.LabelMatchers>>,id='/'}) by (<<.GroupBy>>)",
|
NodeQuery: "sum(container_memory_working_set_bytes{<<.LabelMatchers>>,id='/'}) by (<<.GroupBy>>)",
|
||||||
Resources: ResourceMapping{
|
Resources: ResourceMapping{
|
||||||
Template: fmt.Sprintf("%s<<.Resource>>", labelPrefix),
|
Overrides: map[string]GroupResource{
|
||||||
|
"namespace": {Resource: "namespace"},
|
||||||
|
"pod_name": {Resource: "pod"},
|
||||||
|
"instance": {Resource: "node"},
|
||||||
|
},
|
||||||
},
|
},
|
||||||
ContainerLabel: fmt.Sprintf("%scontainer_name", labelPrefix),
|
ContainerLabel: fmt.Sprintf("%scontainer_name", labelPrefix),
|
||||||
},
|
},
|
||||||
|
|
|
||||||
|
|
@ -17,8 +17,7 @@ data:
|
||||||
name:
|
name:
|
||||||
matches: ^container_(.*)_seconds_total$
|
matches: ^container_(.*)_seconds_total$
|
||||||
as: ""
|
as: ""
|
||||||
metricsQuery: sum(rate(<<.Series>>{<<.LabelMatchers>>,container_name!="POD"}[5m]))
|
metricsQuery: sum(rate(<<.Series>>{<<.LabelMatchers>>,container_name!="POD"}[1m])) by (<<.GroupBy>>)
|
||||||
by (<<.GroupBy>>)
|
|
||||||
- seriesQuery: '{__name__=~"^container_.*",container_name!="POD",namespace!="",pod_name!=""}'
|
- seriesQuery: '{__name__=~"^container_.*",container_name!="POD",namespace!="",pod_name!=""}'
|
||||||
seriesFilters:
|
seriesFilters:
|
||||||
- isNot: ^container_.*_seconds_total$
|
- isNot: ^container_.*_seconds_total$
|
||||||
|
|
@ -31,8 +30,7 @@ data:
|
||||||
name:
|
name:
|
||||||
matches: ^container_(.*)_total$
|
matches: ^container_(.*)_total$
|
||||||
as: ""
|
as: ""
|
||||||
metricsQuery: sum(rate(<<.Series>>{<<.LabelMatchers>>,container_name!="POD"}[5m]))
|
metricsQuery: sum(rate(<<.Series>>{<<.LabelMatchers>>,container_name!="POD"}[1m])) by (<<.GroupBy>>)
|
||||||
by (<<.GroupBy>>)
|
|
||||||
- seriesQuery: '{__name__=~"^container_.*",container_name!="POD",namespace!="",pod_name!=""}'
|
- seriesQuery: '{__name__=~"^container_.*",container_name!="POD",namespace!="",pod_name!=""}'
|
||||||
seriesFilters:
|
seriesFilters:
|
||||||
- isNot: ^container_.*_total$
|
- isNot: ^container_.*_total$
|
||||||
|
|
@ -63,7 +61,7 @@ data:
|
||||||
name:
|
name:
|
||||||
matches: ^(.*)_total$
|
matches: ^(.*)_total$
|
||||||
as: ""
|
as: ""
|
||||||
metricsQuery: sum(rate(<<.Series>>{<<.LabelMatchers>>}[5m])) by (<<.GroupBy>>)
|
metricsQuery: sum(rate(<<.Series>>{<<.LabelMatchers>>}[1m])) by (<<.GroupBy>>)
|
||||||
- seriesQuery: '{namespace!="",__name__!~"^container_.*"}'
|
- seriesQuery: '{namespace!="",__name__!~"^container_.*"}'
|
||||||
seriesFilters: []
|
seriesFilters: []
|
||||||
resources:
|
resources:
|
||||||
|
|
@ -71,4 +69,30 @@ data:
|
||||||
name:
|
name:
|
||||||
matches: ^(.*)_seconds_total$
|
matches: ^(.*)_seconds_total$
|
||||||
as: ""
|
as: ""
|
||||||
metricsQuery: sum(rate(<<.Series>>{<<.LabelMatchers>>}[5m])) by (<<.GroupBy>>)
|
metricsQuery: sum(rate(<<.Series>>{<<.LabelMatchers>>}[1m])) by (<<.GroupBy>>)
|
||||||
|
resourceRules:
|
||||||
|
cpu:
|
||||||
|
containerQuery: sum(rate(container_cpu_usage_seconds_total{<<.LabelMatchers>>}[1m])) by (<<.GroupBy>>)
|
||||||
|
nodeQuery: sum(rate(container_cpu_usage_seconds_total{<<.LabelMatchers>>, id='/'}[1m])) by (<<.GroupBy>>)
|
||||||
|
resources:
|
||||||
|
overrides:
|
||||||
|
instance:
|
||||||
|
resource: node
|
||||||
|
namespace:
|
||||||
|
resource: namespace
|
||||||
|
pod_name:
|
||||||
|
resource: pod
|
||||||
|
containerLabel: container_name
|
||||||
|
memory:
|
||||||
|
containerQuery: sum(container_memory_working_set_bytes{<<.LabelMatchers>>}) by (<<.GroupBy>>)
|
||||||
|
nodeQuery: sum(container_memory_working_set_bytes{<<.LabelMatchers>>,id='/'}) by (<<.GroupBy>>)
|
||||||
|
resources:
|
||||||
|
overrides:
|
||||||
|
instance:
|
||||||
|
resource: node
|
||||||
|
namespace:
|
||||||
|
resource: namespace
|
||||||
|
pod_name:
|
||||||
|
resource: pod
|
||||||
|
containerLabel: container_name
|
||||||
|
window: 1m
|
||||||
|
|
|
||||||
|
|
@ -306,11 +306,11 @@ func (p *resourceProvider) queryBoth(now pmodel.Time, resource schema.GroupResou
|
||||||
wg.Add(2)
|
wg.Add(2)
|
||||||
go func() {
|
go func() {
|
||||||
defer wg.Done()
|
defer wg.Done()
|
||||||
cpuRes, cpuErr = p.runQuery(now, p.cpu, resource, "", names...)
|
cpuRes, cpuErr = p.runQuery(now, p.cpu, resource, namespace, names...)
|
||||||
}()
|
}()
|
||||||
go func() {
|
go func() {
|
||||||
defer wg.Done()
|
defer wg.Done()
|
||||||
memRes, memErr = p.runQuery(now, p.mem, resource, "", names...)
|
memRes, memErr = p.runQuery(now, p.mem, resource, namespace, names...)
|
||||||
}()
|
}()
|
||||||
wg.Wait()
|
wg.Wait()
|
||||||
|
|
||||||
|
|
|
||||||
29
pkg/resourceprovider/provider_suite_test.go
Normal file
29
pkg/resourceprovider/provider_suite_test.go
Normal file
|
|
@ -0,0 +1,29 @@
|
||||||
|
/*
|
||||||
|
Copyright 2018 The Kubernetes Authors.
|
||||||
|
|
||||||
|
Licensed under the Apache License, Version 2.0 (the "License");
|
||||||
|
you may not use this file except in compliance with the License.
|
||||||
|
You may obtain a copy of the License at
|
||||||
|
|
||||||
|
http://www.apache.org/licenses/LICENSE-2.0
|
||||||
|
|
||||||
|
Unless required by applicable law or agreed to in writing, software
|
||||||
|
distributed under the License is distributed on an "AS IS" BASIS,
|
||||||
|
WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
|
||||||
|
See the License for the specific language governing permissions and
|
||||||
|
limitations under the License.
|
||||||
|
*/
|
||||||
|
|
||||||
|
package resourceprovider
|
||||||
|
|
||||||
|
import (
|
||||||
|
"testing"
|
||||||
|
|
||||||
|
. "github.com/onsi/ginkgo"
|
||||||
|
. "github.com/onsi/gomega"
|
||||||
|
)
|
||||||
|
|
||||||
|
func TestProvider(t *testing.T) {
|
||||||
|
RegisterFailHandler(Fail)
|
||||||
|
RunSpecs(t, "Resource Metrics Provider Suite")
|
||||||
|
}
|
||||||
265
pkg/resourceprovider/provider_test.go
Normal file
265
pkg/resourceprovider/provider_test.go
Normal file
|
|
@ -0,0 +1,265 @@
|
||||||
|
/*
|
||||||
|
Copyright 2018 The Kubernetes Authors.
|
||||||
|
|
||||||
|
Licensed under the Apache License, Version 2.0 (the "License");
|
||||||
|
you may not use this file except in compliance with the License.
|
||||||
|
You may obtain a copy of the License at
|
||||||
|
|
||||||
|
http://www.apache.org/licenses/LICENSE-2.0
|
||||||
|
|
||||||
|
Unless required by applicable law or agreed to in writing, software
|
||||||
|
distributed under the License is distributed on an "AS IS" BASIS,
|
||||||
|
WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
|
||||||
|
See the License for the specific language governing permissions and
|
||||||
|
limitations under the License.
|
||||||
|
*/
|
||||||
|
|
||||||
|
package resourceprovider
|
||||||
|
|
||||||
|
import (
|
||||||
|
"time"
|
||||||
|
|
||||||
|
"github.com/kubernetes-incubator/metrics-server/pkg/provider"
|
||||||
|
. "github.com/onsi/ginkgo"
|
||||||
|
. "github.com/onsi/gomega"
|
||||||
|
corev1 "k8s.io/api/core/v1"
|
||||||
|
apimeta "k8s.io/apimachinery/pkg/api/meta"
|
||||||
|
"k8s.io/apimachinery/pkg/api/resource"
|
||||||
|
"k8s.io/apimachinery/pkg/runtime/schema"
|
||||||
|
"k8s.io/apimachinery/pkg/types"
|
||||||
|
"k8s.io/metrics/pkg/apis/metrics"
|
||||||
|
|
||||||
|
config "github.com/directxman12/k8s-prometheus-adapter/cmd/config-gen/utils"
|
||||||
|
prom "github.com/directxman12/k8s-prometheus-adapter/pkg/client"
|
||||||
|
fakeprom "github.com/directxman12/k8s-prometheus-adapter/pkg/client/fake"
|
||||||
|
pmodel "github.com/prometheus/common/model"
|
||||||
|
)
|
||||||
|
|
||||||
|
func restMapper() apimeta.RESTMapper {
|
||||||
|
mapper := apimeta.NewDefaultRESTMapper([]schema.GroupVersion{corev1.SchemeGroupVersion})
|
||||||
|
|
||||||
|
mapper.Add(corev1.SchemeGroupVersion.WithKind("Pod"), apimeta.RESTScopeNamespace)
|
||||||
|
mapper.Add(corev1.SchemeGroupVersion.WithKind("Node"), apimeta.RESTScopeRoot)
|
||||||
|
mapper.Add(corev1.SchemeGroupVersion.WithKind("Namespace"), apimeta.RESTScopeRoot)
|
||||||
|
|
||||||
|
return mapper
|
||||||
|
}
|
||||||
|
|
||||||
|
func buildPodSample(namespace, pod, container string, val float64, ts int64) *pmodel.Sample {
|
||||||
|
return &pmodel.Sample{
|
||||||
|
Metric: pmodel.Metric{
|
||||||
|
"namespace": pmodel.LabelValue(namespace),
|
||||||
|
"pod_name": pmodel.LabelValue(pod),
|
||||||
|
"container_name": pmodel.LabelValue(container),
|
||||||
|
},
|
||||||
|
Value: pmodel.SampleValue(val),
|
||||||
|
Timestamp: pmodel.Time(ts),
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
func buildNodeSample(node string, val float64, ts int64) *pmodel.Sample {
|
||||||
|
return &pmodel.Sample{
|
||||||
|
Metric: pmodel.Metric{
|
||||||
|
"instance": pmodel.LabelValue(node),
|
||||||
|
"id": "/",
|
||||||
|
},
|
||||||
|
Value: pmodel.SampleValue(val),
|
||||||
|
Timestamp: pmodel.Time(ts),
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
func buildQueryRes(metric string, samples ...*pmodel.Sample) prom.QueryResult {
|
||||||
|
for _, sample := range samples {
|
||||||
|
sample.Metric[pmodel.MetricNameLabel] = pmodel.LabelValue(metric)
|
||||||
|
}
|
||||||
|
vec := pmodel.Vector(samples)
|
||||||
|
return prom.QueryResult{
|
||||||
|
Type: pmodel.ValVector,
|
||||||
|
Vector: &vec,
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
func mustBuild(sel prom.Selector, err error) prom.Selector {
|
||||||
|
Expect(err).NotTo(HaveOccurred())
|
||||||
|
return sel
|
||||||
|
}
|
||||||
|
|
||||||
|
func buildResList(cpu, memory float64) corev1.ResourceList {
|
||||||
|
return corev1.ResourceList{
|
||||||
|
corev1.ResourceCPU: *resource.NewMilliQuantity(int64(cpu*1000.0), resource.DecimalSI),
|
||||||
|
corev1.ResourceMemory: *resource.NewMilliQuantity(int64(memory*1000.0), resource.BinarySI),
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
var _ = Describe("Resource Metrics Provider", func() {
|
||||||
|
var (
|
||||||
|
prov provider.MetricsProvider
|
||||||
|
fakeProm *fakeprom.FakePrometheusClient
|
||||||
|
cpuQueries, memQueries resourceQuery
|
||||||
|
)
|
||||||
|
|
||||||
|
BeforeEach(func() {
|
||||||
|
By("setting up a fake prometheus client and provider")
|
||||||
|
mapper := restMapper()
|
||||||
|
|
||||||
|
cfg := config.DefaultConfig(1*time.Minute, "")
|
||||||
|
|
||||||
|
var err error
|
||||||
|
cpuQueries, err = newResourceQuery(cfg.ResourceRules.CPU, mapper)
|
||||||
|
Expect(err).NotTo(HaveOccurred())
|
||||||
|
memQueries, err = newResourceQuery(cfg.ResourceRules.Memory, mapper)
|
||||||
|
Expect(err).NotTo(HaveOccurred())
|
||||||
|
|
||||||
|
fakeProm = &fakeprom.FakePrometheusClient{}
|
||||||
|
fakeProm.AcceptableInterval = pmodel.Interval{End: pmodel.Latest}
|
||||||
|
|
||||||
|
prov, err = NewProvider(fakeProm, restMapper(), cfg.ResourceRules)
|
||||||
|
Expect(err).NotTo(HaveOccurred())
|
||||||
|
})
|
||||||
|
|
||||||
|
It("should be able to list metrics pods across different namespaces", func() {
|
||||||
|
pods := []types.NamespacedName{
|
||||||
|
{Namespace: "some-ns", Name: "pod1"},
|
||||||
|
{Namespace: "some-ns", Name: "pod3"},
|
||||||
|
{Namespace: "other-ns", Name: "pod27"},
|
||||||
|
}
|
||||||
|
fakeProm.QueryResults = map[prom.Selector]prom.QueryResult{
|
||||||
|
mustBuild(cpuQueries.contQuery.Build("", podResource, "some-ns", []string{cpuQueries.containerLabel}, "pod1", "pod3")): buildQueryRes("container_cpu_usage_seconds_total",
|
||||||
|
buildPodSample("some-ns", "pod1", "cont1", 1100.0, 10),
|
||||||
|
buildPodSample("some-ns", "pod1", "cont2", 1110.0, 20),
|
||||||
|
buildPodSample("some-ns", "pod3", "cont1", 1300.0, 10),
|
||||||
|
buildPodSample("some-ns", "pod3", "cont2", 1310.0, 20),
|
||||||
|
),
|
||||||
|
mustBuild(cpuQueries.contQuery.Build("", podResource, "other-ns", []string{cpuQueries.containerLabel}, "pod27")): buildQueryRes("container_cpu_usage_seconds_total",
|
||||||
|
buildPodSample("other-ns", "pod27", "cont1", 2200.0, 270),
|
||||||
|
),
|
||||||
|
mustBuild(memQueries.contQuery.Build("", podResource, "some-ns", []string{cpuQueries.containerLabel}, "pod1", "pod3")): buildQueryRes("container_memory_working_set_bytes",
|
||||||
|
buildPodSample("some-ns", "pod1", "cont1", 3100.0, 11),
|
||||||
|
buildPodSample("some-ns", "pod1", "cont2", 3110.0, 21),
|
||||||
|
buildPodSample("some-ns", "pod3", "cont1", 3300.0, 11),
|
||||||
|
buildPodSample("some-ns", "pod3", "cont2", 3310.0, 21),
|
||||||
|
),
|
||||||
|
mustBuild(memQueries.contQuery.Build("", podResource, "other-ns", []string{cpuQueries.containerLabel}, "pod27")): buildQueryRes("container_memory_working_set_bytes",
|
||||||
|
buildPodSample("other-ns", "pod27", "cont1", 4200.0, 271),
|
||||||
|
),
|
||||||
|
}
|
||||||
|
|
||||||
|
By("querying for metrics for some pods")
|
||||||
|
times, metricVals, err := prov.GetContainerMetrics(pods...)
|
||||||
|
Expect(err).NotTo(HaveOccurred())
|
||||||
|
|
||||||
|
By("verifying that the reported times for each are the earliest times for each pod")
|
||||||
|
Expect(times).To(Equal([]provider.TimeInfo{
|
||||||
|
{Timestamp: pmodel.Time(10).Time(), Window: 1 * time.Minute},
|
||||||
|
{Timestamp: pmodel.Time(10).Time(), Window: 1 * time.Minute},
|
||||||
|
{Timestamp: pmodel.Time(270).Time(), Window: 1 * time.Minute},
|
||||||
|
}))
|
||||||
|
|
||||||
|
By("verifying that the right metrics were fetched")
|
||||||
|
Expect(metricVals).To(HaveLen(3))
|
||||||
|
Expect(metricVals[0]).To(ConsistOf(
|
||||||
|
metrics.ContainerMetrics{Name: "cont1", Usage: buildResList(1100.0, 3100.0)},
|
||||||
|
metrics.ContainerMetrics{Name: "cont2", Usage: buildResList(1110.0, 3110.0)},
|
||||||
|
))
|
||||||
|
Expect(metricVals[1]).To(ConsistOf(
|
||||||
|
metrics.ContainerMetrics{Name: "cont1", Usage: buildResList(1300.0, 3300.0)},
|
||||||
|
metrics.ContainerMetrics{Name: "cont2", Usage: buildResList(1310.0, 3310.0)},
|
||||||
|
))
|
||||||
|
|
||||||
|
Expect(metricVals[2]).To(ConsistOf(
|
||||||
|
metrics.ContainerMetrics{Name: "cont1", Usage: buildResList(2200.0, 4200.0)},
|
||||||
|
))
|
||||||
|
})
|
||||||
|
|
||||||
|
It("should return nil metrics for missing pods, but still return partial results", func() {
|
||||||
|
fakeProm.QueryResults = map[prom.Selector]prom.QueryResult{
|
||||||
|
mustBuild(cpuQueries.contQuery.Build("", podResource, "some-ns", []string{cpuQueries.containerLabel}, "pod1", "pod-nonexistant")): buildQueryRes("container_cpu_usage_seconds_total",
|
||||||
|
buildPodSample("some-ns", "pod1", "cont1", 1100.0, 10),
|
||||||
|
buildPodSample("some-ns", "pod1", "cont2", 1110.0, 20),
|
||||||
|
),
|
||||||
|
mustBuild(memQueries.contQuery.Build("", podResource, "some-ns", []string{cpuQueries.containerLabel}, "pod1", "pod-nonexistant")): buildQueryRes("container_memory_working_set_bytes",
|
||||||
|
buildPodSample("some-ns", "pod1", "cont1", 3100.0, 11),
|
||||||
|
buildPodSample("some-ns", "pod1", "cont2", 3110.0, 21),
|
||||||
|
),
|
||||||
|
}
|
||||||
|
|
||||||
|
By("querying for metrics for some pods, one of which is missing")
|
||||||
|
times, metricVals, err := prov.GetContainerMetrics(
|
||||||
|
types.NamespacedName{Namespace: "some-ns", Name: "pod1"},
|
||||||
|
types.NamespacedName{Namespace: "some-ns", Name: "pod-nonexistant"},
|
||||||
|
)
|
||||||
|
Expect(err).NotTo(HaveOccurred())
|
||||||
|
|
||||||
|
By("verifying that the missing pod had nil metrics")
|
||||||
|
Expect(metricVals).To(HaveLen(2))
|
||||||
|
Expect(metricVals[1]).To(BeNil())
|
||||||
|
|
||||||
|
By("verifying that the rest of time metrics and times are correct")
|
||||||
|
Expect(metricVals[0]).To(ConsistOf(
|
||||||
|
metrics.ContainerMetrics{Name: "cont1", Usage: buildResList(1100.0, 3100.0)},
|
||||||
|
metrics.ContainerMetrics{Name: "cont2", Usage: buildResList(1110.0, 3110.0)},
|
||||||
|
))
|
||||||
|
Expect(times).To(HaveLen(2))
|
||||||
|
Expect(times[0]).To(Equal(provider.TimeInfo{Timestamp: pmodel.Time(10).Time(), Window: 1 * time.Minute}))
|
||||||
|
})
|
||||||
|
|
||||||
|
It("should be able to list metrics for nodes", func() {
|
||||||
|
fakeProm.QueryResults = map[prom.Selector]prom.QueryResult{
|
||||||
|
mustBuild(cpuQueries.nodeQuery.Build("", nodeResource, "", nil, "node1", "node2")): buildQueryRes("container_cpu_usage_seconds_total",
|
||||||
|
buildNodeSample("node1", 1100.0, 10),
|
||||||
|
buildNodeSample("node2", 1200.0, 14),
|
||||||
|
),
|
||||||
|
mustBuild(memQueries.nodeQuery.Build("", nodeResource, "", nil, "node1", "node2")): buildQueryRes("container_memory_working_set_bytes",
|
||||||
|
buildNodeSample("node1", 2100.0, 11),
|
||||||
|
buildNodeSample("node2", 2200.0, 12),
|
||||||
|
),
|
||||||
|
}
|
||||||
|
By("querying for metrics for some nodes")
|
||||||
|
times, metricVals, err := prov.GetNodeMetrics("node1", "node2")
|
||||||
|
Expect(err).NotTo(HaveOccurred())
|
||||||
|
|
||||||
|
By("verifying that the reported times for each are the earliest times for each pod")
|
||||||
|
Expect(times).To(Equal([]provider.TimeInfo{
|
||||||
|
{Timestamp: pmodel.Time(10).Time(), Window: 1 * time.Minute},
|
||||||
|
{Timestamp: pmodel.Time(12).Time(), Window: 1 * time.Minute},
|
||||||
|
}))
|
||||||
|
|
||||||
|
By("verifying that the right metrics were fetched")
|
||||||
|
Expect(metricVals).To(Equal([]corev1.ResourceList{
|
||||||
|
buildResList(1100.0, 2100.0),
|
||||||
|
buildResList(1200.0, 2200.0),
|
||||||
|
}))
|
||||||
|
})
|
||||||
|
|
||||||
|
It("should return nil metrics for missing nodes, but still return partial results", func() {
|
||||||
|
fakeProm.QueryResults = map[prom.Selector]prom.QueryResult{
|
||||||
|
mustBuild(cpuQueries.nodeQuery.Build("", nodeResource, "", nil, "node1", "node2", "node3")): buildQueryRes("container_cpu_usage_seconds_total",
|
||||||
|
buildNodeSample("node1", 1100.0, 10),
|
||||||
|
buildNodeSample("node2", 1200.0, 14),
|
||||||
|
),
|
||||||
|
mustBuild(memQueries.nodeQuery.Build("", nodeResource, "", nil, "node1", "node2", "node3")): buildQueryRes("container_memory_working_set_bytes",
|
||||||
|
buildNodeSample("node1", 2100.0, 11),
|
||||||
|
buildNodeSample("node2", 2200.0, 12),
|
||||||
|
),
|
||||||
|
}
|
||||||
|
By("querying for metrics for some nodes, one of which is missing")
|
||||||
|
times, metricVals, err := prov.GetNodeMetrics("node1", "node2", "node3")
|
||||||
|
Expect(err).NotTo(HaveOccurred())
|
||||||
|
|
||||||
|
By("verifying that the missing pod had nil metrics")
|
||||||
|
Expect(metricVals).To(HaveLen(3))
|
||||||
|
Expect(metricVals[2]).To(BeNil())
|
||||||
|
|
||||||
|
By("verifying that the rest of time metrics and times are correct")
|
||||||
|
Expect(metricVals).To(Equal([]corev1.ResourceList{
|
||||||
|
buildResList(1100.0, 2100.0),
|
||||||
|
buildResList(1200.0, 2200.0),
|
||||||
|
nil,
|
||||||
|
}))
|
||||||
|
Expect(times).To(Equal([]provider.TimeInfo{
|
||||||
|
{Timestamp: pmodel.Time(10).Time(), Window: 1 * time.Minute},
|
||||||
|
{Timestamp: pmodel.Time(12).Time(), Window: 1 * time.Minute},
|
||||||
|
{},
|
||||||
|
}))
|
||||||
|
})
|
||||||
|
})
|
||||||
Loading…
Add table
Add a link
Reference in a new issue