| 123456789101112131415161718192021222324252627282930313233343536373839404142434445464748495051525354555657585960616263646566676869707172737475767778798081828384858687888990919293949596979899100101102103104105106107108109110111112113114115116117118119120121122123124125126127128129130131132133134135136137138139140141142143144145146147148149150151152153154155156157158159160161162163164165166167168169170171172173174175176177178179180181182183184185186187188189190191192193194195196197198199200201202203204205206207208209210211212213214215216217218219220221222223224225226227228229230231232233234235236237238239240241242243244245246247248249250251252253254255256257258259260261262263264265266267268269270271272273274275276277278279280281282283284285286287288289290291292293294295296297298299300301302303304305306307308309310311312313314315316317318319320321322323324325326327328329330331332333334335336337338339340341342343344345346347348349350351352353354355356357358359360361362363364365366367368369370371372373374375376377378379380381382383384385386387388389390391392393394395396397398399400401402403404405406407408409410411412413414415416417418419420421422423424425426427428429430431432433434435436437438439440441442443444445446447448449450451452453454455456457458459460461462463464465466467468469470471472473474475476477478479480481482483484485486487488489490491492493494495496497498499500501502503504505506507508509510511512513514515516517518519520521522523524525526527528529530531532533534535536537538539540541542543544545546547548549550551552553554555556557558559560561562563564565566567568569570571572573574575576577578579580581582583584585586587588589590591592593594595596597598599600601602603604605606607608609610611612613614615616617618619620621622623624625626627628629630631632633634635636637638639640641642643644645646647648649650651652653654655656657658659660661662663664665666667668669670671672673674675676677678679680681682683684685686687688689690691692693694695696697698699700701702703704705706707708709710711712713714715716717718719720721722723724725726727728729730731732733734735736737738739740741742743744745746747748749750751752753754755756757758759760761762763764765766767768769770771772773774775776777778779780781782783784785786787788789790791792793794795796797798799800801802803804805806807808809810811812813814815816817818819820821822823824825826827828829830831832833834835836837838839840841842843844845846847848849850851852853854855856857858859860861862863864865866867868869870871872873874875876877878879880881882883884885886887888889890891892893894895896897898899900901902903904905906907908909910911912913914915916917918919920921922923924925926927928929930931932933934935936937938939940941942943944945946947948949950951952953954955956957958959960961962963964965966967968969970971972973974975976977978979980981982983984985986987988989990991992993994995996997998999100010011002100310041005100610071008100910101011101210131014101510161017101810191020102110221023102410251026102710281029103010311032103310341035103610371038103910401041104210431044104510461047104810491050105110521053105410551056105710581059106010611062106310641065106610671068106910701071107210731074107510761077107810791080108110821083108410851086108710881089109010911092109310941095109610971098109911001101110211031104110511061107110811091110111111121113111411151116111711181119112011211122112311241125112611271128112911301131113211331134113511361137113811391140114111421143114411451146114711481149115011511152115311541155115611571158115911601161116211631164116511661167116811691170117111721173117411751176117711781179118011811182118311841185118611871188118911901191119211931194119511961197119811991200120112021203120412051206120712081209121012111212121312141215121612171218121912201221122212231224122512261227122812291230123112321233123412351236123712381239124012411242124312441245124612471248124912501251125212531254125512561257125812591260126112621263126412651266126712681269127012711272127312741275127612771278127912801281128212831284128512861287128812891290129112921293129412951296129712981299130013011302130313041305130613071308130913101311131213131314131513161317131813191320132113221323132413251326132713281329133013311332133313341335133613371338133913401341134213431344134513461347134813491350135113521353135413551356135713581359136013611362136313641365136613671368136913701371137213731374137513761377137813791380138113821383138413851386138713881389139013911392139313941395139613971398139914001401140214031404140514061407140814091410141114121413141414151416141714181419142014211422142314241425142614271428142914301431143214331434143514361437143814391440144114421443144414451446144714481449145014511452145314541455145614571458145914601461146214631464146514661467146814691470147114721473147414751476147714781479148014811482148314841485148614871488148914901491149214931494149514961497149814991500150115021503150415051506150715081509151015111512151315141515151615171518151915201521152215231524152515261527152815291530153115321533153415351536153715381539154015411542154315441545154615471548154915501551155215531554155515561557155815591560156115621563156415651566156715681569157015711572157315741575157615771578157915801581158215831584158515861587158815891590159115921593159415951596 |
- package costmodel
- import (
- "fmt"
- "net"
- "strconv"
- "time"
- "github.com/opencost/opencost/pkg/cloud/provider"
- prometheus "github.com/prometheus/client_golang/api"
- "golang.org/x/exp/slices"
- "github.com/opencost/opencost/core/pkg/log"
- "github.com/opencost/opencost/core/pkg/opencost"
- "github.com/opencost/opencost/core/pkg/util/timeutil"
- "github.com/opencost/opencost/pkg/cloud/models"
- "github.com/opencost/opencost/pkg/env"
- "github.com/opencost/opencost/pkg/prom"
- )
- const (
- queryClusterCores = `sum(
- avg(avg_over_time(kube_node_status_capacity_cpu_cores{%s}[%s] %s)) by (node, %s) * avg(avg_over_time(node_cpu_hourly_cost{%s}[%s] %s)) by (node, %s) * 730 +
- avg(avg_over_time(node_gpu_hourly_cost{%s}[%s] %s)) by (node, %s) * 730
- ) by (%s)`
- queryClusterRAM = `sum(
- avg(avg_over_time(kube_node_status_capacity_memory_bytes{%s}[%s] %s)) by (node, %s) / 1024 / 1024 / 1024 * avg(avg_over_time(node_ram_hourly_cost{%s}[%s] %s)) by (node, %s) * 730
- ) by (%s)`
- queryStorage = `sum(
- avg(avg_over_time(pv_hourly_cost{%s}[%s] %s)) by (persistentvolume, %s) * 730
- * avg(avg_over_time(kube_persistentvolume_capacity_bytes{%s}[%s] %s)) by (persistentvolume, %s) / 1024 / 1024 / 1024
- ) by (%s) %s`
- queryTotal = `sum(avg(node_total_hourly_cost{%s}) by (node, %s)) * 730 +
- sum(
- avg(avg_over_time(pv_hourly_cost{%s}[1h])) by (persistentvolume, %s) * 730
- * avg(avg_over_time(kube_persistentvolume_capacity_bytes{%s}[1h])) by (persistentvolume, %s) / 1024 / 1024 / 1024
- ) by (%s) %s`
- queryNodes = `sum(avg(node_total_hourly_cost{%s}) by (node, %s)) * 730 %s`
- )
- const maxLocalDiskSize = 200 // AWS limits root disks to 100 Gi, and occasional metric errors in filesystem size should not contribute to large costs.
- // Costs represents cumulative and monthly cluster costs over a given duration. Costs
- // are broken down by cores, memory, and storage.
- type ClusterCosts struct {
- Start *time.Time `json:"startTime"`
- End *time.Time `json:"endTime"`
- CPUCumulative float64 `json:"cpuCumulativeCost"`
- CPUMonthly float64 `json:"cpuMonthlyCost"`
- CPUBreakdown *ClusterCostsBreakdown `json:"cpuBreakdown"`
- GPUCumulative float64 `json:"gpuCumulativeCost"`
- GPUMonthly float64 `json:"gpuMonthlyCost"`
- RAMCumulative float64 `json:"ramCumulativeCost"`
- RAMMonthly float64 `json:"ramMonthlyCost"`
- RAMBreakdown *ClusterCostsBreakdown `json:"ramBreakdown"`
- StorageCumulative float64 `json:"storageCumulativeCost"`
- StorageMonthly float64 `json:"storageMonthlyCost"`
- StorageBreakdown *ClusterCostsBreakdown `json:"storageBreakdown"`
- TotalCumulative float64 `json:"totalCumulativeCost"`
- TotalMonthly float64 `json:"totalMonthlyCost"`
- DataMinutes float64
- }
- // ClusterCostsBreakdown provides percentage-based breakdown of a resource by
- // categories: user for user-space (i.e. non-system) usage, system, and idle.
- type ClusterCostsBreakdown struct {
- Idle float64 `json:"idle"`
- Other float64 `json:"other"`
- System float64 `json:"system"`
- User float64 `json:"user"`
- }
- // NewClusterCostsFromCumulative takes cumulative cost data over a given time range, computes
- // the associated monthly rate data, and returns the Costs.
- func NewClusterCostsFromCumulative(cpu, gpu, ram, storage float64, window, offset time.Duration, dataHours float64) (*ClusterCosts, error) {
- start, end := timeutil.ParseTimeRange(window, offset)
- // If the number of hours is not given (i.e. is zero) compute one from the window and offset
- if dataHours == 0 {
- dataHours = end.Sub(start).Hours()
- }
- // Do not allow zero-length windows to prevent divide-by-zero issues
- if dataHours == 0 {
- return nil, fmt.Errorf("illegal time range: window %s, offset %s", window, offset)
- }
- cc := &ClusterCosts{
- Start: &start,
- End: &end,
- CPUCumulative: cpu,
- GPUCumulative: gpu,
- RAMCumulative: ram,
- StorageCumulative: storage,
- TotalCumulative: cpu + gpu + ram + storage,
- CPUMonthly: cpu / dataHours * (timeutil.HoursPerMonth),
- GPUMonthly: gpu / dataHours * (timeutil.HoursPerMonth),
- RAMMonthly: ram / dataHours * (timeutil.HoursPerMonth),
- StorageMonthly: storage / dataHours * (timeutil.HoursPerMonth),
- }
- cc.TotalMonthly = cc.CPUMonthly + cc.GPUMonthly + cc.RAMMonthly + cc.StorageMonthly
- return cc, nil
- }
- type Disk struct {
- Cluster string
- Name string
- ProviderID string
- StorageClass string
- VolumeName string
- ClaimName string
- ClaimNamespace string
- Cost float64
- Bytes float64
- // These two fields may not be available at all times because they rely on
- // a new set of metrics that may or may not be available. Thus, they must
- // be nilable to represent the complete absence of the data.
- //
- // In other words, nilability here lets us distinguish between
- // "metric is not available" and "metric is available but is 0".
- //
- // They end in "Ptr" to distinguish from an earlier version in order to
- // ensure that all usages are checked for nil.
- BytesUsedAvgPtr *float64
- BytesUsedMaxPtr *float64
- Local bool
- Start time.Time
- End time.Time
- Minutes float64
- Breakdown *ClusterCostsBreakdown
- }
- type DiskIdentifier struct {
- Cluster string
- Name string
- }
- func ClusterDisks(client prometheus.Client, provider models.Provider, start, end time.Time) (map[DiskIdentifier]*Disk, error) {
- // Start from the time "end", querying backwards
- t := end
- // minsPerResolution determines accuracy and resource use for the following
- // queries. Smaller values (higher resolution) result in better accuracy,
- // but more expensive queries, and vice-a-versa.
- resolution := env.GetETLResolution()
- //Ensuring if ETL_RESOLUTION_SECONDS is less than 60s default it to 1m
- var minsPerResolution int
- if minsPerResolution = int(resolution.Minutes()); int(resolution.Minutes()) == 0 {
- minsPerResolution = 1
- log.DedupedWarningf(3, "ClusterDisks(): Configured ETL resolution (%d seconds) is below the 60 seconds threshold. Overriding with 1 minute.", int(resolution.Seconds()))
- }
- durStr := timeutil.DurationString(end.Sub(start))
- if durStr == "" {
- return nil, fmt.Errorf("illegal duration value for %s", opencost.NewClosedWindow(start, end))
- }
- ctx := prom.NewNamedContext(client, prom.ClusterContextName)
- queryPVCost := fmt.Sprintf(`avg(avg_over_time(pv_hourly_cost{%s}[%s])) by (%s, persistentvolume,provider_id)`, env.GetPromClusterFilter(), durStr, env.GetPromClusterLabel())
- queryPVSize := fmt.Sprintf(`avg(avg_over_time(kube_persistentvolume_capacity_bytes{%s}[%s])) by (%s, persistentvolume)`, env.GetPromClusterFilter(), durStr, env.GetPromClusterLabel())
- queryActiveMins := fmt.Sprintf(`avg(kube_persistentvolume_capacity_bytes{%s}) by (%s, persistentvolume)[%s:%dm]`, env.GetPromClusterFilter(), env.GetPromClusterLabel(), durStr, minsPerResolution)
- queryPVStorageClass := fmt.Sprintf(`avg(avg_over_time(kubecost_pv_info{%s}[%s])) by (%s, persistentvolume, storageclass)`, env.GetPromClusterFilter(), durStr, env.GetPromClusterLabel())
- queryPVUsedAvg := fmt.Sprintf(`avg(avg_over_time(kubelet_volume_stats_used_bytes{%s}[%s])) by (%s, persistentvolumeclaim, namespace)`, env.GetPromClusterFilter(), durStr, env.GetPromClusterLabel())
- queryPVUsedMax := fmt.Sprintf(`max(max_over_time(kubelet_volume_stats_used_bytes{%s}[%s])) by (%s, persistentvolumeclaim, namespace)`, env.GetPromClusterFilter(), durStr, env.GetPromClusterLabel())
- queryPVCInfo := fmt.Sprintf(`avg(avg_over_time(kube_persistentvolumeclaim_info{%s}[%s])) by (%s, volumename, persistentvolumeclaim, namespace)`, env.GetPromClusterFilter(), durStr, env.GetPromClusterLabel())
- resChPVCost := ctx.QueryAtTime(queryPVCost, t)
- resChPVSize := ctx.QueryAtTime(queryPVSize, t)
- resChActiveMins := ctx.QueryAtTime(queryActiveMins, t)
- resChPVStorageClass := ctx.QueryAtTime(queryPVStorageClass, t)
- resChPVUsedAvg := ctx.QueryAtTime(queryPVUsedAvg, t)
- resChPVUsedMax := ctx.QueryAtTime(queryPVUsedMax, t)
- resChPVCInfo := ctx.QueryAtTime(queryPVCInfo, t)
- resPVCost, _ := resChPVCost.Await()
- resPVSize, _ := resChPVSize.Await()
- resActiveMins, _ := resChActiveMins.Await()
- resPVStorageClass, _ := resChPVStorageClass.Await()
- resPVUsedAvg, _ := resChPVUsedAvg.Await()
- resPVUsedMax, _ := resChPVUsedMax.Await()
- resPVCInfo, _ := resChPVCInfo.Await()
- // Cloud providers do not always charge for a node's local disk costs (i.e.
- // ephemeral storage). Provide an option to opt out of calculating &
- // allocating local disk costs. Note, that this does not affect
- // PersistentVolume costs.
- //
- // Ref:
- // https://docs.aws.amazon.com/AWSEC2/latest/UserGuide/RootDeviceStorage.html
- // https://learn.microsoft.com/en-us/azure/virtual-machines/managed-disks-overview#temporary-disk
- // https://cloud.google.com/compute/docs/disks/local-ssd
- resLocalStorageCost := []*prom.QueryResult{}
- resLocalStorageUsedCost := []*prom.QueryResult{}
- resLocalStorageUsedAvg := []*prom.QueryResult{}
- resLocalStorageUsedMax := []*prom.QueryResult{}
- resLocalStorageBytes := []*prom.QueryResult{}
- resLocalActiveMins := []*prom.QueryResult{}
- if env.GetAssetIncludeLocalDiskCost() {
- // hourlyToCumulative is a scaling factor that, when multiplied by an
- // hourly value, converts it to a cumulative value; i.e. [$/hr] *
- // [min/res]*[hr/min] = [$/res]
- hourlyToCumulative := float64(minsPerResolution) * (1.0 / 60.0)
- costPerGBHr := 0.04 / 730.0
- queryLocalStorageCost := fmt.Sprintf(`sum_over_time(sum(container_fs_limit_bytes{device!="tmpfs", id="/", %s}) by (instance, %s)[%s:%dm]) / 1024 / 1024 / 1024 * %f * %f`, env.GetPromClusterFilter(), env.GetPromClusterLabel(), durStr, minsPerResolution, hourlyToCumulative, costPerGBHr)
- queryLocalStorageUsedCost := fmt.Sprintf(`sum_over_time(sum(container_fs_usage_bytes{device!="tmpfs", id="/", %s}) by (instance, %s)[%s:%dm]) / 1024 / 1024 / 1024 * %f * %f`, env.GetPromClusterFilter(), env.GetPromClusterLabel(), durStr, minsPerResolution, hourlyToCumulative, costPerGBHr)
- queryLocalStorageUsedAvg := fmt.Sprintf(`avg(sum(avg_over_time(container_fs_usage_bytes{device!="tmpfs", id="/", %s}[%s])) by (instance, %s, job)) by (instance, %s)`, env.GetPromClusterFilter(), durStr, env.GetPromClusterLabel(), env.GetPromClusterLabel())
- queryLocalStorageUsedMax := fmt.Sprintf(`max(sum(max_over_time(container_fs_usage_bytes{device!="tmpfs", id="/", %s}[%s])) by (instance, %s, job)) by (instance, %s)`, env.GetPromClusterFilter(), durStr, env.GetPromClusterLabel(), env.GetPromClusterLabel())
- queryLocalStorageBytes := fmt.Sprintf(`avg_over_time(sum(container_fs_limit_bytes{device!="tmpfs", id="/", %s}) by (instance, %s)[%s:%dm])`, env.GetPromClusterFilter(), env.GetPromClusterLabel(), durStr, minsPerResolution)
- queryLocalActiveMins := fmt.Sprintf(`count(node_total_hourly_cost{%s}) by (%s, node)[%s:%dm]`, env.GetPromClusterFilter(), env.GetPromClusterLabel(), durStr, minsPerResolution)
- resChLocalStorageCost := ctx.QueryAtTime(queryLocalStorageCost, t)
- resChLocalStorageUsedCost := ctx.QueryAtTime(queryLocalStorageUsedCost, t)
- resChLocalStoreageUsedAvg := ctx.QueryAtTime(queryLocalStorageUsedAvg, t)
- resChLocalStoreageUsedMax := ctx.QueryAtTime(queryLocalStorageUsedMax, t)
- resChLocalStorageBytes := ctx.QueryAtTime(queryLocalStorageBytes, t)
- resChLocalActiveMins := ctx.QueryAtTime(queryLocalActiveMins, t)
- resLocalStorageCost, _ = resChLocalStorageCost.Await()
- resLocalStorageUsedCost, _ = resChLocalStorageUsedCost.Await()
- resLocalStorageUsedAvg, _ = resChLocalStoreageUsedAvg.Await()
- resLocalStorageUsedMax, _ = resChLocalStoreageUsedMax.Await()
- resLocalStorageBytes, _ = resChLocalStorageBytes.Await()
- resLocalActiveMins, _ = resChLocalActiveMins.Await()
- }
- if ctx.HasErrors() {
- return nil, ctx.ErrorCollection()
- }
- diskMap := map[DiskIdentifier]*Disk{}
- for _, result := range resPVCInfo {
- cluster, err := result.GetString(env.GetPromClusterLabel())
- if err != nil {
- cluster = env.GetClusterID()
- }
- volumeName, err := result.GetString("volumename")
- if err != nil {
- log.Debugf("ClusterDisks: pv claim data missing volumename")
- continue
- }
- claimName, err := result.GetString("persistentvolumeclaim")
- if err != nil {
- log.Debugf("ClusterDisks: pv claim data missing persistentvolumeclaim")
- continue
- }
- claimNamespace, err := result.GetString("namespace")
- if err != nil {
- log.Debugf("ClusterDisks: pv claim data missing namespace")
- continue
- }
- key := DiskIdentifier{cluster, volumeName}
- if _, ok := diskMap[key]; !ok {
- diskMap[key] = &Disk{
- Cluster: cluster,
- Name: volumeName,
- Breakdown: &ClusterCostsBreakdown{},
- }
- }
- diskMap[key].VolumeName = volumeName
- diskMap[key].ClaimName = claimName
- diskMap[key].ClaimNamespace = claimNamespace
- }
- pvCosts(diskMap, resolution, resActiveMins, resPVSize, resPVCost, resPVUsedAvg, resPVUsedMax, resPVCInfo, provider, opencost.NewClosedWindow(start, end))
- for _, result := range resLocalStorageCost {
- cluster, err := result.GetString(env.GetPromClusterLabel())
- if err != nil {
- cluster = env.GetClusterID()
- }
- name, err := result.GetString("instance")
- if err != nil {
- log.Warnf("ClusterDisks: local storage data missing instance")
- continue
- }
- cost := result.Values[0].Value
- key := DiskIdentifier{cluster, name}
- if _, ok := diskMap[key]; !ok {
- diskMap[key] = &Disk{
- Cluster: cluster,
- Name: name,
- Breakdown: &ClusterCostsBreakdown{},
- Local: true,
- }
- }
- diskMap[key].Cost += cost
- //Assigning explicitly the storage class of local storage to local
- diskMap[key].StorageClass = opencost.LocalStorageClass
- }
- for _, result := range resLocalStorageUsedCost {
- cluster, err := result.GetString(env.GetPromClusterLabel())
- if err != nil {
- cluster = env.GetClusterID()
- }
- name, err := result.GetString("instance")
- if err != nil {
- log.Warnf("ClusterDisks: local storage usage data missing instance")
- continue
- }
- cost := result.Values[0].Value
- key := DiskIdentifier{cluster, name}
- if _, ok := diskMap[key]; !ok {
- diskMap[key] = &Disk{
- Cluster: cluster,
- Name: name,
- Breakdown: &ClusterCostsBreakdown{},
- Local: true,
- }
- }
- diskMap[key].Breakdown.System = cost / diskMap[key].Cost
- }
- for _, result := range resLocalStorageUsedAvg {
- cluster, err := result.GetString(env.GetPromClusterLabel())
- if err != nil {
- cluster = env.GetClusterID()
- }
- name, err := result.GetString("instance")
- if err != nil {
- log.Warnf("ClusterDisks: local storage data missing instance")
- continue
- }
- bytesAvg := result.Values[0].Value
- key := DiskIdentifier{cluster, name}
- if _, ok := diskMap[key]; !ok {
- diskMap[key] = &Disk{
- Cluster: cluster,
- Name: name,
- Breakdown: &ClusterCostsBreakdown{},
- Local: true,
- }
- }
- diskMap[key].BytesUsedAvgPtr = &bytesAvg
- }
- for _, result := range resLocalStorageUsedMax {
- cluster, err := result.GetString(env.GetPromClusterLabel())
- if err != nil {
- cluster = env.GetClusterID()
- }
- name, err := result.GetString("instance")
- if err != nil {
- log.Warnf("ClusterDisks: local storage data missing instance")
- continue
- }
- bytesMax := result.Values[0].Value
- key := DiskIdentifier{cluster, name}
- if _, ok := diskMap[key]; !ok {
- diskMap[key] = &Disk{
- Cluster: cluster,
- Name: name,
- Breakdown: &ClusterCostsBreakdown{},
- Local: true,
- }
- }
- diskMap[key].BytesUsedMaxPtr = &bytesMax
- }
- for _, result := range resLocalStorageBytes {
- cluster, err := result.GetString(env.GetPromClusterLabel())
- if err != nil {
- cluster = env.GetClusterID()
- }
- name, err := result.GetString("instance")
- if err != nil {
- log.Warnf("ClusterDisks: local storage data missing instance")
- continue
- }
- bytes := result.Values[0].Value
- key := DiskIdentifier{cluster, name}
- if _, ok := diskMap[key]; !ok {
- diskMap[key] = &Disk{
- Cluster: cluster,
- Name: name,
- Breakdown: &ClusterCostsBreakdown{},
- Local: true,
- }
- }
- diskMap[key].Bytes = bytes
- if bytes/1024/1024/1024 > maxLocalDiskSize {
- log.DedupedWarningf(5, "Deleting large root disk/localstorage disk from analysis")
- delete(diskMap, key)
- }
- }
- for _, result := range resLocalActiveMins {
- cluster, err := result.GetString(env.GetPromClusterLabel())
- if err != nil {
- cluster = env.GetClusterID()
- }
- name, err := result.GetString("node")
- if err != nil {
- log.DedupedWarningf(5, "ClusterDisks: local active mins data missing instance")
- continue
- }
- key := DiskIdentifier{cluster, name}
- if _, ok := diskMap[key]; !ok {
- log.DedupedWarningf(5, "ClusterDisks: local active mins for unidentified disk or disk deleted from analysis")
- continue
- }
- if len(result.Values) == 0 {
- continue
- }
- s := time.Unix(int64(result.Values[0].Timestamp), 0)
- e := time.Unix(int64(result.Values[len(result.Values)-1].Timestamp), 0)
- mins := e.Sub(s).Minutes()
- // TODO niko/assets if mins >= threshold, interpolate for missing data?
- diskMap[key].End = e
- diskMap[key].Start = s
- diskMap[key].Minutes = mins
- }
- var unTracedDiskLogData []DiskIdentifier
- //Iterating through Persistent Volume given by custom metrics kubecost_pv_info and assign the storage class if known and __unknown__ if not populated.
- for _, result := range resPVStorageClass {
- cluster, err := result.GetString(env.GetPromClusterLabel())
- if err != nil {
- cluster = env.GetClusterID()
- }
- name, _ := result.GetString("persistentvolume")
- key := DiskIdentifier{cluster, name}
- if _, ok := diskMap[key]; !ok {
- if !slices.Contains(unTracedDiskLogData, key) {
- unTracedDiskLogData = append(unTracedDiskLogData, key)
- }
- continue
- }
- if len(result.Values) == 0 {
- continue
- }
- storageClass, err := result.GetString("storageclass")
- if err != nil {
- diskMap[key].StorageClass = opencost.UnknownStorageClass
- } else {
- diskMap[key].StorageClass = storageClass
- }
- }
- // Logging the unidentified disk information outside the loop
- for _, unIdentifiedDisk := range unTracedDiskLogData {
- log.Warnf("ClusterDisks: Cluster %s has Storage Class information for unidentified disk %s or disk deleted from analysis", unIdentifiedDisk.Cluster, unIdentifiedDisk.Name)
- }
- for _, disk := range diskMap {
- // Apply all remaining RAM to Idle
- disk.Breakdown.Idle = 1.0 - (disk.Breakdown.System + disk.Breakdown.Other + disk.Breakdown.User)
- // Set provider Id to the name for reconciliation
- if disk.ProviderID == "" {
- disk.ProviderID = disk.Name
- }
- }
- return diskMap, nil
- }
- type NodeOverhead struct {
- CpuOverheadFraction float64
- RamOverheadFraction float64
- }
- type Node struct {
- Cluster string
- Name string
- ProviderID string
- NodeType string
- CPUCost float64
- CPUCores float64
- GPUCost float64
- GPUCount float64
- RAMCost float64
- RAMBytes float64
- Discount float64
- Preemptible bool
- CPUBreakdown *ClusterCostsBreakdown
- RAMBreakdown *ClusterCostsBreakdown
- Start time.Time
- End time.Time
- Minutes float64
- Labels map[string]string
- CostPerCPUHr float64
- CostPerRAMGiBHr float64
- CostPerGPUHr float64
- Overhead *NodeOverhead
- }
- // GKE lies about the number of cores e2 nodes have. This table
- // contains a mapping from node type -> actual CPU cores
- // for those cases.
- var partialCPUMap = map[string]float64{
- "e2-micro": 0.25,
- "e2-small": 0.5,
- "e2-medium": 1.0,
- }
- type NodeIdentifier struct {
- Cluster string
- Name string
- ProviderID string
- }
- type nodeIdentifierNoProviderID struct {
- Cluster string
- Name string
- }
- func costTimesMinuteAndCount(activeDataMap map[NodeIdentifier]activeData, costMap map[NodeIdentifier]float64, resourceCountMap map[nodeIdentifierNoProviderID]float64) {
- for k, v := range activeDataMap {
- keyNon := nodeIdentifierNoProviderID{
- Cluster: k.Cluster,
- Name: k.Name,
- }
- if cost, ok := costMap[k]; ok {
- minutes := v.minutes
- count := 1.0
- if c, ok := resourceCountMap[keyNon]; ok {
- count = c
- }
- costMap[k] = cost * (minutes / 60) * count
- }
- }
- }
- func costTimesMinute(activeDataMap map[NodeIdentifier]activeData, costMap map[NodeIdentifier]float64) {
- for k, v := range activeDataMap {
- if cost, ok := costMap[k]; ok {
- minutes := v.minutes
- costMap[k] = cost * (minutes / 60)
- }
- }
- }
- func ClusterNodes(cp models.Provider, client prometheus.Client, start, end time.Time) (map[NodeIdentifier]*Node, error) {
- // Start from the time "end", querying backwards
- t := end
- // minsPerResolution determines accuracy and resource use for the following
- // queries. Smaller values (higher resolution) result in better accuracy,
- // but more expensive queries, and vice-a-versa.
- resolution := env.GetETLResolution()
- //Ensuring if ETL_RESOLUTION_SECONDS is less than 60s default it to 1m
- var minsPerResolution int
- if minsPerResolution = int(resolution.Minutes()); int(resolution.Minutes()) == 0 {
- minsPerResolution = 1
- log.DedupedWarningf(3, "ClusterNodes(): Configured ETL resolution (%d seconds) is below the 60 seconds threshold. Overriding with 1 minute.", int(resolution.Seconds()))
- }
- durStr := timeutil.DurationString(end.Sub(start))
- if durStr == "" {
- return nil, fmt.Errorf("illegal duration value for %s", opencost.NewClosedWindow(start, end))
- }
- requiredCtx := prom.NewNamedContext(client, prom.ClusterContextName)
- optionalCtx := prom.NewNamedContext(client, prom.ClusterOptionalContextName)
- queryNodeCPUHourlyCost := fmt.Sprintf(`avg(avg_over_time(node_cpu_hourly_cost{%s}[%s])) by (%s, node, instance_type, provider_id)`, env.GetPromClusterFilter(), durStr, env.GetPromClusterLabel())
- queryNodeCPUCoresCapacity := fmt.Sprintf(`avg(avg_over_time(kube_node_status_capacity_cpu_cores{%s}[%s])) by (%s, node)`, env.GetPromClusterFilter(), durStr, env.GetPromClusterLabel())
- queryNodeCPUCoresAllocatable := fmt.Sprintf(`avg(avg_over_time(kube_node_status_allocatable_cpu_cores{%s}[%s])) by (%s, node)`, env.GetPromClusterFilter(), durStr, env.GetPromClusterLabel())
- queryNodeRAMHourlyCost := fmt.Sprintf(`avg(avg_over_time(node_ram_hourly_cost{%s}[%s])) by (%s, node, instance_type, provider_id) / 1024 / 1024 / 1024`, env.GetPromClusterFilter(), durStr, env.GetPromClusterLabel())
- queryNodeRAMBytesCapacity := fmt.Sprintf(`avg(avg_over_time(kube_node_status_capacity_memory_bytes{%s}[%s])) by (%s, node)`, env.GetPromClusterFilter(), durStr, env.GetPromClusterLabel())
- queryNodeRAMBytesAllocatable := fmt.Sprintf(`avg(avg_over_time(kube_node_status_allocatable_memory_bytes{%s}[%s])) by (%s, node)`, env.GetPromClusterFilter(), durStr, env.GetPromClusterLabel())
- queryNodeGPUCount := fmt.Sprintf(`avg(avg_over_time(node_gpu_count{%s}[%s])) by (%s, node, provider_id)`, env.GetPromClusterFilter(), durStr, env.GetPromClusterLabel())
- queryNodeGPUHourlyCost := fmt.Sprintf(`avg(avg_over_time(node_gpu_hourly_cost{%s}[%s])) by (%s, node, instance_type, provider_id)`, env.GetPromClusterFilter(), durStr, env.GetPromClusterLabel())
- queryNodeCPUModeTotal := fmt.Sprintf(`sum(rate(node_cpu_seconds_total{%s}[%s:%dm])) by (kubernetes_node, %s, mode)`, env.GetPromClusterFilter(), durStr, minsPerResolution, env.GetPromClusterLabel())
- queryNodeRAMSystemPct := fmt.Sprintf(`sum(sum_over_time(container_memory_working_set_bytes{container_name!="POD",container_name!="",namespace="kube-system", %s}[%s:%dm])) by (instance, %s) / avg(label_replace(sum(sum_over_time(kube_node_status_capacity_memory_bytes{%s}[%s:%dm])) by (node, %s), "instance", "$1", "node", "(.*)")) by (instance, %s)`, env.GetPromClusterFilter(), durStr, minsPerResolution, env.GetPromClusterLabel(), env.GetPromClusterFilter(), durStr, minsPerResolution, env.GetPromClusterLabel(), env.GetPromClusterLabel())
- queryNodeRAMUserPct := fmt.Sprintf(`sum(sum_over_time(container_memory_working_set_bytes{container_name!="POD",container_name!="",namespace!="kube-system", %s}[%s:%dm])) by (instance, %s) / avg(label_replace(sum(sum_over_time(kube_node_status_capacity_memory_bytes{%s}[%s:%dm])) by (node, %s), "instance", "$1", "node", "(.*)")) by (instance, %s)`, env.GetPromClusterFilter(), durStr, minsPerResolution, env.GetPromClusterLabel(), env.GetPromClusterFilter(), durStr, minsPerResolution, env.GetPromClusterLabel(), env.GetPromClusterLabel())
- queryActiveMins := fmt.Sprintf(`avg(node_total_hourly_cost{%s}) by (node, %s, provider_id)[%s:%dm]`, env.GetPromClusterFilter(), env.GetPromClusterLabel(), durStr, minsPerResolution)
- queryIsSpot := fmt.Sprintf(`avg_over_time(kubecost_node_is_spot{%s}[%s:%dm])`, env.GetPromClusterFilter(), durStr, minsPerResolution)
- queryLabels := fmt.Sprintf(`count_over_time(kube_node_labels{%s}[%s:%dm])`, env.GetPromClusterFilter(), durStr, minsPerResolution)
- // Return errors if these fail
- resChNodeCPUHourlyCost := requiredCtx.QueryAtTime(queryNodeCPUHourlyCost, t)
- resChNodeCPUCoresCapacity := requiredCtx.QueryAtTime(queryNodeCPUCoresCapacity, t)
- resChNodeCPUCoresAllocatable := requiredCtx.QueryAtTime(queryNodeCPUCoresAllocatable, t)
- resChNodeRAMHourlyCost := requiredCtx.QueryAtTime(queryNodeRAMHourlyCost, t)
- resChNodeRAMBytesCapacity := requiredCtx.QueryAtTime(queryNodeRAMBytesCapacity, t)
- resChNodeRAMBytesAllocatable := requiredCtx.QueryAtTime(queryNodeRAMBytesAllocatable, t)
- resChNodeGPUCount := requiredCtx.QueryAtTime(queryNodeGPUCount, t)
- resChNodeGPUHourlyCost := requiredCtx.QueryAtTime(queryNodeGPUHourlyCost, t)
- resChActiveMins := requiredCtx.QueryAtTime(queryActiveMins, t)
- resChIsSpot := requiredCtx.QueryAtTime(queryIsSpot, t)
- // Do not return errors if these fail, but log warnings
- resChNodeCPUModeTotal := optionalCtx.QueryAtTime(queryNodeCPUModeTotal, t)
- resChNodeRAMSystemPct := optionalCtx.QueryAtTime(queryNodeRAMSystemPct, t)
- resChNodeRAMUserPct := optionalCtx.QueryAtTime(queryNodeRAMUserPct, t)
- resChLabels := optionalCtx.QueryAtTime(queryLabels, t)
- resNodeCPUHourlyCost, _ := resChNodeCPUHourlyCost.Await()
- resNodeCPUCoresCapacity, _ := resChNodeCPUCoresCapacity.Await()
- resNodeCPUCoresAllocatable, _ := resChNodeCPUCoresAllocatable.Await()
- resNodeGPUCount, _ := resChNodeGPUCount.Await()
- resNodeGPUHourlyCost, _ := resChNodeGPUHourlyCost.Await()
- resNodeRAMHourlyCost, _ := resChNodeRAMHourlyCost.Await()
- resNodeRAMBytesCapacity, _ := resChNodeRAMBytesCapacity.Await()
- resNodeRAMBytesAllocatable, _ := resChNodeRAMBytesAllocatable.Await()
- resIsSpot, _ := resChIsSpot.Await()
- resNodeCPUModeTotal, _ := resChNodeCPUModeTotal.Await()
- resNodeRAMSystemPct, _ := resChNodeRAMSystemPct.Await()
- resNodeRAMUserPct, _ := resChNodeRAMUserPct.Await()
- resActiveMins, _ := resChActiveMins.Await()
- resLabels, _ := resChLabels.Await()
- if optionalCtx.HasErrors() {
- for _, err := range optionalCtx.Errors() {
- log.Warnf("ClusterNodes: %s", err)
- }
- }
- if requiredCtx.HasErrors() {
- for _, err := range requiredCtx.Errors() {
- log.Errorf("ClusterNodes: %s", err)
- }
- return nil, requiredCtx.ErrorCollection()
- }
- activeDataMap := buildActiveDataMap(resActiveMins, resolution, opencost.NewClosedWindow(start, end))
- gpuCountMap := buildGPUCountMap(resNodeGPUCount)
- preemptibleMap := buildPreemptibleMap(resIsSpot)
- cpuCostMap, clusterAndNameToType1 := buildCPUCostMap(resNodeCPUHourlyCost, cp, preemptibleMap)
- ramCostMap, clusterAndNameToType2 := buildRAMCostMap(resNodeRAMHourlyCost, cp, preemptibleMap)
- gpuCostMap, clusterAndNameToType3 := buildGPUCostMap(resNodeGPUHourlyCost, gpuCountMap, cp, preemptibleMap)
- clusterAndNameToTypeIntermediate := mergeTypeMaps(clusterAndNameToType1, clusterAndNameToType2)
- clusterAndNameToType := mergeTypeMaps(clusterAndNameToTypeIntermediate, clusterAndNameToType3)
- cpuCoresCapacityMap := buildCPUCoresMap(resNodeCPUCoresCapacity)
- ramBytesCapacityMap := buildRAMBytesMap(resNodeRAMBytesCapacity)
- cpuCoresAllocatableMap := buildCPUCoresMap(resNodeCPUCoresAllocatable)
- ramBytesAllocatableMap := buildRAMBytesMap(resNodeRAMBytesAllocatable)
- overheadMap := buildOverheadMap(ramBytesCapacityMap, ramBytesAllocatableMap, cpuCoresCapacityMap, cpuCoresAllocatableMap)
- ramUserPctMap := buildRAMUserPctMap(resNodeRAMUserPct)
- ramSystemPctMap := buildRAMSystemPctMap(resNodeRAMSystemPct)
- cpuBreakdownMap := buildCPUBreakdownMap(resNodeCPUModeTotal)
- labelsMap := buildLabelsMap(resLabels)
- costTimesMinuteAndCount(activeDataMap, cpuCostMap, cpuCoresCapacityMap)
- costTimesMinuteAndCount(activeDataMap, ramCostMap, ramBytesCapacityMap)
- costTimesMinute(activeDataMap, gpuCostMap) // there's no need to do a weird "nodeIdentifierNoProviderID" type match since gpuCounts have a providerID
- nodeMap := buildNodeMap(
- cpuCostMap, ramCostMap, gpuCostMap, gpuCountMap,
- cpuCoresCapacityMap, ramBytesCapacityMap, ramUserPctMap,
- ramSystemPctMap,
- cpuBreakdownMap,
- activeDataMap,
- preemptibleMap,
- labelsMap,
- clusterAndNameToType,
- resolution,
- overheadMap,
- )
- c, err := cp.GetConfig()
- if err != nil {
- return nil, err
- }
- discount, err := ParsePercentString(c.Discount)
- if err != nil {
- return nil, err
- }
- negotiatedDiscount, err := ParsePercentString(c.NegotiatedDiscount)
- if err != nil {
- return nil, err
- }
- for _, node := range nodeMap {
- // TODO take GKE Reserved Instances into account
- node.Discount = cp.CombinedDiscountForNode(node.NodeType, node.Preemptible, discount, negotiatedDiscount)
- // Apply all remaining resources to Idle
- node.CPUBreakdown.Idle = 1.0 - (node.CPUBreakdown.System + node.CPUBreakdown.Other + node.CPUBreakdown.User)
- node.RAMBreakdown.Idle = 1.0 - (node.RAMBreakdown.System + node.RAMBreakdown.Other + node.RAMBreakdown.User)
- }
- return nodeMap, nil
- }
- type LoadBalancerIdentifier struct {
- Cluster string
- Namespace string
- Name string
- }
- type LoadBalancer struct {
- Cluster string
- Namespace string
- Name string
- ProviderID string
- Cost float64
- Start time.Time
- End time.Time
- Minutes float64
- Private bool
- Ip string
- }
- func ClusterLoadBalancers(client prometheus.Client, start, end time.Time) (map[LoadBalancerIdentifier]*LoadBalancer, error) {
- // Start from the time "end", querying backwards
- t := end
- // minsPerResolution determines accuracy and resource use for the following
- // queries. Smaller values (higher resolution) result in better accuracy,
- // but more expensive queries, and vice-a-versa.
- resolution := env.GetETLResolution()
- //Ensuring if ETL_RESOLUTION_SECONDS is less than 60s default it to 1m
- var minsPerResolution int
- if minsPerResolution = int(resolution.Minutes()); int(resolution.Minutes()) == 0 {
- minsPerResolution = 1
- log.DedupedWarningf(3, "ClusterLoadBalancers(): Configured ETL resolution (%d seconds) is below the 60 seconds threshold. Overriding with 1 minute.", int(resolution.Seconds()))
- }
- // Query for the duration between start and end
- durStr := timeutil.DurationString(end.Sub(start))
- if durStr == "" {
- return nil, fmt.Errorf("illegal duration value for %s", opencost.NewClosedWindow(start, end))
- }
- ctx := prom.NewNamedContext(client, prom.ClusterContextName)
- queryLBCost := fmt.Sprintf(`avg(avg_over_time(kubecost_load_balancer_cost{%s}[%s])) by (namespace, service_name, %s, ingress_ip)`, env.GetPromClusterFilter(), durStr, env.GetPromClusterLabel())
- queryActiveMins := fmt.Sprintf(`avg(kubecost_load_balancer_cost{%s}) by (namespace, service_name, %s, ingress_ip)[%s:%dm]`, env.GetPromClusterFilter(), env.GetPromClusterLabel(), durStr, minsPerResolution)
- resChLBCost := ctx.QueryAtTime(queryLBCost, t)
- resChActiveMins := ctx.QueryAtTime(queryActiveMins, t)
- resLBCost, _ := resChLBCost.Await()
- resActiveMins, _ := resChActiveMins.Await()
- if ctx.HasErrors() {
- return nil, ctx.ErrorCollection()
- }
- loadBalancerMap := make(map[LoadBalancerIdentifier]*LoadBalancer, len(resActiveMins))
- for _, result := range resActiveMins {
- cluster, err := result.GetString(env.GetPromClusterLabel())
- if err != nil {
- cluster = env.GetClusterID()
- }
- namespace, err := result.GetString("namespace")
- if err != nil {
- log.Warnf("ClusterLoadBalancers: LB cost data missing namespace")
- continue
- }
- name, err := result.GetString("service_name")
- if err != nil {
- log.Warnf("ClusterLoadBalancers: LB cost data missing service_name")
- continue
- }
- providerID, err := result.GetString("ingress_ip")
- if err != nil {
- log.DedupedWarningf(5, "ClusterLoadBalancers: LB cost data missing ingress_ip")
- providerID = ""
- }
- key := LoadBalancerIdentifier{
- Cluster: cluster,
- Namespace: namespace,
- Name: name,
- }
- // Skip if there are no data
- if len(result.Values) == 0 {
- continue
- }
- // Add load balancer to the set of load balancers
- if _, ok := loadBalancerMap[key]; !ok {
- loadBalancerMap[key] = &LoadBalancer{
- Cluster: cluster,
- Namespace: namespace,
- Name: fmt.Sprintf("%s/%s", namespace, name), // TODO:ETL this is kept for backwards-compatibility, but not good
- ProviderID: provider.ParseLBID(providerID),
- }
- }
- // Append start, end, and minutes. This should come before all other data.
- s := time.Unix(int64(result.Values[0].Timestamp), 0)
- e := time.Unix(int64(result.Values[len(result.Values)-1].Timestamp), 0)
- loadBalancerMap[key].Start = s
- loadBalancerMap[key].End = e
- loadBalancerMap[key].Minutes = e.Sub(s).Minutes()
- // Fill in Provider ID if it is available and missing in the loadBalancerMap
- // Prevents there from being a duplicate LoadBalancers on the same day
- if providerID != "" && loadBalancerMap[key].ProviderID == "" {
- loadBalancerMap[key].ProviderID = providerID
- }
- }
- for _, result := range resLBCost {
- cluster, err := result.GetString(env.GetPromClusterLabel())
- if err != nil {
- cluster = env.GetClusterID()
- }
- namespace, err := result.GetString("namespace")
- if err != nil {
- log.Warnf("ClusterLoadBalancers: LB cost data missing namespace")
- continue
- }
- name, err := result.GetString("service_name")
- if err != nil {
- log.Warnf("ClusterLoadBalancers: LB cost data missing service_name")
- continue
- }
- providerID, err := result.GetString("ingress_ip")
- if err != nil {
- log.DedupedWarningf(5, "ClusterLoadBalancers: LB cost data missing ingress_ip")
- // only update asset cost when an actual IP was returned
- continue
- }
- key := LoadBalancerIdentifier{
- Cluster: cluster,
- Namespace: namespace,
- Name: name,
- }
- // Apply cost as price-per-hour * hours
- if lb, ok := loadBalancerMap[key]; ok {
- lbPricePerHr := result.Values[0].Value
- // interpolate any missing data
- resultMins := lb.Minutes
- if resultMins > 0 {
- scaleFactor := (resultMins + resolution.Minutes()) / resultMins
- hrs := (lb.Minutes * scaleFactor) / 60.0
- lb.Cost += lbPricePerHr * hrs
- } else {
- log.DedupedWarningf(20, "ClusterLoadBalancers: found zero minutes for key: %v", key)
- }
- if lb.Ip != "" && lb.Ip != providerID {
- log.DedupedWarningf(5, "ClusterLoadBalancers: multiple IPs per load balancer not supported, using most recent IP")
- }
- lb.Ip = providerID
- lb.Private = privateIPCheck(providerID)
- } else {
- log.DedupedWarningf(20, "ClusterLoadBalancers: found minutes for key that does not exist: %v", key)
- }
- }
- return loadBalancerMap, nil
- }
- // Check if an ip is private.
- func privateIPCheck(ip string) bool {
- ipAddress := net.ParseIP(ip)
- return ipAddress.IsPrivate()
- }
- // ComputeClusterCosts gives the cumulative and monthly-rate cluster costs over a window of time for all clusters.
- func (a *Accesses) ComputeClusterCosts(client prometheus.Client, provider models.Provider, window, offset time.Duration, withBreakdown bool) (map[string]*ClusterCosts, error) {
- if window < 10*time.Minute {
- return nil, fmt.Errorf("minimum window of 10m required; got %s", window)
- }
- // Compute number of minutes in the full interval, for use interpolating missed scrapes or scaling missing data
- start, end := timeutil.ParseTimeRange(window, offset)
- mins := end.Sub(start).Minutes()
- // minsPerResolution determines accuracy and resource use for the following
- // queries. Smaller values (higher resolution) result in better accuracy,
- // but more expensive queries, and vice-a-versa.
- resolution := env.GetETLResolution()
- //Ensuring if ETL_RESOLUTION_SECONDS is less than 60s default it to 1m
- var minsPerResolution int
- if minsPerResolution = int(resolution.Minutes()); int(resolution.Minutes()) < 1 {
- minsPerResolution = 1
- log.DedupedWarningf(3, "ComputeClusterCosts(): Configured ETL resolution (%d seconds) is below the 60 seconds threshold. Overriding with 1 minute.", int(resolution.Seconds()))
- }
- windowStr := timeutil.DurationString(window)
- // hourlyToCumulative is a scaling factor that, when multiplied by an hourly
- // value, converts it to a cumulative value; i.e.
- // [$/hr] * [min/res]*[hr/min] = [$/res]
- hourlyToCumulative := float64(minsPerResolution) * (1.0 / 60.0)
- const fmtQueryDataCount = `
- count_over_time(sum(kube_node_status_capacity_cpu_cores{%s}) by (%s)[%s:%dm]%s) * %d
- `
- const fmtQueryTotalGPU = `
- sum(
- sum_over_time(node_gpu_hourly_cost{%s}[%s:%dm]%s) * %f
- ) by (%s)
- `
- const fmtQueryTotalCPU = `
- sum(
- sum_over_time(avg(kube_node_status_capacity_cpu_cores{%s}) by (node, %s)[%s:%dm]%s) *
- avg(avg_over_time(node_cpu_hourly_cost{%s}[%s:%dm]%s)) by (node, %s) * %f
- ) by (%s)
- `
- const fmtQueryTotalRAM = `
- sum(
- sum_over_time(avg(kube_node_status_capacity_memory_bytes{%s}) by (node, %s)[%s:%dm]%s) / 1024 / 1024 / 1024 *
- avg(avg_over_time(node_ram_hourly_cost{%s}[%s:%dm]%s)) by (node, %s) * %f
- ) by (%s)
- `
- const fmtQueryTotalStorage = `
- sum(
- sum_over_time(avg(kube_persistentvolume_capacity_bytes{%s}) by (persistentvolume, %s)[%s:%dm]%s) / 1024 / 1024 / 1024 *
- avg(avg_over_time(pv_hourly_cost{%s}[%s:%dm]%s)) by (persistentvolume, %s) * %f
- ) by (%s)
- `
- const fmtQueryCPUModePct = `
- sum(rate(node_cpu_seconds_total{%s}[%s]%s)) by (%s, mode) / ignoring(mode)
- group_left sum(rate(node_cpu_seconds_total{%s}[%s]%s)) by (%s)
- `
- const fmtQueryRAMSystemPct = `
- sum(sum_over_time(container_memory_usage_bytes{container_name!="",namespace="kube-system", %s}[%s:%dm]%s)) by (%s)
- / sum(sum_over_time(kube_node_status_capacity_memory_bytes{%s}[%s:%dm]%s)) by (%s)
- `
- const fmtQueryRAMUserPct = `
- sum(sum_over_time(kubecost_cluster_memory_working_set_bytes{%s}[%s:%dm]%s)) by (%s)
- / sum(sum_over_time(kube_node_status_capacity_memory_bytes{%s}[%s:%dm]%s)) by (%s)
- `
- // TODO niko/clustercost metric "kubelet_volume_stats_used_bytes" was deprecated in 1.12, then seems to have come back in 1.17
- // const fmtQueryPVStorageUsePct = `(sum(kube_persistentvolumeclaim_info) by (persistentvolumeclaim, storageclass,namespace) + on (persistentvolumeclaim,namespace)
- // group_right(storageclass) sum(kubelet_volume_stats_used_bytes) by (persistentvolumeclaim,namespace))`
- queryUsedLocalStorage := provider.GetLocalStorageQuery(window, offset, false, true)
- queryTotalLocalStorage := provider.GetLocalStorageQuery(window, offset, false, false)
- if queryTotalLocalStorage != "" {
- queryTotalLocalStorage = fmt.Sprintf(" + %s", queryTotalLocalStorage)
- }
- fmtOffset := timeutil.DurationToPromOffsetString(offset)
- queryDataCount := fmt.Sprintf(fmtQueryDataCount, env.GetPromClusterFilter(), env.GetPromClusterLabel(), windowStr, minsPerResolution, fmtOffset, minsPerResolution)
- queryTotalGPU := fmt.Sprintf(fmtQueryTotalGPU, env.GetPromClusterFilter(), windowStr, minsPerResolution, fmtOffset, hourlyToCumulative, env.GetPromClusterLabel())
- queryTotalCPU := fmt.Sprintf(fmtQueryTotalCPU, env.GetPromClusterFilter(), env.GetPromClusterLabel(), windowStr, minsPerResolution, fmtOffset, env.GetPromClusterFilter(), windowStr, minsPerResolution, fmtOffset, env.GetPromClusterLabel(), hourlyToCumulative, env.GetPromClusterLabel())
- queryTotalRAM := fmt.Sprintf(fmtQueryTotalRAM, env.GetPromClusterFilter(), env.GetPromClusterLabel(), windowStr, minsPerResolution, fmtOffset, env.GetPromClusterFilter(), windowStr, minsPerResolution, fmtOffset, env.GetPromClusterLabel(), hourlyToCumulative, env.GetPromClusterLabel())
- queryTotalStorage := fmt.Sprintf(fmtQueryTotalStorage, env.GetPromClusterFilter(), env.GetPromClusterLabel(), windowStr, minsPerResolution, fmtOffset, env.GetPromClusterFilter(), windowStr, minsPerResolution, fmtOffset, env.GetPromClusterLabel(), hourlyToCumulative, env.GetPromClusterLabel())
- ctx := prom.NewNamedContext(client, prom.ClusterContextName)
- resChs := ctx.QueryAll(
- queryDataCount,
- queryTotalGPU,
- queryTotalCPU,
- queryTotalRAM,
- queryTotalStorage,
- )
- // Only submit the local storage query if it is valid. Otherwise Prometheus
- // will return errors. Always append something to resChs, regardless, to
- // maintain indexing.
- if queryTotalLocalStorage != "" {
- resChs = append(resChs, ctx.Query(queryTotalLocalStorage))
- } else {
- resChs = append(resChs, nil)
- }
- if withBreakdown {
- queryCPUModePct := fmt.Sprintf(fmtQueryCPUModePct, env.GetPromClusterFilter(), windowStr, fmtOffset, env.GetPromClusterLabel(), env.GetPromClusterFilter(), windowStr, fmtOffset, env.GetPromClusterLabel())
- queryRAMSystemPct := fmt.Sprintf(fmtQueryRAMSystemPct, env.GetPromClusterFilter(), windowStr, minsPerResolution, fmtOffset, env.GetPromClusterLabel(), env.GetPromClusterFilter(), windowStr, minsPerResolution, fmtOffset, env.GetPromClusterLabel())
- queryRAMUserPct := fmt.Sprintf(fmtQueryRAMUserPct, env.GetPromClusterFilter(), windowStr, minsPerResolution, fmtOffset, env.GetPromClusterLabel(), env.GetPromClusterFilter(), windowStr, minsPerResolution, fmtOffset, env.GetPromClusterLabel())
- bdResChs := ctx.QueryAll(
- queryCPUModePct,
- queryRAMSystemPct,
- queryRAMUserPct,
- )
- // Only submit the local storage query if it is valid. Otherwise Prometheus
- // will return errors. Always append something to resChs, regardless, to
- // maintain indexing.
- if queryUsedLocalStorage != "" {
- bdResChs = append(bdResChs, ctx.Query(queryUsedLocalStorage))
- } else {
- bdResChs = append(bdResChs, nil)
- }
- resChs = append(resChs, bdResChs...)
- }
- resDataCount, _ := resChs[0].Await()
- resTotalGPU, _ := resChs[1].Await()
- resTotalCPU, _ := resChs[2].Await()
- resTotalRAM, _ := resChs[3].Await()
- resTotalStorage, _ := resChs[4].Await()
- if ctx.HasErrors() {
- return nil, ctx.ErrorCollection()
- }
- defaultClusterID := env.GetClusterID()
- dataMinsByCluster := map[string]float64{}
- for _, result := range resDataCount {
- clusterID, _ := result.GetString(env.GetPromClusterLabel())
- if clusterID == "" {
- clusterID = defaultClusterID
- }
- dataMins := mins
- if len(result.Values) > 0 {
- dataMins = result.Values[0].Value
- } else {
- log.Warnf("Cluster cost data count returned no results for cluster %s", clusterID)
- }
- dataMinsByCluster[clusterID] = dataMins
- }
- // Determine combined discount
- discount, customDiscount := 0.0, 0.0
- c, err := a.CloudProvider.GetConfig()
- if err == nil {
- discount, err = ParsePercentString(c.Discount)
- if err != nil {
- discount = 0.0
- }
- customDiscount, err = ParsePercentString(c.NegotiatedDiscount)
- if err != nil {
- customDiscount = 0.0
- }
- }
- // Intermediate structure storing mapping of [clusterID][type ∈ {cpu, ram, storage, total}]=cost
- costData := make(map[string]map[string]float64)
- // Helper function to iterate over Prom query results, parsing the raw values into
- // the intermediate costData structure.
- setCostsFromResults := func(costData map[string]map[string]float64, results []*prom.QueryResult, name string, discount float64, customDiscount float64) {
- for _, result := range results {
- clusterID, _ := result.GetString(env.GetPromClusterLabel())
- if clusterID == "" {
- clusterID = defaultClusterID
- }
- if _, ok := costData[clusterID]; !ok {
- costData[clusterID] = map[string]float64{}
- }
- if len(result.Values) > 0 {
- costData[clusterID][name] += result.Values[0].Value * (1.0 - discount) * (1.0 - customDiscount)
- costData[clusterID]["total"] += result.Values[0].Value * (1.0 - discount) * (1.0 - customDiscount)
- }
- }
- }
- // Apply both sustained use and custom discounts to RAM and CPU
- setCostsFromResults(costData, resTotalCPU, "cpu", discount, customDiscount)
- setCostsFromResults(costData, resTotalRAM, "ram", discount, customDiscount)
- // Apply only custom discount to GPU and storage
- setCostsFromResults(costData, resTotalGPU, "gpu", 0.0, customDiscount)
- setCostsFromResults(costData, resTotalStorage, "storage", 0.0, customDiscount)
- if queryTotalLocalStorage != "" {
- resTotalLocalStorage, err := resChs[5].Await()
- if err != nil {
- return nil, err
- }
- setCostsFromResults(costData, resTotalLocalStorage, "localstorage", 0.0, customDiscount)
- }
- cpuBreakdownMap := map[string]*ClusterCostsBreakdown{}
- ramBreakdownMap := map[string]*ClusterCostsBreakdown{}
- pvUsedCostMap := map[string]float64{}
- if withBreakdown {
- resCPUModePct, _ := resChs[6].Await()
- resRAMSystemPct, _ := resChs[7].Await()
- resRAMUserPct, _ := resChs[8].Await()
- if ctx.HasErrors() {
- return nil, ctx.ErrorCollection()
- }
- for _, result := range resCPUModePct {
- clusterID, _ := result.GetString(env.GetPromClusterLabel())
- if clusterID == "" {
- clusterID = defaultClusterID
- }
- if _, ok := cpuBreakdownMap[clusterID]; !ok {
- cpuBreakdownMap[clusterID] = &ClusterCostsBreakdown{}
- }
- cpuBD := cpuBreakdownMap[clusterID]
- mode, err := result.GetString("mode")
- if err != nil {
- log.Warnf("ComputeClusterCosts: unable to read CPU mode: %s", err)
- mode = "other"
- }
- switch mode {
- case "idle":
- cpuBD.Idle += result.Values[0].Value
- case "system":
- cpuBD.System += result.Values[0].Value
- case "user":
- cpuBD.User += result.Values[0].Value
- default:
- cpuBD.Other += result.Values[0].Value
- }
- }
- for _, result := range resRAMSystemPct {
- clusterID, _ := result.GetString(env.GetPromClusterLabel())
- if clusterID == "" {
- clusterID = defaultClusterID
- }
- if _, ok := ramBreakdownMap[clusterID]; !ok {
- ramBreakdownMap[clusterID] = &ClusterCostsBreakdown{}
- }
- ramBD := ramBreakdownMap[clusterID]
- ramBD.System += result.Values[0].Value
- }
- for _, result := range resRAMUserPct {
- clusterID, _ := result.GetString(env.GetPromClusterLabel())
- if clusterID == "" {
- clusterID = defaultClusterID
- }
- if _, ok := ramBreakdownMap[clusterID]; !ok {
- ramBreakdownMap[clusterID] = &ClusterCostsBreakdown{}
- }
- ramBD := ramBreakdownMap[clusterID]
- ramBD.User += result.Values[0].Value
- }
- for _, ramBD := range ramBreakdownMap {
- remaining := 1.0
- remaining -= ramBD.Other
- remaining -= ramBD.System
- remaining -= ramBD.User
- ramBD.Idle = remaining
- }
- if queryUsedLocalStorage != "" {
- resUsedLocalStorage, err := resChs[9].Await()
- if err != nil {
- return nil, err
- }
- for _, result := range resUsedLocalStorage {
- clusterID, _ := result.GetString(env.GetPromClusterLabel())
- if clusterID == "" {
- clusterID = defaultClusterID
- }
- pvUsedCostMap[clusterID] += result.Values[0].Value
- }
- }
- }
- if ctx.HasErrors() {
- for _, err := range ctx.Errors() {
- log.Errorf("ComputeClusterCosts: %s", err)
- }
- return nil, ctx.ErrorCollection()
- }
- // Convert intermediate structure to Costs instances
- costsByCluster := map[string]*ClusterCosts{}
- for id, cd := range costData {
- dataMins, ok := dataMinsByCluster[id]
- if !ok {
- dataMins = mins
- log.Warnf("Cluster cost data count not found for cluster %s", id)
- }
- costs, err := NewClusterCostsFromCumulative(cd["cpu"], cd["gpu"], cd["ram"], cd["storage"]+cd["localstorage"], window, offset, dataMins/timeutil.MinsPerHour)
- if err != nil {
- log.Warnf("Failed to parse cluster costs on %s (%s) from cumulative data: %+v", window, offset, cd)
- return nil, err
- }
- if cpuBD, ok := cpuBreakdownMap[id]; ok {
- costs.CPUBreakdown = cpuBD
- }
- if ramBD, ok := ramBreakdownMap[id]; ok {
- costs.RAMBreakdown = ramBD
- }
- costs.StorageBreakdown = &ClusterCostsBreakdown{}
- if pvUC, ok := pvUsedCostMap[id]; ok {
- costs.StorageBreakdown.Idle = (costs.StorageCumulative - pvUC) / costs.StorageCumulative
- costs.StorageBreakdown.User = pvUC / costs.StorageCumulative
- }
- costs.DataMinutes = dataMins
- costsByCluster[id] = costs
- }
- return costsByCluster, nil
- }
- type Totals struct {
- TotalCost [][]string `json:"totalcost"`
- CPUCost [][]string `json:"cpucost"`
- MemCost [][]string `json:"memcost"`
- StorageCost [][]string `json:"storageCost"`
- }
- func resultToTotals(qrs []*prom.QueryResult) ([][]string, error) {
- if len(qrs) == 0 {
- return [][]string{}, fmt.Errorf("Not enough data available in the selected time range")
- }
- result := qrs[0]
- totals := [][]string{}
- for _, value := range result.Values {
- d0 := fmt.Sprintf("%f", value.Timestamp)
- d1 := fmt.Sprintf("%f", value.Value)
- toAppend := []string{
- d0,
- d1,
- }
- totals = append(totals, toAppend)
- }
- return totals, nil
- }
- // ClusterCostsOverTime gives the full cluster costs over time
- func ClusterCostsOverTime(cli prometheus.Client, provider models.Provider, startString, endString string, window, offset time.Duration) (*Totals, error) {
- localStorageQuery := provider.GetLocalStorageQuery(window, offset, true, false)
- if localStorageQuery != "" {
- localStorageQuery = fmt.Sprintf("+ %s", localStorageQuery)
- }
- layout := "2006-01-02T15:04:05.000Z"
- start, err := time.Parse(layout, startString)
- if err != nil {
- log.Errorf("Error parsing time %s. Error: %s", startString, err.Error())
- return nil, err
- }
- end, err := time.Parse(layout, endString)
- if err != nil {
- log.Errorf("Error parsing time %s. Error: %s", endString, err.Error())
- return nil, err
- }
- fmtWindow := timeutil.DurationString(window)
- if fmtWindow == "" {
- err := fmt.Errorf("window value invalid or missing")
- log.Errorf("Error parsing time %v. Error: %s", window, err.Error())
- return nil, err
- }
- fmtOffset := timeutil.DurationToPromOffsetString(offset)
- qCores := fmt.Sprintf(queryClusterCores, env.GetPromClusterFilter(), fmtWindow, fmtOffset, env.GetPromClusterLabel(), env.GetPromClusterFilter(), fmtWindow, fmtOffset, env.GetPromClusterLabel(), env.GetPromClusterFilter(), fmtWindow, fmtOffset, env.GetPromClusterLabel(), env.GetPromClusterLabel())
- qRAM := fmt.Sprintf(queryClusterRAM, env.GetPromClusterFilter(), fmtWindow, fmtOffset, env.GetPromClusterLabel(), env.GetPromClusterFilter(), fmtWindow, fmtOffset, env.GetPromClusterLabel(), env.GetPromClusterLabel())
- qStorage := fmt.Sprintf(queryStorage, env.GetPromClusterFilter(), fmtWindow, fmtOffset, env.GetPromClusterLabel(), env.GetPromClusterFilter(), fmtWindow, fmtOffset, env.GetPromClusterLabel(), env.GetPromClusterLabel(), localStorageQuery)
- qTotal := fmt.Sprintf(queryTotal, env.GetPromClusterFilter(), env.GetPromClusterLabel(), env.GetPromClusterFilter(), env.GetPromClusterLabel(), env.GetPromClusterFilter(), env.GetPromClusterLabel(), env.GetPromClusterLabel(), localStorageQuery)
- ctx := prom.NewNamedContext(cli, prom.ClusterContextName)
- resChClusterCores := ctx.QueryRange(qCores, start, end, window)
- resChClusterRAM := ctx.QueryRange(qRAM, start, end, window)
- resChStorage := ctx.QueryRange(qStorage, start, end, window)
- resChTotal := ctx.QueryRange(qTotal, start, end, window)
- resultClusterCores, err := resChClusterCores.Await()
- if err != nil {
- return nil, err
- }
- resultClusterRAM, err := resChClusterRAM.Await()
- if err != nil {
- return nil, err
- }
- resultStorage, err := resChStorage.Await()
- if err != nil {
- return nil, err
- }
- resultTotal, err := resChTotal.Await()
- if err != nil {
- return nil, err
- }
- coreTotal, err := resultToTotals(resultClusterCores)
- if err != nil {
- log.Infof("[Warning] ClusterCostsOverTime: no cpu data: %s", err)
- return nil, err
- }
- ramTotal, err := resultToTotals(resultClusterRAM)
- if err != nil {
- log.Infof("[Warning] ClusterCostsOverTime: no ram data: %s", err)
- return nil, err
- }
- storageTotal, err := resultToTotals(resultStorage)
- if err != nil {
- log.Infof("[Warning] ClusterCostsOverTime: no storage data: %s", err)
- }
- clusterTotal, err := resultToTotals(resultTotal)
- if err != nil {
- // If clusterTotal query failed, it's likely because there are no PVs, which
- // causes the qTotal query to return no data. Instead, query only node costs.
- // If that fails, return an error because something is actually wrong.
- qNodes := fmt.Sprintf(queryNodes, env.GetPromClusterFilter(), env.GetPromClusterLabel(), localStorageQuery)
- resultNodes, warnings, err := ctx.QueryRangeSync(qNodes, start, end, window)
- for _, warning := range warnings {
- log.Warnf(warning)
- }
- if err != nil {
- return nil, err
- }
- clusterTotal, err = resultToTotals(resultNodes)
- if err != nil {
- log.Infof("[Warning] ClusterCostsOverTime: no node data: %s", err)
- return nil, err
- }
- }
- return &Totals{
- TotalCost: clusterTotal,
- CPUCost: coreTotal,
- MemCost: ramTotal,
- StorageCost: storageTotal,
- }, nil
- }
- func pvCosts(diskMap map[DiskIdentifier]*Disk, resolution time.Duration, resActiveMins, resPVSize, resPVCost, resPVUsedAvg, resPVUsedMax, resPVCInfo []*prom.QueryResult, cp models.Provider, window opencost.Window) {
- for _, result := range resActiveMins {
- cluster, err := result.GetString(env.GetPromClusterLabel())
- if err != nil {
- cluster = env.GetClusterID()
- }
- name, err := result.GetString("persistentvolume")
- if err != nil {
- log.Warnf("ClusterDisks: active mins missing pv name")
- continue
- }
- if len(result.Values) == 0 {
- continue
- }
- key := DiskIdentifier{cluster, name}
- if _, ok := diskMap[key]; !ok {
- diskMap[key] = &Disk{
- Cluster: cluster,
- Name: name,
- Breakdown: &ClusterCostsBreakdown{},
- }
- }
- s, e := calculateStartAndEnd(result, resolution, window)
- mins := e.Sub(s).Minutes()
- diskMap[key].End = e
- diskMap[key].Start = s
- diskMap[key].Minutes = mins
- }
- for _, result := range resPVSize {
- cluster, err := result.GetString(env.GetPromClusterLabel())
- if err != nil {
- cluster = env.GetClusterID()
- }
- name, err := result.GetString("persistentvolume")
- if err != nil {
- log.Warnf("ClusterDisks: PV size data missing persistentvolume")
- continue
- }
- // TODO niko/assets storage class
- bytes := result.Values[0].Value
- key := DiskIdentifier{cluster, name}
- if _, ok := diskMap[key]; !ok {
- diskMap[key] = &Disk{
- Cluster: cluster,
- Name: name,
- Breakdown: &ClusterCostsBreakdown{},
- }
- }
- diskMap[key].Bytes = bytes
- }
- customPricingEnabled := provider.CustomPricesEnabled(cp)
- customPricingConfig, err := cp.GetConfig()
- if err != nil {
- log.Warnf("ClusterDisks: failed to load custom pricing: %s", err)
- }
- for _, result := range resPVCost {
- cluster, err := result.GetString(env.GetPromClusterLabel())
- if err != nil {
- cluster = env.GetClusterID()
- }
- name, err := result.GetString("persistentvolume")
- if err != nil {
- log.Warnf("ClusterDisks: PV cost data missing persistentvolume")
- continue
- }
- // TODO niko/assets storage class
- var cost float64
- if customPricingEnabled && customPricingConfig != nil {
- customPVCostStr := customPricingConfig.Storage
- customPVCost, err := strconv.ParseFloat(customPVCostStr, 64)
- if err != nil {
- log.Warnf("ClusterDisks: error parsing custom PV price: %s", customPVCostStr)
- }
- cost = customPVCost
- } else {
- cost = result.Values[0].Value
- }
- key := DiskIdentifier{cluster, name}
- if _, ok := diskMap[key]; !ok {
- diskMap[key] = &Disk{
- Cluster: cluster,
- Name: name,
- Breakdown: &ClusterCostsBreakdown{},
- }
- }
- diskMap[key].Cost = cost * (diskMap[key].Bytes / 1024 / 1024 / 1024) * (diskMap[key].Minutes / 60)
- providerID, _ := result.GetString("provider_id") // just put the providerID set up here, it's the simplest query.
- if providerID != "" {
- diskMap[key].ProviderID = provider.ParsePVID(providerID)
- }
- }
- for _, result := range resPVUsedAvg {
- cluster, err := result.GetString(env.GetPromClusterLabel())
- if err != nil {
- cluster = env.GetClusterID()
- }
- claimName, err := result.GetString("persistentvolumeclaim")
- if err != nil {
- log.Debugf("ClusterDisks: pv usage data missing persistentvolumeclaim")
- continue
- }
- claimNamespace, err := result.GetString("namespace")
- if err != nil {
- log.Debugf("ClusterDisks: pv usage data missing namespace")
- continue
- }
- var volumeName string
- for _, thatRes := range resPVCInfo {
- thatCluster, err := thatRes.GetString(env.GetPromClusterLabel())
- if err != nil {
- thatCluster = env.GetClusterID()
- }
- thatVolumeName, err := thatRes.GetString("volumename")
- if err != nil {
- log.Debugf("ClusterDisks: pv claim data missing volumename")
- continue
- }
- thatClaimName, err := thatRes.GetString("persistentvolumeclaim")
- if err != nil {
- log.Debugf("ClusterDisks: pv claim data missing persistentvolumeclaim")
- continue
- }
- thatClaimNamespace, err := thatRes.GetString("namespace")
- if err != nil {
- log.Debugf("ClusterDisks: pv claim data missing namespace")
- continue
- }
- if cluster == thatCluster && claimName == thatClaimName && claimNamespace == thatClaimNamespace {
- volumeName = thatVolumeName
- }
- }
- usage := result.Values[0].Value
- key := DiskIdentifier{cluster, volumeName}
- if _, ok := diskMap[key]; !ok {
- diskMap[key] = &Disk{
- Cluster: cluster,
- Name: volumeName,
- Breakdown: &ClusterCostsBreakdown{},
- }
- }
- diskMap[key].BytesUsedAvgPtr = &usage
- }
- for _, result := range resPVUsedMax {
- cluster, err := result.GetString(env.GetPromClusterLabel())
- if err != nil {
- cluster = env.GetClusterID()
- }
- claimName, err := result.GetString("persistentvolumeclaim")
- if err != nil {
- log.Debugf("ClusterDisks: pv usage data missing persistentvolumeclaim")
- continue
- }
- claimNamespace, err := result.GetString("namespace")
- if err != nil {
- log.Debugf("ClusterDisks: pv usage data missing namespace")
- continue
- }
- var volumeName string
- for _, thatRes := range resPVCInfo {
- thatCluster, err := thatRes.GetString(env.GetPromClusterLabel())
- if err != nil {
- thatCluster = env.GetClusterID()
- }
- thatVolumeName, err := thatRes.GetString("volumename")
- if err != nil {
- log.Debugf("ClusterDisks: pv claim data missing volumename")
- continue
- }
- thatClaimName, err := thatRes.GetString("persistentvolumeclaim")
- if err != nil {
- log.Debugf("ClusterDisks: pv claim data missing persistentvolumeclaim")
- continue
- }
- thatClaimNamespace, err := thatRes.GetString("namespace")
- if err != nil {
- log.Debugf("ClusterDisks: pv claim data missing namespace")
- continue
- }
- if cluster == thatCluster && claimName == thatClaimName && claimNamespace == thatClaimNamespace {
- volumeName = thatVolumeName
- }
- }
- usage := result.Values[0].Value
- key := DiskIdentifier{cluster, volumeName}
- if _, ok := diskMap[key]; !ok {
- diskMap[key] = &Disk{
- Cluster: cluster,
- Name: volumeName,
- Breakdown: &ClusterCostsBreakdown{},
- }
- }
- diskMap[key].BytesUsedMaxPtr = &usage
- }
- }
|