mirror of
https://github.com/ClusterCockpit/cc-backend
synced 2024-11-10 08:57:25 +01:00
6a1cb51c2f
- Adapt to new metricConfig logic - Footprint-Metrics generalized for bar card - Footprint-Metrics in stats filter and sorting - Frontend always uses GQL, except adminOptions - Job View will load scopes for all metrics on request
765 lines
21 KiB
Go
765 lines
21 KiB
Go
// Copyright (C) NHR@FAU, University Erlangen-Nuremberg.
|
|
// All rights reserved.
|
|
// Use of this source code is governed by a MIT-style
|
|
// license that can be found in the LICENSE file.
|
|
package repository
|
|
|
|
import (
|
|
"context"
|
|
"database/sql"
|
|
"fmt"
|
|
"math"
|
|
"time"
|
|
|
|
"github.com/ClusterCockpit/cc-backend/internal/config"
|
|
"github.com/ClusterCockpit/cc-backend/internal/graph/model"
|
|
"github.com/ClusterCockpit/cc-backend/internal/metricdata"
|
|
"github.com/ClusterCockpit/cc-backend/pkg/archive"
|
|
"github.com/ClusterCockpit/cc-backend/pkg/log"
|
|
"github.com/ClusterCockpit/cc-backend/pkg/schema"
|
|
sq "github.com/Masterminds/squirrel"
|
|
)
|
|
|
|
// GraphQL validation should make sure that no unkown values can be specified.
|
|
var groupBy2column = map[model.Aggregate]string{
|
|
model.AggregateUser: "job.user",
|
|
model.AggregateProject: "job.project",
|
|
model.AggregateCluster: "job.cluster",
|
|
}
|
|
|
|
var sortBy2column = map[model.SortByAggregate]string{
|
|
model.SortByAggregateTotaljobs: "totalJobs",
|
|
model.SortByAggregateTotalwalltime: "totalWalltime",
|
|
model.SortByAggregateTotalnodes: "totalNodes",
|
|
model.SortByAggregateTotalnodehours: "totalNodeHours",
|
|
model.SortByAggregateTotalcores: "totalCores",
|
|
model.SortByAggregateTotalcorehours: "totalCoreHours",
|
|
model.SortByAggregateTotalaccs: "totalAccs",
|
|
model.SortByAggregateTotalacchours: "totalAccHours",
|
|
}
|
|
|
|
func (r *JobRepository) buildCountQuery(
|
|
filter []*model.JobFilter,
|
|
kind string,
|
|
col string,
|
|
) sq.SelectBuilder {
|
|
var query sq.SelectBuilder
|
|
|
|
if col != "" {
|
|
// Scan columns: id, cnt
|
|
query = sq.Select(col, "COUNT(job.id)").From("job").GroupBy(col)
|
|
} else {
|
|
// Scan columns: cnt
|
|
query = sq.Select("COUNT(job.id)").From("job")
|
|
}
|
|
|
|
switch kind {
|
|
case "running":
|
|
query = query.Where("job.job_state = ?", "running")
|
|
case "short":
|
|
query = query.Where("job.duration < ?", config.Keys.ShortRunningJobsDuration)
|
|
}
|
|
|
|
for _, f := range filter {
|
|
query = BuildWhereClause(f, query)
|
|
}
|
|
|
|
return query
|
|
}
|
|
|
|
func (r *JobRepository) buildStatsQuery(
|
|
filter []*model.JobFilter,
|
|
col string,
|
|
) sq.SelectBuilder {
|
|
var query sq.SelectBuilder
|
|
castType := r.getCastType()
|
|
|
|
// fmt.Sprintf(`CAST(ROUND((CASE WHEN job.job_state = "running" THEN %d - job.start_time ELSE job.duration END) / 3600) as %s) as value`, time.Now().Unix(), castType)
|
|
|
|
if col != "" {
|
|
// Scan columns: id, totalJobs, totalWalltime, totalNodes, totalNodeHours, totalCores, totalCoreHours, totalAccs, totalAccHours
|
|
query = sq.Select(col, "COUNT(job.id) as totalJobs",
|
|
fmt.Sprintf(`CAST(ROUND(SUM((CASE WHEN job.job_state = "running" THEN %d - job.start_time ELSE job.duration END)) / 3600) as %s) as totalWalltime`, time.Now().Unix(), castType),
|
|
fmt.Sprintf(`CAST(SUM(job.num_nodes) as %s) as totalNodes`, castType),
|
|
fmt.Sprintf(`CAST(ROUND(SUM((CASE WHEN job.job_state = "running" THEN %d - job.start_time ELSE job.duration END) * job.num_nodes) / 3600) as %s) as totalNodeHours`, time.Now().Unix(), castType),
|
|
fmt.Sprintf(`CAST(SUM(job.num_hwthreads) as %s) as totalCores`, castType),
|
|
fmt.Sprintf(`CAST(ROUND(SUM((CASE WHEN job.job_state = "running" THEN %d - job.start_time ELSE job.duration END) * job.num_hwthreads) / 3600) as %s) as totalCoreHours`, time.Now().Unix(), castType),
|
|
fmt.Sprintf(`CAST(SUM(job.num_acc) as %s) as totalAccs`, castType),
|
|
fmt.Sprintf(`CAST(ROUND(SUM((CASE WHEN job.job_state = "running" THEN %d - job.start_time ELSE job.duration END) * job.num_acc) / 3600) as %s) as totalAccHours`, time.Now().Unix(), castType),
|
|
).From("job").GroupBy(col)
|
|
} else {
|
|
// Scan columns: totalJobs, totalWalltime, totalNodes, totalNodeHours, totalCores, totalCoreHours, totalAccs, totalAccHours
|
|
query = sq.Select("COUNT(job.id)",
|
|
fmt.Sprintf(`CAST(ROUND(SUM((CASE WHEN job.job_state = "running" THEN %d - job.start_time ELSE job.duration END)) / 3600) as %s)`, time.Now().Unix(), castType),
|
|
fmt.Sprintf(`CAST(SUM(job.num_nodes) as %s)`, castType),
|
|
fmt.Sprintf(`CAST(ROUND(SUM((CASE WHEN job.job_state = "running" THEN %d - job.start_time ELSE job.duration END) * job.num_nodes) / 3600) as %s)`, time.Now().Unix(), castType),
|
|
fmt.Sprintf(`CAST(SUM(job.num_hwthreads) as %s)`, castType),
|
|
fmt.Sprintf(`CAST(ROUND(SUM((CASE WHEN job.job_state = "running" THEN %d - job.start_time ELSE job.duration END) * job.num_hwthreads) / 3600) as %s)`, time.Now().Unix(), castType),
|
|
fmt.Sprintf(`CAST(SUM(job.num_acc) as %s)`, castType),
|
|
fmt.Sprintf(`CAST(ROUND(SUM((CASE WHEN job.job_state = "running" THEN %d - job.start_time ELSE job.duration END) * job.num_acc) / 3600) as %s)`, time.Now().Unix(), castType),
|
|
).From("job")
|
|
}
|
|
|
|
for _, f := range filter {
|
|
query = BuildWhereClause(f, query)
|
|
}
|
|
|
|
return query
|
|
}
|
|
|
|
func (r *JobRepository) getUserName(ctx context.Context, id string) string {
|
|
user := GetUserFromContext(ctx)
|
|
name, _ := r.FindColumnValue(user, id, "user", "name", "username", false)
|
|
if name != "" {
|
|
return name
|
|
} else {
|
|
return "-"
|
|
}
|
|
}
|
|
|
|
func (r *JobRepository) getCastType() string {
|
|
var castType string
|
|
|
|
switch r.driver {
|
|
case "sqlite3":
|
|
castType = "int"
|
|
case "mysql":
|
|
castType = "unsigned"
|
|
default:
|
|
castType = ""
|
|
}
|
|
|
|
return castType
|
|
}
|
|
|
|
func (r *JobRepository) JobsStatsGrouped(
|
|
ctx context.Context,
|
|
filter []*model.JobFilter,
|
|
page *model.PageRequest,
|
|
sortBy *model.SortByAggregate,
|
|
groupBy *model.Aggregate,
|
|
) ([]*model.JobsStatistics, error) {
|
|
start := time.Now()
|
|
col := groupBy2column[*groupBy]
|
|
query := r.buildStatsQuery(filter, col)
|
|
|
|
query, err := SecurityCheck(ctx, query)
|
|
if err != nil {
|
|
return nil, err
|
|
}
|
|
|
|
if sortBy != nil {
|
|
sortBy := sortBy2column[*sortBy]
|
|
query = query.OrderBy(fmt.Sprintf("%s DESC", sortBy))
|
|
}
|
|
if page != nil && page.ItemsPerPage != -1 {
|
|
limit := uint64(page.ItemsPerPage)
|
|
query = query.Offset((uint64(page.Page) - 1) * limit).Limit(limit)
|
|
}
|
|
|
|
rows, err := query.RunWith(r.DB).Query()
|
|
if err != nil {
|
|
log.Warn("Error while querying DB for job statistics")
|
|
return nil, err
|
|
}
|
|
|
|
stats := make([]*model.JobsStatistics, 0, 100)
|
|
|
|
for rows.Next() {
|
|
var id sql.NullString
|
|
var jobs, walltime, nodes, nodeHours, cores, coreHours, accs, accHours sql.NullInt64
|
|
if err := rows.Scan(&id, &jobs, &walltime, &nodes, &nodeHours, &cores, &coreHours, &accs, &accHours); err != nil {
|
|
log.Warn("Error while scanning rows")
|
|
return nil, err
|
|
}
|
|
|
|
if id.Valid {
|
|
var totalJobs, totalWalltime, totalNodes, totalNodeHours, totalCores, totalCoreHours, totalAccs, totalAccHours int
|
|
|
|
if jobs.Valid {
|
|
totalJobs = int(jobs.Int64)
|
|
}
|
|
|
|
if walltime.Valid {
|
|
totalWalltime = int(walltime.Int64)
|
|
}
|
|
|
|
if nodes.Valid {
|
|
totalNodes = int(nodes.Int64)
|
|
}
|
|
if cores.Valid {
|
|
totalCores = int(cores.Int64)
|
|
}
|
|
if accs.Valid {
|
|
totalAccs = int(accs.Int64)
|
|
}
|
|
|
|
if nodeHours.Valid {
|
|
totalNodeHours = int(nodeHours.Int64)
|
|
}
|
|
if coreHours.Valid {
|
|
totalCoreHours = int(coreHours.Int64)
|
|
}
|
|
if accHours.Valid {
|
|
totalAccHours = int(accHours.Int64)
|
|
}
|
|
|
|
if col == "job.user" {
|
|
name := r.getUserName(ctx, id.String)
|
|
stats = append(stats,
|
|
&model.JobsStatistics{
|
|
ID: id.String,
|
|
Name: name,
|
|
TotalJobs: totalJobs,
|
|
TotalWalltime: totalWalltime,
|
|
TotalNodes: totalNodes,
|
|
TotalNodeHours: totalNodeHours,
|
|
TotalCores: totalCores,
|
|
TotalCoreHours: totalCoreHours,
|
|
TotalAccs: totalAccs,
|
|
TotalAccHours: totalAccHours,
|
|
})
|
|
} else {
|
|
stats = append(stats,
|
|
&model.JobsStatistics{
|
|
ID: id.String,
|
|
TotalJobs: int(jobs.Int64),
|
|
TotalWalltime: int(walltime.Int64),
|
|
TotalNodes: totalNodes,
|
|
TotalNodeHours: totalNodeHours,
|
|
TotalCores: totalCores,
|
|
TotalCoreHours: totalCoreHours,
|
|
TotalAccs: totalAccs,
|
|
TotalAccHours: totalAccHours,
|
|
})
|
|
}
|
|
}
|
|
}
|
|
|
|
log.Debugf("Timer JobsStatsGrouped %s", time.Since(start))
|
|
return stats, nil
|
|
}
|
|
|
|
func (r *JobRepository) JobsStats(
|
|
ctx context.Context,
|
|
filter []*model.JobFilter,
|
|
) ([]*model.JobsStatistics, error) {
|
|
start := time.Now()
|
|
query := r.buildStatsQuery(filter, "")
|
|
query, err := SecurityCheck(ctx, query)
|
|
if err != nil {
|
|
return nil, err
|
|
}
|
|
|
|
row := query.RunWith(r.DB).QueryRow()
|
|
stats := make([]*model.JobsStatistics, 0, 1)
|
|
|
|
var jobs, walltime, nodes, nodeHours, cores, coreHours, accs, accHours sql.NullInt64
|
|
if err := row.Scan(&jobs, &walltime, &nodes, &nodeHours, &cores, &coreHours, &accs, &accHours); err != nil {
|
|
log.Warn("Error while scanning rows")
|
|
return nil, err
|
|
}
|
|
|
|
if jobs.Valid {
|
|
var totalNodeHours, totalCoreHours, totalAccHours int
|
|
|
|
if nodeHours.Valid {
|
|
totalNodeHours = int(nodeHours.Int64)
|
|
}
|
|
if coreHours.Valid {
|
|
totalCoreHours = int(coreHours.Int64)
|
|
}
|
|
if accHours.Valid {
|
|
totalAccHours = int(accHours.Int64)
|
|
}
|
|
stats = append(stats,
|
|
&model.JobsStatistics{
|
|
TotalJobs: int(jobs.Int64),
|
|
TotalWalltime: int(walltime.Int64),
|
|
TotalNodeHours: totalNodeHours,
|
|
TotalCoreHours: totalCoreHours,
|
|
TotalAccHours: totalAccHours,
|
|
})
|
|
}
|
|
|
|
log.Debugf("Timer JobStats %s", time.Since(start))
|
|
return stats, nil
|
|
}
|
|
|
|
// FIXME: Make generic
|
|
func LoadJobStat(job *schema.JobMeta, metric string) float64 {
|
|
if stats, ok := job.Statistics[metric]; ok {
|
|
if metric == "mem_used" {
|
|
return stats.Max
|
|
} else {
|
|
return stats.Avg
|
|
}
|
|
}
|
|
|
|
return 0.0
|
|
}
|
|
|
|
func (r *JobRepository) JobCountGrouped(
|
|
ctx context.Context,
|
|
filter []*model.JobFilter,
|
|
groupBy *model.Aggregate,
|
|
) ([]*model.JobsStatistics, error) {
|
|
start := time.Now()
|
|
col := groupBy2column[*groupBy]
|
|
query := r.buildCountQuery(filter, "", col)
|
|
query, err := SecurityCheck(ctx, query)
|
|
if err != nil {
|
|
return nil, err
|
|
}
|
|
rows, err := query.RunWith(r.DB).Query()
|
|
if err != nil {
|
|
log.Warn("Error while querying DB for job statistics")
|
|
return nil, err
|
|
}
|
|
|
|
stats := make([]*model.JobsStatistics, 0, 100)
|
|
|
|
for rows.Next() {
|
|
var id sql.NullString
|
|
var cnt sql.NullInt64
|
|
if err := rows.Scan(&id, &cnt); err != nil {
|
|
log.Warn("Error while scanning rows")
|
|
return nil, err
|
|
}
|
|
if id.Valid {
|
|
stats = append(stats,
|
|
&model.JobsStatistics{
|
|
ID: id.String,
|
|
TotalJobs: int(cnt.Int64),
|
|
})
|
|
}
|
|
}
|
|
|
|
log.Debugf("Timer JobCountGrouped %s", time.Since(start))
|
|
return stats, nil
|
|
}
|
|
|
|
func (r *JobRepository) AddJobCountGrouped(
|
|
ctx context.Context,
|
|
filter []*model.JobFilter,
|
|
groupBy *model.Aggregate,
|
|
stats []*model.JobsStatistics,
|
|
kind string,
|
|
) ([]*model.JobsStatistics, error) {
|
|
start := time.Now()
|
|
col := groupBy2column[*groupBy]
|
|
query := r.buildCountQuery(filter, kind, col)
|
|
query, err := SecurityCheck(ctx, query)
|
|
if err != nil {
|
|
return nil, err
|
|
}
|
|
rows, err := query.RunWith(r.DB).Query()
|
|
if err != nil {
|
|
log.Warn("Error while querying DB for job statistics")
|
|
return nil, err
|
|
}
|
|
|
|
counts := make(map[string]int)
|
|
|
|
for rows.Next() {
|
|
var id sql.NullString
|
|
var cnt sql.NullInt64
|
|
if err := rows.Scan(&id, &cnt); err != nil {
|
|
log.Warn("Error while scanning rows")
|
|
return nil, err
|
|
}
|
|
if id.Valid {
|
|
counts[id.String] = int(cnt.Int64)
|
|
}
|
|
}
|
|
|
|
switch kind {
|
|
case "running":
|
|
for _, s := range stats {
|
|
s.RunningJobs = counts[s.ID]
|
|
}
|
|
case "short":
|
|
for _, s := range stats {
|
|
s.ShortJobs = counts[s.ID]
|
|
}
|
|
}
|
|
|
|
log.Debugf("Timer AddJobCountGrouped %s", time.Since(start))
|
|
return stats, nil
|
|
}
|
|
|
|
func (r *JobRepository) AddJobCount(
|
|
ctx context.Context,
|
|
filter []*model.JobFilter,
|
|
stats []*model.JobsStatistics,
|
|
kind string,
|
|
) ([]*model.JobsStatistics, error) {
|
|
start := time.Now()
|
|
query := r.buildCountQuery(filter, kind, "")
|
|
query, err := SecurityCheck(ctx, query)
|
|
if err != nil {
|
|
return nil, err
|
|
}
|
|
rows, err := query.RunWith(r.DB).Query()
|
|
if err != nil {
|
|
log.Warn("Error while querying DB for job statistics")
|
|
return nil, err
|
|
}
|
|
|
|
var count int
|
|
|
|
for rows.Next() {
|
|
var cnt sql.NullInt64
|
|
if err := rows.Scan(&cnt); err != nil {
|
|
log.Warn("Error while scanning rows")
|
|
return nil, err
|
|
}
|
|
|
|
count = int(cnt.Int64)
|
|
}
|
|
|
|
switch kind {
|
|
case "running":
|
|
for _, s := range stats {
|
|
s.RunningJobs = count
|
|
}
|
|
case "short":
|
|
for _, s := range stats {
|
|
s.ShortJobs = count
|
|
}
|
|
}
|
|
|
|
log.Debugf("Timer AddJobCount %s", time.Since(start))
|
|
return stats, nil
|
|
}
|
|
|
|
func (r *JobRepository) AddHistograms(
|
|
ctx context.Context,
|
|
filter []*model.JobFilter,
|
|
stat *model.JobsStatistics,
|
|
) (*model.JobsStatistics, error) {
|
|
start := time.Now()
|
|
|
|
castType := r.getCastType()
|
|
var err error
|
|
value := fmt.Sprintf(`CAST(ROUND((CASE WHEN job.job_state = "running" THEN %d - job.start_time ELSE job.duration END) / 3600) as %s) as value`, time.Now().Unix(), castType)
|
|
stat.HistDuration, err = r.jobsStatisticsHistogram(ctx, value, filter)
|
|
if err != nil {
|
|
log.Warn("Error while loading job statistics histogram: running jobs")
|
|
return nil, err
|
|
}
|
|
|
|
stat.HistNumNodes, err = r.jobsStatisticsHistogram(ctx, "job.num_nodes as value", filter)
|
|
if err != nil {
|
|
log.Warn("Error while loading job statistics histogram: num nodes")
|
|
return nil, err
|
|
}
|
|
|
|
stat.HistNumCores, err = r.jobsStatisticsHistogram(ctx, "job.num_hwthreads as value", filter)
|
|
if err != nil {
|
|
log.Warn("Error while loading job statistics histogram: num hwthreads")
|
|
return nil, err
|
|
}
|
|
|
|
stat.HistNumAccs, err = r.jobsStatisticsHistogram(ctx, "job.num_acc as value", filter)
|
|
if err != nil {
|
|
log.Warn("Error while loading job statistics histogram: num acc")
|
|
return nil, err
|
|
}
|
|
|
|
log.Debugf("Timer AddHistograms %s", time.Since(start))
|
|
return stat, nil
|
|
}
|
|
|
|
// Requires thresholds for metric from config for cluster? Of all clusters and use largest? split to 10 + 1 for artifacts?
|
|
func (r *JobRepository) AddMetricHistograms(
|
|
ctx context.Context,
|
|
filter []*model.JobFilter,
|
|
metrics []string,
|
|
stat *model.JobsStatistics,
|
|
) (*model.JobsStatistics, error) {
|
|
start := time.Now()
|
|
|
|
// Running Jobs Only: First query jobdata from sqlite, then query data and make bins
|
|
for _, f := range filter {
|
|
if f.State != nil {
|
|
if len(f.State) == 1 && f.State[0] == "running" {
|
|
stat.HistMetrics = r.runningJobsMetricStatisticsHistogram(ctx, metrics, filter)
|
|
log.Debugf("Timer AddMetricHistograms %s", time.Since(start))
|
|
return stat, nil
|
|
}
|
|
}
|
|
}
|
|
|
|
// All other cases: Query and make bins in sqlite directly
|
|
for _, m := range metrics {
|
|
metricHisto, err := r.jobsMetricStatisticsHistogram(ctx, m, filter)
|
|
if err != nil {
|
|
log.Warnf("Error while loading job metric statistics histogram: %s", m)
|
|
continue
|
|
}
|
|
stat.HistMetrics = append(stat.HistMetrics, metricHisto)
|
|
}
|
|
|
|
log.Debugf("Timer AddMetricHistograms %s", time.Since(start))
|
|
return stat, nil
|
|
}
|
|
|
|
// `value` must be the column grouped by, but renamed to "value"
|
|
func (r *JobRepository) jobsStatisticsHistogram(
|
|
ctx context.Context,
|
|
value string,
|
|
filters []*model.JobFilter,
|
|
) ([]*model.HistoPoint, error) {
|
|
start := time.Now()
|
|
query, qerr := SecurityCheck(ctx,
|
|
sq.Select(value, "COUNT(job.id) AS count").From("job"))
|
|
|
|
if qerr != nil {
|
|
return nil, qerr
|
|
}
|
|
|
|
for _, f := range filters {
|
|
query = BuildWhereClause(f, query)
|
|
}
|
|
|
|
rows, err := query.GroupBy("value").RunWith(r.DB).Query()
|
|
if err != nil {
|
|
log.Error("Error while running query")
|
|
return nil, err
|
|
}
|
|
|
|
points := make([]*model.HistoPoint, 0)
|
|
for rows.Next() {
|
|
point := model.HistoPoint{}
|
|
if err := rows.Scan(&point.Value, &point.Count); err != nil {
|
|
log.Warn("Error while scanning rows")
|
|
return nil, err
|
|
}
|
|
|
|
points = append(points, &point)
|
|
}
|
|
log.Debugf("Timer jobsStatisticsHistogram %s", time.Since(start))
|
|
return points, nil
|
|
}
|
|
|
|
func (r *JobRepository) jobsMetricStatisticsHistogram(
|
|
ctx context.Context,
|
|
metric string,
|
|
filters []*model.JobFilter,
|
|
) (*model.MetricHistoPoints, error) {
|
|
// Get specific Peak or largest Peak
|
|
var metricConfig *schema.MetricConfig
|
|
var peak float64 = 0.0
|
|
var unit string = ""
|
|
var footprintStat string = ""
|
|
|
|
for _, f := range filters {
|
|
if f.Cluster != nil {
|
|
metricConfig = archive.GetMetricConfig(*f.Cluster.Eq, metric)
|
|
peak = metricConfig.Peak
|
|
unit = metricConfig.Unit.Prefix + metricConfig.Unit.Base
|
|
footprintStat = metricConfig.Footprint
|
|
log.Debugf("Cluster %s filter found with peak %f for %s", *f.Cluster.Eq, peak, metric)
|
|
}
|
|
}
|
|
|
|
if peak == 0.0 {
|
|
for _, c := range archive.Clusters {
|
|
for _, m := range c.MetricConfig {
|
|
if m.Name == metric {
|
|
if m.Peak > peak {
|
|
peak = m.Peak
|
|
}
|
|
if unit == "" {
|
|
unit = m.Unit.Prefix + m.Unit.Base
|
|
}
|
|
if footprintStat == "" {
|
|
footprintStat = m.Footprint
|
|
}
|
|
}
|
|
}
|
|
}
|
|
}
|
|
|
|
// log.Debugf("Metric %s, Peak %f, Unit %s, Aggregation %s", metric, peak, unit, aggreg)
|
|
// Make bins, see https://jereze.com/code/sql-histogram/
|
|
|
|
start := time.Now()
|
|
jm := fmt.Sprintf(`json_extract(footprint, "$.%s")`, (metric + "_" + footprintStat))
|
|
|
|
crossJoinQuery := sq.Select(
|
|
fmt.Sprintf(`max(%s) as max`, jm),
|
|
fmt.Sprintf(`min(%s) as min`, jm),
|
|
).From("job").Where(
|
|
"JSON_VALID(footprint)",
|
|
).Where(
|
|
fmt.Sprintf(`%s is not null`, jm),
|
|
).Where(
|
|
fmt.Sprintf(`%s <= %f`, jm, peak),
|
|
)
|
|
|
|
crossJoinQuery, cjqerr := SecurityCheck(ctx, crossJoinQuery)
|
|
|
|
if cjqerr != nil {
|
|
return nil, cjqerr
|
|
}
|
|
|
|
for _, f := range filters {
|
|
crossJoinQuery = BuildWhereClause(f, crossJoinQuery)
|
|
}
|
|
|
|
crossJoinQuerySql, crossJoinQueryArgs, sqlerr := crossJoinQuery.ToSql()
|
|
if sqlerr != nil {
|
|
return nil, sqlerr
|
|
}
|
|
|
|
bins := 10
|
|
binQuery := fmt.Sprintf(`CAST( (case when %s = value.max
|
|
then value.max*0.999999999 else %s end - value.min) / (value.max -
|
|
value.min) * %d as INTEGER )`, jm, jm, bins)
|
|
|
|
mainQuery := sq.Select(
|
|
fmt.Sprintf(`%s + 1 as bin`, binQuery),
|
|
fmt.Sprintf(`count(%s) as count`, jm),
|
|
fmt.Sprintf(`CAST(((value.max / %d) * (%s )) as INTEGER ) as min`, bins, binQuery),
|
|
fmt.Sprintf(`CAST(((value.max / %d) * (%s + 1 )) as INTEGER ) as max`, bins, binQuery),
|
|
).From("job").CrossJoin(
|
|
fmt.Sprintf(`(%s) as value`, crossJoinQuerySql), crossJoinQueryArgs...,
|
|
).Where(fmt.Sprintf(`%s is not null and %s <= %f`, jm, jm, peak))
|
|
|
|
mainQuery, qerr := SecurityCheck(ctx, mainQuery)
|
|
|
|
if qerr != nil {
|
|
return nil, qerr
|
|
}
|
|
|
|
for _, f := range filters {
|
|
mainQuery = BuildWhereClause(f, mainQuery)
|
|
}
|
|
|
|
// Finalize query with Grouping and Ordering
|
|
mainQuery = mainQuery.GroupBy("bin").OrderBy("bin")
|
|
|
|
rows, err := mainQuery.RunWith(r.DB).Query()
|
|
if err != nil {
|
|
log.Errorf("Error while running mainQuery: %s", err)
|
|
return nil, err
|
|
}
|
|
|
|
points := make([]*model.MetricHistoPoint, 0)
|
|
for rows.Next() {
|
|
point := model.MetricHistoPoint{}
|
|
if err := rows.Scan(&point.Bin, &point.Count, &point.Min, &point.Max); err != nil {
|
|
log.Warnf("Error while scanning rows for %s", jm)
|
|
return nil, err // Totally bricks cc-backend if returned and if all metrics requested?
|
|
}
|
|
|
|
points = append(points, &point)
|
|
}
|
|
|
|
result := model.MetricHistoPoints{Metric: metric, Unit: unit, Stat: &footprintStat, Data: points}
|
|
|
|
log.Debugf("Timer jobsStatisticsHistogram %s", time.Since(start))
|
|
return &result, nil
|
|
}
|
|
|
|
func (r *JobRepository) runningJobsMetricStatisticsHistogram(
|
|
ctx context.Context,
|
|
metrics []string,
|
|
filters []*model.JobFilter,
|
|
) []*model.MetricHistoPoints {
|
|
// Get Jobs
|
|
jobs, err := r.QueryJobs(ctx, filters, &model.PageRequest{Page: 1, ItemsPerPage: 500 + 1}, nil)
|
|
if err != nil {
|
|
log.Errorf("Error while querying jobs for footprint: %s", err)
|
|
return nil
|
|
}
|
|
if len(jobs) > 500 {
|
|
log.Errorf("too many jobs matched (max: %d)", 500)
|
|
return nil
|
|
}
|
|
|
|
// Get AVGs from metric repo
|
|
avgs := make([][]schema.Float, len(metrics))
|
|
for i := range avgs {
|
|
avgs[i] = make([]schema.Float, 0, len(jobs))
|
|
}
|
|
|
|
for _, job := range jobs {
|
|
if job.MonitoringStatus == schema.MonitoringStatusDisabled || job.MonitoringStatus == schema.MonitoringStatusArchivingFailed {
|
|
continue
|
|
}
|
|
|
|
if err := metricdata.LoadAverages(job, metrics, avgs, ctx); err != nil {
|
|
log.Errorf("Error while loading averages for histogram: %s", err)
|
|
return nil
|
|
}
|
|
}
|
|
|
|
// Iterate metrics to fill endresult
|
|
data := make([]*model.MetricHistoPoints, 0)
|
|
for idx, metric := range metrics {
|
|
// Get specific Peak or largest Peak
|
|
var metricConfig *schema.MetricConfig
|
|
var peak float64 = 0.0
|
|
var unit string = ""
|
|
|
|
for _, f := range filters {
|
|
if f.Cluster != nil {
|
|
metricConfig = archive.GetMetricConfig(*f.Cluster.Eq, metric)
|
|
peak = metricConfig.Peak
|
|
unit = metricConfig.Unit.Prefix + metricConfig.Unit.Base
|
|
log.Debugf("Cluster %s filter found with peak %f for %s", *f.Cluster.Eq, peak, metric)
|
|
}
|
|
}
|
|
|
|
if peak == 0.0 {
|
|
for _, c := range archive.Clusters {
|
|
for _, m := range c.MetricConfig {
|
|
if m.Name == metric {
|
|
if m.Peak > peak {
|
|
peak = m.Peak
|
|
}
|
|
if unit == "" {
|
|
unit = m.Unit.Prefix + m.Unit.Base
|
|
}
|
|
}
|
|
}
|
|
}
|
|
}
|
|
|
|
// Make and fill bins
|
|
bins := 10.0
|
|
peakBin := peak / bins
|
|
|
|
points := make([]*model.MetricHistoPoint, 0)
|
|
for b := 0; b < 10; b++ {
|
|
count := 0
|
|
bindex := b + 1
|
|
bmin := math.Round(peakBin * float64(b))
|
|
bmax := math.Round(peakBin * (float64(b) + 1.0))
|
|
|
|
// Iterate AVG values for indexed metric and count for bins
|
|
for _, val := range avgs[idx] {
|
|
if float64(val) >= bmin && float64(val) < bmax {
|
|
count += 1
|
|
}
|
|
}
|
|
|
|
bminint := int(bmin)
|
|
bmaxint := int(bmax)
|
|
|
|
// Append Bin to Metric Result Array
|
|
point := model.MetricHistoPoint{Bin: &bindex, Count: count, Min: &bminint, Max: &bmaxint}
|
|
points = append(points, &point)
|
|
}
|
|
|
|
// Append Metric Result Array to final results array
|
|
result := model.MetricHistoPoints{Metric: metric, Unit: unit, Data: points}
|
|
data = append(data, &result)
|
|
}
|
|
|
|
return data
|
|
}
|