mirror of
https://github.com/ClusterCockpit/cc-backend
synced 2024-12-25 04:49:05 +01:00
Refactor directory structure
This commit is contained in:
parent
45359cca9d
commit
81819db436
2
.gitmodules
vendored
2
.gitmodules
vendored
@ -1,5 +1,5 @@
|
|||||||
[submodule "frontend"]
|
[submodule "frontend"]
|
||||||
path = frontend
|
path = web/frontend
|
||||||
url = git@github.com:ClusterCockpit/cc-frontend.git
|
url = git@github.com:ClusterCockpit/cc-frontend.git
|
||||||
branch = main
|
branch = main
|
||||||
update = merge
|
update = merge
|
||||||
|
@ -22,26 +22,25 @@ import (
|
|||||||
|
|
||||||
"github.com/99designs/gqlgen/graphql/handler"
|
"github.com/99designs/gqlgen/graphql/handler"
|
||||||
"github.com/99designs/gqlgen/graphql/playground"
|
"github.com/99designs/gqlgen/graphql/playground"
|
||||||
"github.com/ClusterCockpit/cc-backend/api"
|
"github.com/ClusterCockpit/cc-backend/internal/api"
|
||||||
"github.com/ClusterCockpit/cc-backend/auth"
|
"github.com/ClusterCockpit/cc-backend/internal/auth"
|
||||||
"github.com/ClusterCockpit/cc-backend/config"
|
"github.com/ClusterCockpit/cc-backend/internal/config"
|
||||||
"github.com/ClusterCockpit/cc-backend/graph"
|
"github.com/ClusterCockpit/cc-backend/internal/graph"
|
||||||
"github.com/ClusterCockpit/cc-backend/graph/generated"
|
"github.com/ClusterCockpit/cc-backend/internal/graph/generated"
|
||||||
"github.com/ClusterCockpit/cc-backend/log"
|
"github.com/ClusterCockpit/cc-backend/internal/metricdata"
|
||||||
"github.com/ClusterCockpit/cc-backend/metricdata"
|
"github.com/ClusterCockpit/cc-backend/internal/repository"
|
||||||
"github.com/ClusterCockpit/cc-backend/repository"
|
"github.com/ClusterCockpit/cc-backend/internal/routerConfig"
|
||||||
"github.com/ClusterCockpit/cc-backend/templates"
|
"github.com/ClusterCockpit/cc-backend/internal/runtimeEnv"
|
||||||
|
"github.com/ClusterCockpit/cc-backend/internal/templates"
|
||||||
|
"github.com/ClusterCockpit/cc-backend/pkg/log"
|
||||||
"github.com/google/gops/agent"
|
"github.com/google/gops/agent"
|
||||||
"github.com/gorilla/handlers"
|
"github.com/gorilla/handlers"
|
||||||
"github.com/gorilla/mux"
|
"github.com/gorilla/mux"
|
||||||
"github.com/jmoiron/sqlx"
|
|
||||||
|
|
||||||
_ "github.com/go-sql-driver/mysql"
|
_ "github.com/go-sql-driver/mysql"
|
||||||
_ "github.com/mattn/go-sqlite3"
|
_ "github.com/mattn/go-sqlite3"
|
||||||
)
|
)
|
||||||
|
|
||||||
var jobRepo *repository.JobRepository
|
|
||||||
|
|
||||||
// Format of the configurartion (file). See below for the defaults.
|
// Format of the configurartion (file). See below for the defaults.
|
||||||
type ProgramConfig struct {
|
type ProgramConfig struct {
|
||||||
// Address where the http (or https) server will listen on (for example: 'localhost:80').
|
// Address where the http (or https) server will listen on (for example: 'localhost:80').
|
||||||
@ -152,7 +151,7 @@ func main() {
|
|||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
if err := loadEnv("./.env"); err != nil && !os.IsNotExist(err) {
|
if err := runtimeEnv.LoadEnv("./.env"); err != nil && !os.IsNotExist(err) {
|
||||||
log.Fatalf("parsing './.env' file failed: %s", err.Error())
|
log.Fatalf("parsing './.env' file failed: %s", err.Error())
|
||||||
}
|
}
|
||||||
|
|
||||||
@ -178,28 +177,8 @@ func main() {
|
|||||||
}
|
}
|
||||||
|
|
||||||
var err error
|
var err error
|
||||||
var db *sqlx.DB
|
repository.Connect(programConfig.DBDriver, programConfig.DB)
|
||||||
if programConfig.DBDriver == "sqlite3" {
|
db := repository.GetConnection()
|
||||||
db, err = sqlx.Open("sqlite3", fmt.Sprintf("%s?_foreign_keys=on", programConfig.DB))
|
|
||||||
if err != nil {
|
|
||||||
log.Fatal(err)
|
|
||||||
}
|
|
||||||
|
|
||||||
// sqlite does not multithread. Having more than one connection open would just mean
|
|
||||||
// waiting for locks.
|
|
||||||
db.SetMaxOpenConns(1)
|
|
||||||
} else if programConfig.DBDriver == "mysql" {
|
|
||||||
db, err = sqlx.Open("mysql", fmt.Sprintf("%s?multiStatements=true", programConfig.DB))
|
|
||||||
if err != nil {
|
|
||||||
log.Fatal(err)
|
|
||||||
}
|
|
||||||
|
|
||||||
db.SetConnMaxLifetime(time.Minute * 3)
|
|
||||||
db.SetMaxOpenConns(10)
|
|
||||||
db.SetMaxIdleConns(10)
|
|
||||||
} else {
|
|
||||||
log.Fatalf("unsupported database driver: %s", programConfig.DBDriver)
|
|
||||||
}
|
|
||||||
|
|
||||||
// Initialize sub-modules and handle all command line flags.
|
// Initialize sub-modules and handle all command line flags.
|
||||||
// The order here is important! For example, the metricdata package
|
// The order here is important! For example, the metricdata package
|
||||||
@ -215,7 +194,7 @@ func main() {
|
|||||||
authentication.JwtMaxAge = d
|
authentication.JwtMaxAge = d
|
||||||
}
|
}
|
||||||
|
|
||||||
if err := authentication.Init(db, programConfig.LdapConfig); err != nil {
|
if err := authentication.Init(db.DB, programConfig.LdapConfig); err != nil {
|
||||||
log.Fatal(err)
|
log.Fatal(err)
|
||||||
}
|
}
|
||||||
|
|
||||||
@ -257,7 +236,7 @@ func main() {
|
|||||||
log.Fatal("arguments --add-user and --del-user can only be used if authentication is enabled")
|
log.Fatal("arguments --add-user and --del-user can only be used if authentication is enabled")
|
||||||
}
|
}
|
||||||
|
|
||||||
if err := config.Init(db, !programConfig.DisableAuthentication, programConfig.UiDefaults, programConfig.JobArchive); err != nil {
|
if err := config.Init(db.DB, !programConfig.DisableAuthentication, programConfig.UiDefaults, programConfig.JobArchive); err != nil {
|
||||||
log.Fatal(err)
|
log.Fatal(err)
|
||||||
}
|
}
|
||||||
|
|
||||||
@ -266,15 +245,12 @@ func main() {
|
|||||||
}
|
}
|
||||||
|
|
||||||
if flagReinitDB {
|
if flagReinitDB {
|
||||||
if err := repository.InitDB(db, programConfig.JobArchive); err != nil {
|
if err := repository.InitDB(db.DB, programConfig.JobArchive); err != nil {
|
||||||
log.Fatal(err)
|
log.Fatal(err)
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
jobRepo = &repository.JobRepository{DB: db}
|
jobRepo := repository.GetRepository()
|
||||||
if err := jobRepo.Init(); err != nil {
|
|
||||||
log.Fatal(err)
|
|
||||||
}
|
|
||||||
|
|
||||||
if flagImportJob != "" {
|
if flagImportJob != "" {
|
||||||
if err := jobRepo.HandleImportFlag(flagImportJob); err != nil {
|
if err := jobRepo.HandleImportFlag(flagImportJob); err != nil {
|
||||||
@ -288,7 +264,7 @@ func main() {
|
|||||||
|
|
||||||
// Setup the http.Handler/Router used by the server
|
// Setup the http.Handler/Router used by the server
|
||||||
|
|
||||||
resolver := &graph.Resolver{DB: db, Repo: jobRepo}
|
resolver := &graph.Resolver{DB: db.DB, Repo: jobRepo}
|
||||||
graphQLEndpoint := handler.NewDefaultServer(generated.NewExecutableSchema(generated.Config{Resolvers: resolver}))
|
graphQLEndpoint := handler.NewDefaultServer(generated.NewExecutableSchema(generated.Config{Resolvers: resolver}))
|
||||||
if os.Getenv("DEBUG") != "1" {
|
if os.Getenv("DEBUG") != "1" {
|
||||||
// Having this handler means that a error message is returned via GraphQL instead of the connection simply beeing closed.
|
// Having this handler means that a error message is returned via GraphQL instead of the connection simply beeing closed.
|
||||||
@ -394,7 +370,7 @@ func main() {
|
|||||||
})
|
})
|
||||||
|
|
||||||
// Mount all /monitoring/... and /api/... routes.
|
// Mount all /monitoring/... and /api/... routes.
|
||||||
setupRoutes(secured, routes)
|
routerConfig.SetupRoutes(secured)
|
||||||
api.MountRoutes(secured)
|
api.MountRoutes(secured)
|
||||||
|
|
||||||
r.PathPrefix("/").Handler(http.FileServer(http.Dir(programConfig.StaticFiles)))
|
r.PathPrefix("/").Handler(http.FileServer(http.Dir(programConfig.StaticFiles)))
|
||||||
@ -461,7 +437,7 @@ func main() {
|
|||||||
// Because this program will want to bind to a privileged port (like 80), the listener must
|
// Because this program will want to bind to a privileged port (like 80), the listener must
|
||||||
// be established first, then the user can be changed, and after that,
|
// be established first, then the user can be changed, and after that,
|
||||||
// the actuall http server can be started.
|
// the actuall http server can be started.
|
||||||
if err := dropPrivileges(); err != nil {
|
if err := runtimeEnv.DropPrivileges(programConfig.Group, programConfig.User); err != nil {
|
||||||
log.Fatalf("error while changing user: %s", err.Error())
|
log.Fatalf("error while changing user: %s", err.Error())
|
||||||
}
|
}
|
||||||
|
|
||||||
@ -479,7 +455,7 @@ func main() {
|
|||||||
go func() {
|
go func() {
|
||||||
defer wg.Done()
|
defer wg.Done()
|
||||||
<-sigs
|
<-sigs
|
||||||
systemdNotifiy(false, "shutting down")
|
runtimeEnv.SystemdNotifiy(false, "shutting down")
|
||||||
|
|
||||||
// First shut down the server gracefully (waiting for all ongoing requests)
|
// First shut down the server gracefully (waiting for all ongoing requests)
|
||||||
server.Shutdown(context.Background())
|
server.Shutdown(context.Background())
|
||||||
@ -503,7 +479,7 @@ func main() {
|
|||||||
if os.Getenv("GOGC") == "" {
|
if os.Getenv("GOGC") == "" {
|
||||||
debug.SetGCPercent(25)
|
debug.SetGCPercent(25)
|
||||||
}
|
}
|
||||||
systemdNotifiy(true, "running")
|
runtimeEnv.SystemdNotifiy(true, "running")
|
||||||
wg.Wait()
|
wg.Wait()
|
||||||
log.Print("Gracefull shutdown completed!")
|
log.Print("Gracefull shutdown completed!")
|
||||||
}
|
}
|
2
frontend
2
frontend
@ -1 +1 @@
|
|||||||
Subproject commit 94ef11aa9fc3c194f1df497e3e06c60a7125883d
|
Subproject commit 4d698c519a56dd411dde7001beb0b73eb60157b9
|
@ -1,10 +1,10 @@
|
|||||||
# Where are all the schema files located? globs are supported eg src/**/*.graphqls
|
# Where are all the schema files located? globs are supported eg src/**/*.graphqls
|
||||||
schema:
|
schema:
|
||||||
- graph/*.graphqls
|
- api/*.graphqls
|
||||||
|
|
||||||
# Where should the generated server code go?
|
# Where should the generated server code go?
|
||||||
exec:
|
exec:
|
||||||
filename: graph/generated/generated.go
|
filename: internal/graph/generated/generated.go
|
||||||
package: generated
|
package: generated
|
||||||
|
|
||||||
# Uncomment to enable federation
|
# Uncomment to enable federation
|
||||||
@ -14,7 +14,7 @@ exec:
|
|||||||
|
|
||||||
# Where should any generated models go?
|
# Where should any generated models go?
|
||||||
model:
|
model:
|
||||||
filename: graph/model/models_gen.go
|
filename: internal/graph/model/models_gen.go
|
||||||
package: model
|
package: model
|
||||||
|
|
||||||
# Where should the resolver implementations go?
|
# Where should the resolver implementations go?
|
||||||
@ -75,5 +75,3 @@ models:
|
|||||||
Series: { model: "github.com/ClusterCockpit/cc-backend/schema.Series" }
|
Series: { model: "github.com/ClusterCockpit/cc-backend/schema.Series" }
|
||||||
MetricStatistics: { model: "github.com/ClusterCockpit/cc-backend/schema.MetricStatistics" }
|
MetricStatistics: { model: "github.com/ClusterCockpit/cc-backend/schema.MetricStatistics" }
|
||||||
StatsSeries: { model: "github.com/ClusterCockpit/cc-backend/schema.StatsSeries" }
|
StatsSeries: { model: "github.com/ClusterCockpit/cc-backend/schema.StatsSeries" }
|
||||||
|
|
||||||
|
|
||||||
|
@ -16,14 +16,14 @@ import (
|
|||||||
"sync"
|
"sync"
|
||||||
"time"
|
"time"
|
||||||
|
|
||||||
"github.com/ClusterCockpit/cc-backend/auth"
|
"github.com/ClusterCockpit/cc-backend/internal/auth"
|
||||||
"github.com/ClusterCockpit/cc-backend/config"
|
"github.com/ClusterCockpit/cc-backend/internal/config"
|
||||||
"github.com/ClusterCockpit/cc-backend/graph"
|
"github.com/ClusterCockpit/cc-backend/internal/graph"
|
||||||
"github.com/ClusterCockpit/cc-backend/graph/model"
|
"github.com/ClusterCockpit/cc-backend/internal/graph/model"
|
||||||
"github.com/ClusterCockpit/cc-backend/log"
|
"github.com/ClusterCockpit/cc-backend/internal/metricdata"
|
||||||
"github.com/ClusterCockpit/cc-backend/metricdata"
|
"github.com/ClusterCockpit/cc-backend/internal/repository"
|
||||||
"github.com/ClusterCockpit/cc-backend/repository"
|
"github.com/ClusterCockpit/cc-backend/pkg/log"
|
||||||
"github.com/ClusterCockpit/cc-backend/schema"
|
"github.com/ClusterCockpit/cc-backend/pkg/schema"
|
||||||
"github.com/gorilla/mux"
|
"github.com/gorilla/mux"
|
||||||
)
|
)
|
||||||
|
|
@ -14,8 +14,8 @@ import (
|
|||||||
"strings"
|
"strings"
|
||||||
"time"
|
"time"
|
||||||
|
|
||||||
"github.com/ClusterCockpit/cc-backend/graph/model"
|
"github.com/ClusterCockpit/cc-backend/internal/graph/model"
|
||||||
"github.com/ClusterCockpit/cc-backend/log"
|
"github.com/ClusterCockpit/cc-backend/pkg/log"
|
||||||
sq "github.com/Masterminds/squirrel"
|
sq "github.com/Masterminds/squirrel"
|
||||||
"github.com/golang-jwt/jwt/v4"
|
"github.com/golang-jwt/jwt/v4"
|
||||||
"github.com/gorilla/sessions"
|
"github.com/gorilla/sessions"
|
@ -6,8 +6,7 @@ import (
|
|||||||
"strings"
|
"strings"
|
||||||
"time"
|
"time"
|
||||||
|
|
||||||
"github.com/ClusterCockpit/cc-backend/log"
|
"github.com/ClusterCockpit/cc-backend/pkg/log"
|
||||||
|
|
||||||
"github.com/go-ldap/ldap/v3"
|
"github.com/go-ldap/ldap/v3"
|
||||||
)
|
)
|
||||||
|
|
@ -11,9 +11,9 @@ import (
|
|||||||
"sync"
|
"sync"
|
||||||
"time"
|
"time"
|
||||||
|
|
||||||
"github.com/ClusterCockpit/cc-backend/auth"
|
"github.com/ClusterCockpit/cc-backend/internal/auth"
|
||||||
"github.com/ClusterCockpit/cc-backend/graph/model"
|
"github.com/ClusterCockpit/cc-backend/internal/graph/model"
|
||||||
"github.com/ClusterCockpit/cc-backend/schema"
|
"github.com/ClusterCockpit/cc-backend/pkg/schema"
|
||||||
"github.com/iamlouk/lrucache"
|
"github.com/iamlouk/lrucache"
|
||||||
"github.com/jmoiron/sqlx"
|
"github.com/jmoiron/sqlx"
|
||||||
)
|
)
|
@ -5,7 +5,7 @@ import (
|
|||||||
"strconv"
|
"strconv"
|
||||||
"strings"
|
"strings"
|
||||||
|
|
||||||
"github.com/ClusterCockpit/cc-backend/log"
|
"github.com/ClusterCockpit/cc-backend/pkg/log"
|
||||||
)
|
)
|
||||||
|
|
||||||
type NodeList [][]interface {
|
type NodeList [][]interface {
|
@ -13,8 +13,8 @@ import (
|
|||||||
|
|
||||||
"github.com/99designs/gqlgen/graphql"
|
"github.com/99designs/gqlgen/graphql"
|
||||||
"github.com/99designs/gqlgen/graphql/introspection"
|
"github.com/99designs/gqlgen/graphql/introspection"
|
||||||
"github.com/ClusterCockpit/cc-backend/graph/model"
|
"github.com/ClusterCockpit/cc-backend/internal/graph/model"
|
||||||
"github.com/ClusterCockpit/cc-backend/schema"
|
"github.com/ClusterCockpit/cc-backend/pkg/schema"
|
||||||
gqlparser "github.com/vektah/gqlparser/v2"
|
gqlparser "github.com/vektah/gqlparser/v2"
|
||||||
"github.com/vektah/gqlparser/v2/ast"
|
"github.com/vektah/gqlparser/v2/ast"
|
||||||
)
|
)
|
@ -8,7 +8,7 @@ import (
|
|||||||
"strconv"
|
"strconv"
|
||||||
"time"
|
"time"
|
||||||
|
|
||||||
"github.com/ClusterCockpit/cc-backend/schema"
|
"github.com/ClusterCockpit/cc-backend/pkg/schema"
|
||||||
)
|
)
|
||||||
|
|
||||||
type Accelerator struct {
|
type Accelerator struct {
|
@ -1,7 +1,7 @@
|
|||||||
package graph
|
package graph
|
||||||
|
|
||||||
import (
|
import (
|
||||||
"github.com/ClusterCockpit/cc-backend/repository"
|
"github.com/ClusterCockpit/cc-backend/internal/repository"
|
||||||
"github.com/jmoiron/sqlx"
|
"github.com/jmoiron/sqlx"
|
||||||
)
|
)
|
||||||
|
|
275
internal/graph/schema.graphqls
Normal file
275
internal/graph/schema.graphqls
Normal file
@ -0,0 +1,275 @@
|
|||||||
|
scalar Time
|
||||||
|
scalar Any
|
||||||
|
|
||||||
|
scalar NullableFloat
|
||||||
|
scalar MetricScope
|
||||||
|
scalar JobState
|
||||||
|
|
||||||
|
type Job {
|
||||||
|
id: ID!
|
||||||
|
jobId: Int!
|
||||||
|
user: String!
|
||||||
|
project: String!
|
||||||
|
cluster: String!
|
||||||
|
subCluster: String!
|
||||||
|
startTime: Time!
|
||||||
|
duration: Int!
|
||||||
|
walltime: Int!
|
||||||
|
numNodes: Int!
|
||||||
|
numHWThreads: Int!
|
||||||
|
numAcc: Int!
|
||||||
|
SMT: Int!
|
||||||
|
exclusive: Int!
|
||||||
|
partition: String!
|
||||||
|
arrayJobId: Int!
|
||||||
|
monitoringStatus: Int!
|
||||||
|
state: JobState!
|
||||||
|
tags: [Tag!]!
|
||||||
|
resources: [Resource!]!
|
||||||
|
|
||||||
|
metaData: Any
|
||||||
|
userData: User
|
||||||
|
}
|
||||||
|
|
||||||
|
type Cluster {
|
||||||
|
name: String!
|
||||||
|
partitions: [String!]! # Slurm partitions
|
||||||
|
metricConfig: [MetricConfig!]!
|
||||||
|
filterRanges: FilterRanges!
|
||||||
|
subClusters: [SubCluster!]! # Hardware partitions/subclusters
|
||||||
|
}
|
||||||
|
|
||||||
|
type SubCluster {
|
||||||
|
name: String!
|
||||||
|
nodes: String!
|
||||||
|
numberOfNodes: Int!
|
||||||
|
processorType: String!
|
||||||
|
socketsPerNode: Int!
|
||||||
|
coresPerSocket: Int!
|
||||||
|
threadsPerCore: Int!
|
||||||
|
flopRateScalar: Int!
|
||||||
|
flopRateSimd: Int!
|
||||||
|
memoryBandwidth: Int!
|
||||||
|
topology: Topology!
|
||||||
|
}
|
||||||
|
|
||||||
|
type Topology {
|
||||||
|
node: [Int!]
|
||||||
|
socket: [[Int!]!]
|
||||||
|
memoryDomain: [[Int!]!]
|
||||||
|
die: [[Int!]!]
|
||||||
|
core: [[Int!]!]
|
||||||
|
accelerators: [Accelerator!]
|
||||||
|
}
|
||||||
|
|
||||||
|
type Accelerator {
|
||||||
|
id: String!
|
||||||
|
type: String!
|
||||||
|
model: String!
|
||||||
|
}
|
||||||
|
|
||||||
|
type SubClusterConfig {
|
||||||
|
name: String!
|
||||||
|
peak: Float!
|
||||||
|
normal: Float!
|
||||||
|
caution: Float!
|
||||||
|
alert: Float!
|
||||||
|
}
|
||||||
|
|
||||||
|
type MetricConfig {
|
||||||
|
name: String!
|
||||||
|
unit: String!
|
||||||
|
scope: MetricScope!
|
||||||
|
aggregation: String
|
||||||
|
timestep: Int!
|
||||||
|
peak: Float
|
||||||
|
normal: Float
|
||||||
|
caution: Float
|
||||||
|
alert: Float
|
||||||
|
subClusters: [SubClusterConfig]
|
||||||
|
}
|
||||||
|
|
||||||
|
type Tag {
|
||||||
|
id: ID!
|
||||||
|
type: String!
|
||||||
|
name: String!
|
||||||
|
}
|
||||||
|
|
||||||
|
type Resource {
|
||||||
|
hostname: String!
|
||||||
|
hwthreads: [Int!]
|
||||||
|
accelerators: [String!]
|
||||||
|
configuration: String
|
||||||
|
}
|
||||||
|
|
||||||
|
type JobMetricWithName {
|
||||||
|
name: String!
|
||||||
|
metric: JobMetric!
|
||||||
|
}
|
||||||
|
|
||||||
|
type JobMetric {
|
||||||
|
unit: String!
|
||||||
|
scope: MetricScope!
|
||||||
|
timestep: Int!
|
||||||
|
series: [Series!]
|
||||||
|
statisticsSeries: StatsSeries
|
||||||
|
}
|
||||||
|
|
||||||
|
type Series {
|
||||||
|
hostname: String!
|
||||||
|
id: Int
|
||||||
|
statistics: MetricStatistics
|
||||||
|
data: [NullableFloat!]!
|
||||||
|
}
|
||||||
|
|
||||||
|
type MetricStatistics {
|
||||||
|
avg: Float!
|
||||||
|
min: Float!
|
||||||
|
max: Float!
|
||||||
|
}
|
||||||
|
|
||||||
|
type StatsSeries {
|
||||||
|
mean: [NullableFloat!]!
|
||||||
|
min: [NullableFloat!]!
|
||||||
|
max: [NullableFloat!]!
|
||||||
|
}
|
||||||
|
|
||||||
|
type MetricFootprints {
|
||||||
|
metric: String!
|
||||||
|
data: [NullableFloat!]!
|
||||||
|
}
|
||||||
|
|
||||||
|
type Footprints {
|
||||||
|
nodehours: [NullableFloat!]!
|
||||||
|
metrics: [MetricFootprints!]!
|
||||||
|
}
|
||||||
|
|
||||||
|
enum Aggregate { USER, PROJECT, CLUSTER }
|
||||||
|
enum Weights { NODE_COUNT, NODE_HOURS }
|
||||||
|
|
||||||
|
type NodeMetrics {
|
||||||
|
host: String!
|
||||||
|
subCluster: String!
|
||||||
|
metrics: [JobMetricWithName!]!
|
||||||
|
}
|
||||||
|
|
||||||
|
type Count {
|
||||||
|
name: String!
|
||||||
|
count: Int!
|
||||||
|
}
|
||||||
|
|
||||||
|
type User {
|
||||||
|
username: String!
|
||||||
|
name: String!
|
||||||
|
email: String!
|
||||||
|
}
|
||||||
|
|
||||||
|
type Query {
|
||||||
|
clusters: [Cluster!]! # List of all clusters
|
||||||
|
tags: [Tag!]! # List of all tags
|
||||||
|
|
||||||
|
user(username: String!): User
|
||||||
|
allocatedNodes(cluster: String!): [Count!]!
|
||||||
|
|
||||||
|
job(id: ID!): Job
|
||||||
|
jobMetrics(id: ID!, metrics: [String!], scopes: [MetricScope!]): [JobMetricWithName!]!
|
||||||
|
jobsFootprints(filter: [JobFilter!], metrics: [String!]!): Footprints
|
||||||
|
|
||||||
|
jobs(filter: [JobFilter!], page: PageRequest, order: OrderByInput): JobResultList!
|
||||||
|
jobsStatistics(filter: [JobFilter!], groupBy: Aggregate): [JobsStatistics!]!
|
||||||
|
jobsCount(filter: [JobFilter]!, groupBy: Aggregate!, weight: Weights, limit: Int): [Count!]!
|
||||||
|
|
||||||
|
rooflineHeatmap(filter: [JobFilter!]!, rows: Int!, cols: Int!, minX: Float!, minY: Float!, maxX: Float!, maxY: Float!): [[Float!]!]!
|
||||||
|
|
||||||
|
nodeMetrics(cluster: String!, nodes: [String!], scopes: [MetricScope!], metrics: [String!], from: Time!, to: Time!): [NodeMetrics!]!
|
||||||
|
}
|
||||||
|
|
||||||
|
type Mutation {
|
||||||
|
createTag(type: String!, name: String!): Tag!
|
||||||
|
deleteTag(id: ID!): ID!
|
||||||
|
addTagsToJob(job: ID!, tagIds: [ID!]!): [Tag!]!
|
||||||
|
removeTagsFromJob(job: ID!, tagIds: [ID!]!): [Tag!]!
|
||||||
|
|
||||||
|
updateConfiguration(name: String!, value: String!): String
|
||||||
|
}
|
||||||
|
|
||||||
|
type IntRangeOutput { from: Int!, to: Int! }
|
||||||
|
type TimeRangeOutput { from: Time!, to: Time! }
|
||||||
|
|
||||||
|
type FilterRanges {
|
||||||
|
duration: IntRangeOutput!
|
||||||
|
numNodes: IntRangeOutput!
|
||||||
|
startTime: TimeRangeOutput!
|
||||||
|
}
|
||||||
|
|
||||||
|
input JobFilter {
|
||||||
|
tags: [ID!]
|
||||||
|
jobId: StringInput
|
||||||
|
arrayJobId: Int
|
||||||
|
user: StringInput
|
||||||
|
project: StringInput
|
||||||
|
cluster: StringInput
|
||||||
|
partition: StringInput
|
||||||
|
duration: IntRange
|
||||||
|
|
||||||
|
minRunningFor: Int
|
||||||
|
|
||||||
|
numNodes: IntRange
|
||||||
|
numAccelerators: IntRange
|
||||||
|
numHWThreads: IntRange
|
||||||
|
|
||||||
|
startTime: TimeRange
|
||||||
|
state: [JobState!]
|
||||||
|
flopsAnyAvg: FloatRange
|
||||||
|
memBwAvg: FloatRange
|
||||||
|
loadAvg: FloatRange
|
||||||
|
memUsedMax: FloatRange
|
||||||
|
}
|
||||||
|
|
||||||
|
input OrderByInput {
|
||||||
|
field: String!
|
||||||
|
order: SortDirectionEnum! = ASC
|
||||||
|
}
|
||||||
|
|
||||||
|
enum SortDirectionEnum {
|
||||||
|
DESC
|
||||||
|
ASC
|
||||||
|
}
|
||||||
|
|
||||||
|
input StringInput {
|
||||||
|
eq: String
|
||||||
|
contains: String
|
||||||
|
startsWith: String
|
||||||
|
endsWith: String
|
||||||
|
}
|
||||||
|
|
||||||
|
input IntRange { from: Int!, to: Int! }
|
||||||
|
input FloatRange { from: Float!, to: Float! }
|
||||||
|
input TimeRange { from: Time, to: Time }
|
||||||
|
|
||||||
|
type JobResultList {
|
||||||
|
items: [Job!]!
|
||||||
|
offset: Int
|
||||||
|
limit: Int
|
||||||
|
count: Int
|
||||||
|
}
|
||||||
|
|
||||||
|
type HistoPoint {
|
||||||
|
count: Int!
|
||||||
|
value: Int!
|
||||||
|
}
|
||||||
|
|
||||||
|
type JobsStatistics {
|
||||||
|
id: ID! # If `groupBy` was used, ID of the user/project/cluster
|
||||||
|
totalJobs: Int! # Number of jobs that matched
|
||||||
|
shortJobs: Int! # Number of jobs with a duration of less than 2 minutes
|
||||||
|
totalWalltime: Int! # Sum of the duration of all matched jobs in hours
|
||||||
|
totalCoreHours: Int! # Sum of the core hours of all matched jobs
|
||||||
|
histDuration: [HistoPoint!]! # value: hour, count: number of jobs with a rounded duration of value
|
||||||
|
histNumNodes: [HistoPoint!]! # value: number of nodes, count: number of jobs with that number of nodes
|
||||||
|
}
|
||||||
|
|
||||||
|
input PageRequest {
|
||||||
|
itemsPerPage: Int!
|
||||||
|
page: Int!
|
||||||
|
}
|
@ -10,12 +10,12 @@ import (
|
|||||||
"strconv"
|
"strconv"
|
||||||
"time"
|
"time"
|
||||||
|
|
||||||
"github.com/ClusterCockpit/cc-backend/auth"
|
"github.com/ClusterCockpit/cc-backend/internal/auth"
|
||||||
"github.com/ClusterCockpit/cc-backend/config"
|
"github.com/ClusterCockpit/cc-backend/internal/config"
|
||||||
"github.com/ClusterCockpit/cc-backend/graph/generated"
|
"github.com/ClusterCockpit/cc-backend/internal/graph/generated"
|
||||||
"github.com/ClusterCockpit/cc-backend/graph/model"
|
"github.com/ClusterCockpit/cc-backend/internal/graph/model"
|
||||||
"github.com/ClusterCockpit/cc-backend/metricdata"
|
"github.com/ClusterCockpit/cc-backend/internal/metricdata"
|
||||||
"github.com/ClusterCockpit/cc-backend/schema"
|
"github.com/ClusterCockpit/cc-backend/pkg/schema"
|
||||||
)
|
)
|
||||||
|
|
||||||
func (r *clusterResolver) Partitions(ctx context.Context, obj *model.Cluster) ([]string, error) {
|
func (r *clusterResolver) Partitions(ctx context.Context, obj *model.Cluster) ([]string, error) {
|
@ -9,11 +9,11 @@ import (
|
|||||||
"time"
|
"time"
|
||||||
|
|
||||||
"github.com/99designs/gqlgen/graphql"
|
"github.com/99designs/gqlgen/graphql"
|
||||||
"github.com/ClusterCockpit/cc-backend/config"
|
"github.com/ClusterCockpit/cc-backend/internal/config"
|
||||||
"github.com/ClusterCockpit/cc-backend/graph/model"
|
"github.com/ClusterCockpit/cc-backend/internal/graph/model"
|
||||||
"github.com/ClusterCockpit/cc-backend/metricdata"
|
"github.com/ClusterCockpit/cc-backend/internal/metricdata"
|
||||||
"github.com/ClusterCockpit/cc-backend/repository"
|
"github.com/ClusterCockpit/cc-backend/internal/repository"
|
||||||
"github.com/ClusterCockpit/cc-backend/schema"
|
"github.com/ClusterCockpit/cc-backend/pkg/schema"
|
||||||
sq "github.com/Masterminds/squirrel"
|
sq "github.com/Masterminds/squirrel"
|
||||||
)
|
)
|
||||||
|
|
@ -13,8 +13,8 @@ import (
|
|||||||
"strconv"
|
"strconv"
|
||||||
"time"
|
"time"
|
||||||
|
|
||||||
"github.com/ClusterCockpit/cc-backend/config"
|
"github.com/ClusterCockpit/cc-backend/internal/config"
|
||||||
"github.com/ClusterCockpit/cc-backend/schema"
|
"github.com/ClusterCockpit/cc-backend/pkg/schema"
|
||||||
)
|
)
|
||||||
|
|
||||||
// For a given job, return the path of the `data.json`/`meta.json` file.
|
// For a given job, return the path of the `data.json`/`meta.json` file.
|
@ -11,8 +11,8 @@ import (
|
|||||||
"strings"
|
"strings"
|
||||||
"time"
|
"time"
|
||||||
|
|
||||||
"github.com/ClusterCockpit/cc-backend/config"
|
"github.com/ClusterCockpit/cc-backend/internal/config"
|
||||||
"github.com/ClusterCockpit/cc-backend/schema"
|
"github.com/ClusterCockpit/cc-backend/pkg/schema"
|
||||||
)
|
)
|
||||||
|
|
||||||
type CCMetricStoreConfig struct {
|
type CCMetricStoreConfig struct {
|
308
internal/metricdata/influxdb-v2.go
Normal file
308
internal/metricdata/influxdb-v2.go
Normal file
@ -0,0 +1,308 @@
|
|||||||
|
package metricdata
|
||||||
|
|
||||||
|
import (
|
||||||
|
"context"
|
||||||
|
"crypto/tls"
|
||||||
|
"encoding/json"
|
||||||
|
"errors"
|
||||||
|
"fmt"
|
||||||
|
"log"
|
||||||
|
"strings"
|
||||||
|
"time"
|
||||||
|
|
||||||
|
"github.com/ClusterCockpit/cc-backend/internal/config"
|
||||||
|
"github.com/ClusterCockpit/cc-backend/pkg/schema"
|
||||||
|
influxdb2 "github.com/influxdata/influxdb-client-go/v2"
|
||||||
|
influxdb2Api "github.com/influxdata/influxdb-client-go/v2/api"
|
||||||
|
)
|
||||||
|
|
||||||
|
type InfluxDBv2DataRepositoryConfig struct {
|
||||||
|
Url string `json:"url"`
|
||||||
|
Token string `json:"token"`
|
||||||
|
Bucket string `json:"bucket"`
|
||||||
|
Org string `json:"org"`
|
||||||
|
SkipTls bool `json:"skiptls"`
|
||||||
|
}
|
||||||
|
|
||||||
|
type InfluxDBv2DataRepository struct {
|
||||||
|
client influxdb2.Client
|
||||||
|
queryClient influxdb2Api.QueryAPI
|
||||||
|
bucket, measurement string
|
||||||
|
}
|
||||||
|
|
||||||
|
func (idb *InfluxDBv2DataRepository) Init(rawConfig json.RawMessage) error {
|
||||||
|
var config InfluxDBv2DataRepositoryConfig
|
||||||
|
if err := json.Unmarshal(rawConfig, &config); err != nil {
|
||||||
|
return err
|
||||||
|
}
|
||||||
|
|
||||||
|
idb.client = influxdb2.NewClientWithOptions(config.Url, config.Token, influxdb2.DefaultOptions().SetTLSConfig(&tls.Config{InsecureSkipVerify: config.SkipTls}))
|
||||||
|
idb.queryClient = idb.client.QueryAPI(config.Org)
|
||||||
|
idb.bucket = config.Bucket
|
||||||
|
|
||||||
|
return nil
|
||||||
|
}
|
||||||
|
|
||||||
|
func (idb *InfluxDBv2DataRepository) formatTime(t time.Time) string {
|
||||||
|
return t.Format(time.RFC3339) // Like “2006-01-02T15:04:05Z07:00”
|
||||||
|
}
|
||||||
|
|
||||||
|
func (idb *InfluxDBv2DataRepository) epochToTime(epoch int64) time.Time {
|
||||||
|
return time.Unix(epoch, 0)
|
||||||
|
}
|
||||||
|
|
||||||
|
func (idb *InfluxDBv2DataRepository) LoadData(job *schema.Job, metrics []string, scopes []schema.MetricScope, ctx context.Context) (schema.JobData, error) {
|
||||||
|
|
||||||
|
measurementsConds := make([]string, 0, len(metrics))
|
||||||
|
for _, m := range metrics {
|
||||||
|
measurementsConds = append(measurementsConds, fmt.Sprintf(`r["_measurement"] == "%s"`, m))
|
||||||
|
}
|
||||||
|
measurementsCond := strings.Join(measurementsConds, " or ")
|
||||||
|
|
||||||
|
hostsConds := make([]string, 0, len(job.Resources))
|
||||||
|
for _, h := range job.Resources {
|
||||||
|
if h.HWThreads != nil || h.Accelerators != nil {
|
||||||
|
// TODO
|
||||||
|
return nil, errors.New("the InfluxDB metric data repository does not yet support HWThreads or Accelerators")
|
||||||
|
}
|
||||||
|
hostsConds = append(hostsConds, fmt.Sprintf(`r["hostname"] == "%s"`, h.Hostname))
|
||||||
|
}
|
||||||
|
hostsCond := strings.Join(hostsConds, " or ")
|
||||||
|
|
||||||
|
jobData := make(schema.JobData) // Empty Schema: map[<string>FIELD]map[<MetricScope>SCOPE]<*JobMetric>METRIC
|
||||||
|
// Requested Scopes
|
||||||
|
for _, scope := range scopes {
|
||||||
|
query := ""
|
||||||
|
switch scope {
|
||||||
|
case "node":
|
||||||
|
// Get Finest Granularity, Groupy By Measurement and Hostname (== Metric / Node), Calculate Mean for 60s windows
|
||||||
|
// log.Println("Note: Scope 'node' requested. ")
|
||||||
|
query = fmt.Sprintf(`
|
||||||
|
from(bucket: "%s")
|
||||||
|
|> range(start: %s, stop: %s)
|
||||||
|
|> filter(fn: (r) => (%s) and (%s) )
|
||||||
|
|> drop(columns: ["_start", "_stop"])
|
||||||
|
|> group(columns: ["hostname", "_measurement"])
|
||||||
|
|> aggregateWindow(every: 60s, fn: mean)
|
||||||
|
|> drop(columns: ["_time"])`,
|
||||||
|
idb.bucket,
|
||||||
|
idb.formatTime(job.StartTime), idb.formatTime(idb.epochToTime(job.StartTimeUnix+int64(job.Duration)+int64(1))),
|
||||||
|
measurementsCond, hostsCond)
|
||||||
|
case "socket":
|
||||||
|
log.Println("Note: Scope 'socket' requested, but not yet supported: Will return 'node' scope only. ")
|
||||||
|
continue
|
||||||
|
case "core":
|
||||||
|
log.Println("Note: Scope 'core' requested, but not yet supported: Will return 'node' scope only. ")
|
||||||
|
continue
|
||||||
|
// Get Finest Granularity only, Set NULL to 0.0
|
||||||
|
// query = fmt.Sprintf(`
|
||||||
|
// from(bucket: "%s")
|
||||||
|
// |> range(start: %s, stop: %s)
|
||||||
|
// |> filter(fn: (r) => %s )
|
||||||
|
// |> filter(fn: (r) => %s )
|
||||||
|
// |> drop(columns: ["_start", "_stop", "cluster"])
|
||||||
|
// |> map(fn: (r) => (if exists r._value then {r with _value: r._value} else {r with _value: 0.0}))`,
|
||||||
|
// idb.bucket,
|
||||||
|
// idb.formatTime(job.StartTime), idb.formatTime(idb.epochToTime(job.StartTimeUnix + int64(job.Duration) + int64(1) )),
|
||||||
|
// measurementsCond, hostsCond)
|
||||||
|
default:
|
||||||
|
log.Println("Note: Unknown Scope requested: Will return 'node' scope. ")
|
||||||
|
continue
|
||||||
|
// return nil, errors.New("the InfluxDB metric data repository does not yet support other scopes than 'node'")
|
||||||
|
}
|
||||||
|
|
||||||
|
rows, err := idb.queryClient.Query(ctx, query)
|
||||||
|
if err != nil {
|
||||||
|
return nil, err
|
||||||
|
}
|
||||||
|
|
||||||
|
// Init Metrics: Only Node level now -> TODO: Matching /check on scope level ...
|
||||||
|
for _, metric := range metrics {
|
||||||
|
jobMetric, ok := jobData[metric]
|
||||||
|
if !ok {
|
||||||
|
mc := config.GetMetricConfig(job.Cluster, metric)
|
||||||
|
jobMetric = map[schema.MetricScope]*schema.JobMetric{
|
||||||
|
scope: { // uses scope var from above!
|
||||||
|
Unit: mc.Unit,
|
||||||
|
Scope: scope,
|
||||||
|
Timestep: mc.Timestep,
|
||||||
|
Series: make([]schema.Series, 0, len(job.Resources)),
|
||||||
|
StatisticsSeries: nil, // Should be: &schema.StatsSeries{},
|
||||||
|
},
|
||||||
|
}
|
||||||
|
}
|
||||||
|
jobData[metric] = jobMetric
|
||||||
|
}
|
||||||
|
|
||||||
|
// Process Result: Time-Data
|
||||||
|
field, host, hostSeries := "", "", schema.Series{}
|
||||||
|
// typeId := 0
|
||||||
|
switch scope {
|
||||||
|
case "node":
|
||||||
|
for rows.Next() {
|
||||||
|
row := rows.Record()
|
||||||
|
if host == "" || host != row.ValueByKey("hostname").(string) || rows.TableChanged() {
|
||||||
|
if host != "" {
|
||||||
|
// Append Series before reset
|
||||||
|
jobData[field][scope].Series = append(jobData[field][scope].Series, hostSeries)
|
||||||
|
}
|
||||||
|
field, host = row.Measurement(), row.ValueByKey("hostname").(string)
|
||||||
|
hostSeries = schema.Series{
|
||||||
|
Hostname: host,
|
||||||
|
Statistics: nil,
|
||||||
|
Data: make([]schema.Float, 0),
|
||||||
|
}
|
||||||
|
}
|
||||||
|
val, ok := row.Value().(float64)
|
||||||
|
if ok {
|
||||||
|
hostSeries.Data = append(hostSeries.Data, schema.Float(val))
|
||||||
|
} else {
|
||||||
|
hostSeries.Data = append(hostSeries.Data, schema.Float(0))
|
||||||
|
}
|
||||||
|
}
|
||||||
|
case "socket":
|
||||||
|
continue
|
||||||
|
case "core":
|
||||||
|
continue
|
||||||
|
// Include Series.Id in hostSeries
|
||||||
|
// for rows.Next() {
|
||||||
|
// row := rows.Record()
|
||||||
|
// if ( host == "" || host != row.ValueByKey("hostname").(string) || typeId != row.ValueByKey("type-id").(int) || rows.TableChanged() ) {
|
||||||
|
// if ( host != "" ) {
|
||||||
|
// // Append Series before reset
|
||||||
|
// jobData[field][scope].Series = append(jobData[field][scope].Series, hostSeries)
|
||||||
|
// }
|
||||||
|
// field, host, typeId = row.Measurement(), row.ValueByKey("hostname").(string), row.ValueByKey("type-id").(int)
|
||||||
|
// hostSeries = schema.Series{
|
||||||
|
// Hostname: host,
|
||||||
|
// Id: &typeId,
|
||||||
|
// Statistics: nil,
|
||||||
|
// Data: make([]schema.Float, 0),
|
||||||
|
// }
|
||||||
|
// }
|
||||||
|
// val := row.Value().(float64)
|
||||||
|
// hostSeries.Data = append(hostSeries.Data, schema.Float(val))
|
||||||
|
// }
|
||||||
|
default:
|
||||||
|
continue
|
||||||
|
// return nil, errors.New("the InfluxDB metric data repository does not yet support other scopes than 'node, core'")
|
||||||
|
}
|
||||||
|
// Append last Series
|
||||||
|
jobData[field][scope].Series = append(jobData[field][scope].Series, hostSeries)
|
||||||
|
}
|
||||||
|
|
||||||
|
// Get Stats
|
||||||
|
stats, err := idb.LoadStats(job, metrics, ctx)
|
||||||
|
if err != nil {
|
||||||
|
return nil, err
|
||||||
|
}
|
||||||
|
|
||||||
|
for _, scope := range scopes {
|
||||||
|
if scope == "node" { // No 'socket/core' support yet
|
||||||
|
for metric, nodes := range stats {
|
||||||
|
// log.Println(fmt.Sprintf("<< Add Stats for : Field %s >>", metric))
|
||||||
|
for node, stats := range nodes {
|
||||||
|
// log.Println(fmt.Sprintf("<< Add Stats for : Host %s : Min %.2f, Max %.2f, Avg %.2f >>", node, stats.Min, stats.Max, stats.Avg ))
|
||||||
|
for index, _ := range jobData[metric][scope].Series {
|
||||||
|
// log.Println(fmt.Sprintf("<< Try to add Stats to Series in Position %d >>", index))
|
||||||
|
if jobData[metric][scope].Series[index].Hostname == node {
|
||||||
|
// log.Println(fmt.Sprintf("<< Match for Series in Position %d : Host %s >>", index, jobData[metric][scope].Series[index].Hostname))
|
||||||
|
jobData[metric][scope].Series[index].Statistics = &schema.MetricStatistics{Avg: stats.Avg, Min: stats.Min, Max: stats.Max}
|
||||||
|
// log.Println(fmt.Sprintf("<< Result Inner: Min %.2f, Max %.2f, Avg %.2f >>", jobData[metric][scope].Series[index].Statistics.Min, jobData[metric][scope].Series[index].Statistics.Max, jobData[metric][scope].Series[index].Statistics.Avg))
|
||||||
|
}
|
||||||
|
}
|
||||||
|
}
|
||||||
|
}
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
// DEBUG:
|
||||||
|
// for _, scope := range scopes {
|
||||||
|
// for _, met := range metrics {
|
||||||
|
// for _, series := range jobData[met][scope].Series {
|
||||||
|
// log.Println(fmt.Sprintf("<< Result: %d data points for metric %s on %s with scope %s, Stats: Min %.2f, Max %.2f, Avg %.2f >>",
|
||||||
|
// len(series.Data), met, series.Hostname, scope,
|
||||||
|
// series.Statistics.Min, series.Statistics.Max, series.Statistics.Avg))
|
||||||
|
// }
|
||||||
|
// }
|
||||||
|
// }
|
||||||
|
|
||||||
|
return jobData, nil
|
||||||
|
}
|
||||||
|
|
||||||
|
func (idb *InfluxDBv2DataRepository) LoadStats(job *schema.Job, metrics []string, ctx context.Context) (map[string]map[string]schema.MetricStatistics, error) {
|
||||||
|
|
||||||
|
stats := map[string]map[string]schema.MetricStatistics{}
|
||||||
|
|
||||||
|
hostsConds := make([]string, 0, len(job.Resources))
|
||||||
|
for _, h := range job.Resources {
|
||||||
|
if h.HWThreads != nil || h.Accelerators != nil {
|
||||||
|
// TODO
|
||||||
|
return nil, errors.New("the InfluxDB metric data repository does not yet support HWThreads or Accelerators")
|
||||||
|
}
|
||||||
|
hostsConds = append(hostsConds, fmt.Sprintf(`r["hostname"] == "%s"`, h.Hostname))
|
||||||
|
}
|
||||||
|
hostsCond := strings.Join(hostsConds, " or ")
|
||||||
|
|
||||||
|
// lenMet := len(metrics)
|
||||||
|
|
||||||
|
for _, metric := range metrics {
|
||||||
|
// log.Println(fmt.Sprintf("<< You are here: %s (Index %d of %d metrics)", metric, index, lenMet))
|
||||||
|
|
||||||
|
query := fmt.Sprintf(`
|
||||||
|
data = from(bucket: "%s")
|
||||||
|
|> range(start: %s, stop: %s)
|
||||||
|
|> filter(fn: (r) => r._measurement == "%s" and r._field == "value" and (%s))
|
||||||
|
union(tables: [data |> mean(column: "_value") |> set(key: "_field", value: "avg"),
|
||||||
|
data |> min(column: "_value") |> set(key: "_field", value: "min"),
|
||||||
|
data |> max(column: "_value") |> set(key: "_field", value: "max")])
|
||||||
|
|> pivot(rowKey: ["hostname"], columnKey: ["_field"], valueColumn: "_value")
|
||||||
|
|> group()`,
|
||||||
|
idb.bucket,
|
||||||
|
idb.formatTime(job.StartTime), idb.formatTime(idb.epochToTime(job.StartTimeUnix+int64(job.Duration)+int64(1))),
|
||||||
|
metric, hostsCond)
|
||||||
|
|
||||||
|
rows, err := idb.queryClient.Query(ctx, query)
|
||||||
|
if err != nil {
|
||||||
|
return nil, err
|
||||||
|
}
|
||||||
|
|
||||||
|
nodes := map[string]schema.MetricStatistics{}
|
||||||
|
for rows.Next() {
|
||||||
|
row := rows.Record()
|
||||||
|
host := row.ValueByKey("hostname").(string)
|
||||||
|
|
||||||
|
avg, avgok := row.ValueByKey("avg").(float64)
|
||||||
|
if !avgok {
|
||||||
|
// log.Println(fmt.Sprintf(">> Assertion error for metric %s, statistic AVG. Expected 'float64', got %v", metric, avg))
|
||||||
|
avg = 0.0
|
||||||
|
}
|
||||||
|
min, minok := row.ValueByKey("min").(float64)
|
||||||
|
if !minok {
|
||||||
|
// log.Println(fmt.Sprintf(">> Assertion error for metric %s, statistic MIN. Expected 'float64', got %v", metric, min))
|
||||||
|
min = 0.0
|
||||||
|
}
|
||||||
|
max, maxok := row.ValueByKey("max").(float64)
|
||||||
|
if !maxok {
|
||||||
|
// log.Println(fmt.Sprintf(">> Assertion error for metric %s, statistic MAX. Expected 'float64', got %v", metric, max))
|
||||||
|
max = 0.0
|
||||||
|
}
|
||||||
|
|
||||||
|
nodes[host] = schema.MetricStatistics{
|
||||||
|
Avg: avg,
|
||||||
|
Min: min,
|
||||||
|
Max: max,
|
||||||
|
}
|
||||||
|
}
|
||||||
|
stats[metric] = nodes
|
||||||
|
}
|
||||||
|
|
||||||
|
return stats, nil
|
||||||
|
}
|
||||||
|
|
||||||
|
func (idb *InfluxDBv2DataRepository) LoadNodeData(cluster string, metrics, nodes []string, scopes []schema.MetricScope, from, to time.Time, ctx context.Context) (map[string]map[string][]*schema.JobMetric, error) {
|
||||||
|
// TODO : Implement to be used in Analysis- und System/Node-View
|
||||||
|
log.Println(fmt.Sprintf("LoadNodeData unimplemented for InfluxDBv2DataRepository, Args: cluster %s, metrics %v, nodes %v, scopes %v", cluster, metrics, nodes, scopes))
|
||||||
|
|
||||||
|
return nil, errors.New("unimplemented for InfluxDBv2DataRepository")
|
||||||
|
}
|
@ -6,9 +6,9 @@ import (
|
|||||||
"fmt"
|
"fmt"
|
||||||
"time"
|
"time"
|
||||||
|
|
||||||
"github.com/ClusterCockpit/cc-backend/config"
|
"github.com/ClusterCockpit/cc-backend/internal/config"
|
||||||
"github.com/ClusterCockpit/cc-backend/log"
|
"github.com/ClusterCockpit/cc-backend/pkg/log"
|
||||||
"github.com/ClusterCockpit/cc-backend/schema"
|
"github.com/ClusterCockpit/cc-backend/pkg/schema"
|
||||||
"github.com/iamlouk/lrucache"
|
"github.com/iamlouk/lrucache"
|
||||||
)
|
)
|
||||||
|
|
@ -5,7 +5,7 @@ import (
|
|||||||
"encoding/json"
|
"encoding/json"
|
||||||
"time"
|
"time"
|
||||||
|
|
||||||
"github.com/ClusterCockpit/cc-backend/schema"
|
"github.com/ClusterCockpit/cc-backend/pkg/schema"
|
||||||
)
|
)
|
||||||
|
|
||||||
var TestLoadDataCallback func(job *schema.Job, metrics []string, scopes []schema.MetricScope, ctx context.Context) (schema.JobData, error) = func(job *schema.Job, metrics []string, scopes []schema.MetricScope, ctx context.Context) (schema.JobData, error) {
|
var TestLoadDataCallback func(job *schema.Job, metrics []string, scopes []schema.MetricScope, ctx context.Context) (schema.JobData, error) = func(job *schema.Job, metrics []string, scopes []schema.MetricScope, ctx context.Context) (schema.JobData, error) {
|
58
internal/repository/dbConnection.go
Normal file
58
internal/repository/dbConnection.go
Normal file
@ -0,0 +1,58 @@
|
|||||||
|
package repository
|
||||||
|
|
||||||
|
import (
|
||||||
|
"fmt"
|
||||||
|
"log"
|
||||||
|
"sync"
|
||||||
|
"time"
|
||||||
|
|
||||||
|
"github.com/jmoiron/sqlx"
|
||||||
|
)
|
||||||
|
|
||||||
|
var (
|
||||||
|
dbConnOnce sync.Once
|
||||||
|
dbConnInstance *DBConnection
|
||||||
|
)
|
||||||
|
|
||||||
|
type DBConnection struct {
|
||||||
|
DB *sqlx.DB
|
||||||
|
}
|
||||||
|
|
||||||
|
func Connect(driver string, db string) {
|
||||||
|
var err error
|
||||||
|
var dbHandle *sqlx.DB
|
||||||
|
|
||||||
|
dbConnOnce.Do(func() {
|
||||||
|
if driver == "sqlite3" {
|
||||||
|
dbHandle, err = sqlx.Open("sqlite3", fmt.Sprintf("%s?_foreign_keys=on", db))
|
||||||
|
if err != nil {
|
||||||
|
log.Fatal(err)
|
||||||
|
}
|
||||||
|
|
||||||
|
// sqlite does not multithread. Having more than one connection open would just mean
|
||||||
|
// waiting for locks.
|
||||||
|
dbHandle.SetMaxOpenConns(1)
|
||||||
|
} else if driver == "mysql" {
|
||||||
|
dbHandle, err = sqlx.Open("mysql", fmt.Sprintf("%s?multiStatements=true", db))
|
||||||
|
if err != nil {
|
||||||
|
log.Fatal(err)
|
||||||
|
}
|
||||||
|
|
||||||
|
dbHandle.SetConnMaxLifetime(time.Minute * 3)
|
||||||
|
dbHandle.SetMaxOpenConns(10)
|
||||||
|
dbHandle.SetMaxIdleConns(10)
|
||||||
|
} else {
|
||||||
|
log.Fatalf("unsupported database driver: %s", driver)
|
||||||
|
}
|
||||||
|
|
||||||
|
dbConnInstance = &DBConnection{DB: dbHandle}
|
||||||
|
})
|
||||||
|
}
|
||||||
|
|
||||||
|
func GetConnection() *DBConnection {
|
||||||
|
if dbConnInstance == nil {
|
||||||
|
log.Fatalf("Database connection not initialized!")
|
||||||
|
}
|
||||||
|
|
||||||
|
return dbConnInstance
|
||||||
|
}
|
@ -9,10 +9,10 @@ import (
|
|||||||
"strings"
|
"strings"
|
||||||
"time"
|
"time"
|
||||||
|
|
||||||
"github.com/ClusterCockpit/cc-backend/config"
|
"github.com/ClusterCockpit/cc-backend/internal/config"
|
||||||
"github.com/ClusterCockpit/cc-backend/log"
|
"github.com/ClusterCockpit/cc-backend/internal/metricdata"
|
||||||
"github.com/ClusterCockpit/cc-backend/metricdata"
|
"github.com/ClusterCockpit/cc-backend/pkg/log"
|
||||||
"github.com/ClusterCockpit/cc-backend/schema"
|
"github.com/ClusterCockpit/cc-backend/pkg/schema"
|
||||||
)
|
)
|
||||||
|
|
||||||
const NamedJobInsert string = `INSERT INTO job (
|
const NamedJobInsert string = `INSERT INTO job (
|
@ -8,8 +8,8 @@ import (
|
|||||||
"path/filepath"
|
"path/filepath"
|
||||||
"time"
|
"time"
|
||||||
|
|
||||||
"github.com/ClusterCockpit/cc-backend/log"
|
"github.com/ClusterCockpit/cc-backend/pkg/log"
|
||||||
"github.com/ClusterCockpit/cc-backend/schema"
|
"github.com/ClusterCockpit/cc-backend/pkg/schema"
|
||||||
"github.com/jmoiron/sqlx"
|
"github.com/jmoiron/sqlx"
|
||||||
)
|
)
|
||||||
|
|
@ -7,17 +7,23 @@ import (
|
|||||||
"errors"
|
"errors"
|
||||||
"fmt"
|
"fmt"
|
||||||
"strconv"
|
"strconv"
|
||||||
|
"sync"
|
||||||
"time"
|
"time"
|
||||||
|
|
||||||
"github.com/ClusterCockpit/cc-backend/auth"
|
"github.com/ClusterCockpit/cc-backend/internal/auth"
|
||||||
"github.com/ClusterCockpit/cc-backend/graph/model"
|
"github.com/ClusterCockpit/cc-backend/internal/graph/model"
|
||||||
"github.com/ClusterCockpit/cc-backend/log"
|
"github.com/ClusterCockpit/cc-backend/pkg/log"
|
||||||
"github.com/ClusterCockpit/cc-backend/schema"
|
"github.com/ClusterCockpit/cc-backend/pkg/schema"
|
||||||
sq "github.com/Masterminds/squirrel"
|
sq "github.com/Masterminds/squirrel"
|
||||||
"github.com/iamlouk/lrucache"
|
"github.com/iamlouk/lrucache"
|
||||||
"github.com/jmoiron/sqlx"
|
"github.com/jmoiron/sqlx"
|
||||||
)
|
)
|
||||||
|
|
||||||
|
var (
|
||||||
|
jobRepoOnce sync.Once
|
||||||
|
jobRepoInstance *JobRepository
|
||||||
|
)
|
||||||
|
|
||||||
type JobRepository struct {
|
type JobRepository struct {
|
||||||
DB *sqlx.DB
|
DB *sqlx.DB
|
||||||
|
|
||||||
@ -25,10 +31,18 @@ type JobRepository struct {
|
|||||||
cache *lrucache.Cache
|
cache *lrucache.Cache
|
||||||
}
|
}
|
||||||
|
|
||||||
func (r *JobRepository) Init() error {
|
func GetRepository() *JobRepository {
|
||||||
r.stmtCache = sq.NewStmtCache(r.DB)
|
jobRepoOnce.Do(func() {
|
||||||
r.cache = lrucache.New(1024 * 1024)
|
db := GetConnection()
|
||||||
return nil
|
|
||||||
|
jobRepoInstance = &JobRepository{
|
||||||
|
DB: db.DB,
|
||||||
|
stmtCache: sq.NewStmtCache(db.DB),
|
||||||
|
cache: lrucache.New(1024 * 1024),
|
||||||
|
}
|
||||||
|
})
|
||||||
|
|
||||||
|
return jobRepoInstance
|
||||||
}
|
}
|
||||||
|
|
||||||
var jobColumns []string = []string{
|
var jobColumns []string = []string{
|
@ -11,22 +11,11 @@ import (
|
|||||||
var db *sqlx.DB
|
var db *sqlx.DB
|
||||||
|
|
||||||
func init() {
|
func init() {
|
||||||
var err error
|
Connect("sqlite3", "../../test/test.db")
|
||||||
db, err = sqlx.Open("sqlite3", "../test/test.db")
|
|
||||||
if err != nil {
|
|
||||||
fmt.Println(err)
|
|
||||||
}
|
|
||||||
}
|
}
|
||||||
|
|
||||||
func setup(t *testing.T) *JobRepository {
|
func setup(t *testing.T) *JobRepository {
|
||||||
r := &JobRepository{
|
return GetRepository()
|
||||||
DB: db,
|
|
||||||
}
|
|
||||||
if err := r.Init(); err != nil {
|
|
||||||
t.Fatal(err)
|
|
||||||
}
|
|
||||||
|
|
||||||
return r
|
|
||||||
}
|
}
|
||||||
|
|
||||||
func TestFind(t *testing.T) {
|
func TestFind(t *testing.T) {
|
@ -8,10 +8,10 @@ import (
|
|||||||
"strings"
|
"strings"
|
||||||
"time"
|
"time"
|
||||||
|
|
||||||
"github.com/ClusterCockpit/cc-backend/auth"
|
"github.com/ClusterCockpit/cc-backend/internal/auth"
|
||||||
"github.com/ClusterCockpit/cc-backend/graph/model"
|
"github.com/ClusterCockpit/cc-backend/internal/graph/model"
|
||||||
"github.com/ClusterCockpit/cc-backend/log"
|
"github.com/ClusterCockpit/cc-backend/pkg/log"
|
||||||
"github.com/ClusterCockpit/cc-backend/schema"
|
"github.com/ClusterCockpit/cc-backend/pkg/schema"
|
||||||
sq "github.com/Masterminds/squirrel"
|
sq "github.com/Masterminds/squirrel"
|
||||||
)
|
)
|
||||||
|
|
@ -1,8 +1,8 @@
|
|||||||
package repository
|
package repository
|
||||||
|
|
||||||
import (
|
import (
|
||||||
"github.com/ClusterCockpit/cc-backend/metricdata"
|
"github.com/ClusterCockpit/cc-backend/internal/metricdata"
|
||||||
"github.com/ClusterCockpit/cc-backend/schema"
|
"github.com/ClusterCockpit/cc-backend/pkg/schema"
|
||||||
sq "github.com/Masterminds/squirrel"
|
sq "github.com/Masterminds/squirrel"
|
||||||
)
|
)
|
||||||
|
|
@ -1,4 +1,4 @@
|
|||||||
package main
|
package routerConfig
|
||||||
|
|
||||||
import (
|
import (
|
||||||
"fmt"
|
"fmt"
|
||||||
@ -8,13 +8,14 @@ import (
|
|||||||
"strings"
|
"strings"
|
||||||
"time"
|
"time"
|
||||||
|
|
||||||
"github.com/ClusterCockpit/cc-backend/auth"
|
"github.com/ClusterCockpit/cc-backend/internal/auth"
|
||||||
"github.com/ClusterCockpit/cc-backend/config"
|
"github.com/ClusterCockpit/cc-backend/internal/config"
|
||||||
"github.com/ClusterCockpit/cc-backend/graph"
|
"github.com/ClusterCockpit/cc-backend/internal/graph"
|
||||||
"github.com/ClusterCockpit/cc-backend/graph/model"
|
"github.com/ClusterCockpit/cc-backend/internal/graph/model"
|
||||||
"github.com/ClusterCockpit/cc-backend/log"
|
"github.com/ClusterCockpit/cc-backend/internal/repository"
|
||||||
"github.com/ClusterCockpit/cc-backend/schema"
|
"github.com/ClusterCockpit/cc-backend/internal/templates"
|
||||||
"github.com/ClusterCockpit/cc-backend/templates"
|
"github.com/ClusterCockpit/cc-backend/pkg/log"
|
||||||
|
"github.com/ClusterCockpit/cc-backend/pkg/schema"
|
||||||
"github.com/gorilla/mux"
|
"github.com/gorilla/mux"
|
||||||
)
|
)
|
||||||
|
|
||||||
@ -50,6 +51,7 @@ func setupHomeRoute(i InfoType, r *http.Request) InfoType {
|
|||||||
TotalJobs int
|
TotalJobs int
|
||||||
RecentShortJobs int
|
RecentShortJobs int
|
||||||
}
|
}
|
||||||
|
jobRepo := repository.GetRepository()
|
||||||
|
|
||||||
runningJobs, err := jobRepo.CountGroupedJobs(r.Context(), model.AggregateCluster, []*model.JobFilter{{
|
runningJobs, err := jobRepo.CountGroupedJobs(r.Context(), model.AggregateCluster, []*model.JobFilter{{
|
||||||
State: []schema.JobState{schema.JobStateRunning},
|
State: []schema.JobState{schema.JobStateRunning},
|
||||||
@ -93,6 +95,7 @@ func setupJobRoute(i InfoType, r *http.Request) InfoType {
|
|||||||
}
|
}
|
||||||
|
|
||||||
func setupUserRoute(i InfoType, r *http.Request) InfoType {
|
func setupUserRoute(i InfoType, r *http.Request) InfoType {
|
||||||
|
jobRepo := repository.GetRepository()
|
||||||
username := mux.Vars(r)["id"]
|
username := mux.Vars(r)["id"]
|
||||||
i["id"] = username
|
i["id"] = username
|
||||||
i["username"] = username
|
i["username"] = username
|
||||||
@ -135,6 +138,7 @@ func setupAnalysisRoute(i InfoType, r *http.Request) InfoType {
|
|||||||
|
|
||||||
func setupTaglistRoute(i InfoType, r *http.Request) InfoType {
|
func setupTaglistRoute(i InfoType, r *http.Request) InfoType {
|
||||||
var username *string = nil
|
var username *string = nil
|
||||||
|
jobRepo := repository.GetRepository()
|
||||||
if user := auth.GetUser(r.Context()); user != nil && !user.HasRole(auth.RoleAdmin) {
|
if user := auth.GetUser(r.Context()); user != nil && !user.HasRole(auth.RoleAdmin) {
|
||||||
username = &user.Username
|
username = &user.Username
|
||||||
}
|
}
|
||||||
@ -245,7 +249,7 @@ func buildFilterPresets(query url.Values) map[string]interface{} {
|
|||||||
return filterPresets
|
return filterPresets
|
||||||
}
|
}
|
||||||
|
|
||||||
func setupRoutes(router *mux.Router, routes []Route) {
|
func SetupRoutes(router *mux.Router) {
|
||||||
for _, route := range routes {
|
for _, route := range routes {
|
||||||
route := route
|
route := route
|
||||||
router.HandleFunc(route.Route, func(rw http.ResponseWriter, r *http.Request) {
|
router.HandleFunc(route.Route, func(rw http.ResponseWriter, r *http.Request) {
|
@ -1,4 +1,4 @@
|
|||||||
package main
|
package runtimeEnv
|
||||||
|
|
||||||
import (
|
import (
|
||||||
"bufio"
|
"bufio"
|
||||||
@ -15,7 +15,7 @@ import (
|
|||||||
// Very simple and limited .env file reader.
|
// Very simple and limited .env file reader.
|
||||||
// All variable definitions found are directly
|
// All variable definitions found are directly
|
||||||
// added to the processes environment.
|
// added to the processes environment.
|
||||||
func loadEnv(file string) error {
|
func LoadEnv(file string) error {
|
||||||
f, err := os.Open(file)
|
f, err := os.Open(file)
|
||||||
if err != nil {
|
if err != nil {
|
||||||
return err
|
return err
|
||||||
@ -81,9 +81,9 @@ func loadEnv(file string) error {
|
|||||||
// specified in the config.json. The go runtime
|
// specified in the config.json. The go runtime
|
||||||
// takes care of all threads (and not only the calling one)
|
// takes care of all threads (and not only the calling one)
|
||||||
// executing the underlying systemcall.
|
// executing the underlying systemcall.
|
||||||
func dropPrivileges() error {
|
func DropPrivileges(username string, group string) error {
|
||||||
if programConfig.Group != "" {
|
if group != "" {
|
||||||
g, err := user.LookupGroup(programConfig.Group)
|
g, err := user.LookupGroup(group)
|
||||||
if err != nil {
|
if err != nil {
|
||||||
return err
|
return err
|
||||||
}
|
}
|
||||||
@ -94,8 +94,8 @@ func dropPrivileges() error {
|
|||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
if programConfig.User != "" {
|
if username != "" {
|
||||||
u, err := user.Lookup(programConfig.User)
|
u, err := user.Lookup(username)
|
||||||
if err != nil {
|
if err != nil {
|
||||||
return err
|
return err
|
||||||
}
|
}
|
||||||
@ -111,7 +111,7 @@ func dropPrivileges() error {
|
|||||||
|
|
||||||
// If started via systemd, inform systemd that we are running:
|
// If started via systemd, inform systemd that we are running:
|
||||||
// https://www.freedesktop.org/software/systemd/man/sd_notify.html
|
// https://www.freedesktop.org/software/systemd/man/sd_notify.html
|
||||||
func systemdNotifiy(ready bool, status string) {
|
func SystemdNotifiy(ready bool, status string) {
|
||||||
if os.Getenv("NOTIFY_SOCKET") == "" {
|
if os.Getenv("NOTIFY_SOCKET") == "" {
|
||||||
// Not started using systemd
|
// Not started using systemd
|
||||||
return
|
return
|
@ -5,8 +5,8 @@ import (
|
|||||||
"net/http"
|
"net/http"
|
||||||
"os"
|
"os"
|
||||||
|
|
||||||
"github.com/ClusterCockpit/cc-backend/config"
|
"github.com/ClusterCockpit/cc-backend/internal/config"
|
||||||
"github.com/ClusterCockpit/cc-backend/log"
|
"github.com/ClusterCockpit/cc-backend/pkg/log"
|
||||||
)
|
)
|
||||||
|
|
||||||
var templatesDir string
|
var templatesDir string
|
||||||
@ -36,7 +36,7 @@ func init() {
|
|||||||
if ebp != "" {
|
if ebp != "" {
|
||||||
bp = ebp
|
bp = ebp
|
||||||
}
|
}
|
||||||
templatesDir = bp + "templates/"
|
templatesDir = bp + "web/templates/"
|
||||||
base := template.Must(template.ParseFiles(templatesDir + "base.tmpl"))
|
base := template.Must(template.ParseFiles(templatesDir + "base.tmpl"))
|
||||||
files := []string{
|
files := []string{
|
||||||
"home.tmpl", "404.tmpl", "login.tmpl",
|
"home.tmpl", "404.tmpl", "login.tmpl",
|
@ -1,308 +0,0 @@
|
|||||||
package metricdata
|
|
||||||
|
|
||||||
import (
|
|
||||||
"context"
|
|
||||||
"errors"
|
|
||||||
"fmt"
|
|
||||||
"log"
|
|
||||||
"strings"
|
|
||||||
"time"
|
|
||||||
"crypto/tls"
|
|
||||||
"encoding/json"
|
|
||||||
|
|
||||||
"github.com/ClusterCockpit/cc-backend/config"
|
|
||||||
"github.com/ClusterCockpit/cc-backend/schema"
|
|
||||||
influxdb2 "github.com/influxdata/influxdb-client-go/v2"
|
|
||||||
influxdb2Api "github.com/influxdata/influxdb-client-go/v2/api"
|
|
||||||
)
|
|
||||||
|
|
||||||
type InfluxDBv2DataRepositoryConfig struct {
|
|
||||||
Url string `json:"url"`
|
|
||||||
Token string `json:"token"`
|
|
||||||
Bucket string `json:"bucket"`
|
|
||||||
Org string `json:"org"`
|
|
||||||
SkipTls bool `json:"skiptls"`
|
|
||||||
}
|
|
||||||
|
|
||||||
type InfluxDBv2DataRepository struct {
|
|
||||||
client influxdb2.Client
|
|
||||||
queryClient influxdb2Api.QueryAPI
|
|
||||||
bucket, measurement string
|
|
||||||
}
|
|
||||||
|
|
||||||
func (idb *InfluxDBv2DataRepository) Init(rawConfig json.RawMessage) error {
|
|
||||||
var config InfluxDBv2DataRepositoryConfig
|
|
||||||
if err := json.Unmarshal(rawConfig, &config); err != nil {
|
|
||||||
return err
|
|
||||||
}
|
|
||||||
|
|
||||||
idb.client = influxdb2.NewClientWithOptions(config.Url, config.Token, influxdb2.DefaultOptions().SetTLSConfig(&tls.Config {InsecureSkipVerify: config.SkipTls,} ))
|
|
||||||
idb.queryClient = idb.client.QueryAPI(config.Org)
|
|
||||||
idb.bucket = config.Bucket
|
|
||||||
|
|
||||||
return nil
|
|
||||||
}
|
|
||||||
|
|
||||||
func (idb *InfluxDBv2DataRepository) formatTime(t time.Time) string {
|
|
||||||
return t.Format(time.RFC3339) // Like “2006-01-02T15:04:05Z07:00”
|
|
||||||
}
|
|
||||||
|
|
||||||
func (idb *InfluxDBv2DataRepository) epochToTime(epoch int64) time.Time {
|
|
||||||
return time.Unix(epoch, 0)
|
|
||||||
}
|
|
||||||
|
|
||||||
func (idb *InfluxDBv2DataRepository) LoadData(job *schema.Job, metrics []string, scopes []schema.MetricScope, ctx context.Context) (schema.JobData, error) {
|
|
||||||
|
|
||||||
measurementsConds := make([]string, 0, len(metrics))
|
|
||||||
for _, m := range metrics {
|
|
||||||
measurementsConds = append(measurementsConds, fmt.Sprintf(`r["_measurement"] == "%s"`, m))
|
|
||||||
}
|
|
||||||
measurementsCond := strings.Join(measurementsConds, " or ")
|
|
||||||
|
|
||||||
hostsConds := make([]string, 0, len(job.Resources))
|
|
||||||
for _, h := range job.Resources {
|
|
||||||
if h.HWThreads != nil || h.Accelerators != nil {
|
|
||||||
// TODO
|
|
||||||
return nil, errors.New("the InfluxDB metric data repository does not yet support HWThreads or Accelerators")
|
|
||||||
}
|
|
||||||
hostsConds = append(hostsConds, fmt.Sprintf(`r["hostname"] == "%s"`, h.Hostname))
|
|
||||||
}
|
|
||||||
hostsCond := strings.Join(hostsConds, " or ")
|
|
||||||
|
|
||||||
jobData := make(schema.JobData) // Empty Schema: map[<string>FIELD]map[<MetricScope>SCOPE]<*JobMetric>METRIC
|
|
||||||
// Requested Scopes
|
|
||||||
for _, scope := range scopes {
|
|
||||||
query := ""
|
|
||||||
switch scope {
|
|
||||||
case "node":
|
|
||||||
// Get Finest Granularity, Groupy By Measurement and Hostname (== Metric / Node), Calculate Mean for 60s windows
|
|
||||||
// log.Println("Note: Scope 'node' requested. ")
|
|
||||||
query = fmt.Sprintf(`
|
|
||||||
from(bucket: "%s")
|
|
||||||
|> range(start: %s, stop: %s)
|
|
||||||
|> filter(fn: (r) => (%s) and (%s) )
|
|
||||||
|> drop(columns: ["_start", "_stop"])
|
|
||||||
|> group(columns: ["hostname", "_measurement"])
|
|
||||||
|> aggregateWindow(every: 60s, fn: mean)
|
|
||||||
|> drop(columns: ["_time"])`,
|
|
||||||
idb.bucket,
|
|
||||||
idb.formatTime(job.StartTime), idb.formatTime(idb.epochToTime(job.StartTimeUnix + int64(job.Duration) + int64(1) )),
|
|
||||||
measurementsCond, hostsCond)
|
|
||||||
case "socket":
|
|
||||||
log.Println("Note: Scope 'socket' requested, but not yet supported: Will return 'node' scope only. ")
|
|
||||||
continue
|
|
||||||
case "core":
|
|
||||||
log.Println("Note: Scope 'core' requested, but not yet supported: Will return 'node' scope only. ")
|
|
||||||
continue
|
|
||||||
// Get Finest Granularity only, Set NULL to 0.0
|
|
||||||
// query = fmt.Sprintf(`
|
|
||||||
// from(bucket: "%s")
|
|
||||||
// |> range(start: %s, stop: %s)
|
|
||||||
// |> filter(fn: (r) => %s )
|
|
||||||
// |> filter(fn: (r) => %s )
|
|
||||||
// |> drop(columns: ["_start", "_stop", "cluster"])
|
|
||||||
// |> map(fn: (r) => (if exists r._value then {r with _value: r._value} else {r with _value: 0.0}))`,
|
|
||||||
// idb.bucket,
|
|
||||||
// idb.formatTime(job.StartTime), idb.formatTime(idb.epochToTime(job.StartTimeUnix + int64(job.Duration) + int64(1) )),
|
|
||||||
// measurementsCond, hostsCond)
|
|
||||||
default:
|
|
||||||
log.Println("Note: Unknown Scope requested: Will return 'node' scope. ")
|
|
||||||
continue
|
|
||||||
// return nil, errors.New("the InfluxDB metric data repository does not yet support other scopes than 'node'")
|
|
||||||
}
|
|
||||||
|
|
||||||
rows, err := idb.queryClient.Query(ctx, query)
|
|
||||||
if err != nil {
|
|
||||||
return nil, err
|
|
||||||
}
|
|
||||||
|
|
||||||
// Init Metrics: Only Node level now -> TODO: Matching /check on scope level ...
|
|
||||||
for _, metric := range metrics {
|
|
||||||
jobMetric, ok := jobData[metric]
|
|
||||||
if !ok {
|
|
||||||
mc := config.GetMetricConfig(job.Cluster, metric)
|
|
||||||
jobMetric = map[schema.MetricScope]*schema.JobMetric{
|
|
||||||
scope: { // uses scope var from above!
|
|
||||||
Unit: mc.Unit,
|
|
||||||
Scope: scope,
|
|
||||||
Timestep: mc.Timestep,
|
|
||||||
Series: make([]schema.Series, 0, len(job.Resources)),
|
|
||||||
StatisticsSeries: nil, // Should be: &schema.StatsSeries{},
|
|
||||||
},
|
|
||||||
}
|
|
||||||
}
|
|
||||||
jobData[metric] = jobMetric
|
|
||||||
}
|
|
||||||
|
|
||||||
// Process Result: Time-Data
|
|
||||||
field, host, hostSeries := "", "", schema.Series{}
|
|
||||||
// typeId := 0
|
|
||||||
switch scope {
|
|
||||||
case "node":
|
|
||||||
for rows.Next() {
|
|
||||||
row := rows.Record()
|
|
||||||
if ( host == "" || host != row.ValueByKey("hostname").(string) || rows.TableChanged() ) {
|
|
||||||
if ( host != "" ) {
|
|
||||||
// Append Series before reset
|
|
||||||
jobData[field][scope].Series = append(jobData[field][scope].Series, hostSeries)
|
|
||||||
}
|
|
||||||
field, host = row.Measurement(), row.ValueByKey("hostname").(string)
|
|
||||||
hostSeries = schema.Series{
|
|
||||||
Hostname: host,
|
|
||||||
Statistics: nil,
|
|
||||||
Data: make([]schema.Float, 0),
|
|
||||||
}
|
|
||||||
}
|
|
||||||
val, ok := row.Value().(float64)
|
|
||||||
if ok {
|
|
||||||
hostSeries.Data = append(hostSeries.Data, schema.Float(val))
|
|
||||||
} else {
|
|
||||||
hostSeries.Data = append(hostSeries.Data, schema.Float(0))
|
|
||||||
}
|
|
||||||
}
|
|
||||||
case "socket":
|
|
||||||
continue
|
|
||||||
case "core":
|
|
||||||
continue
|
|
||||||
// Include Series.Id in hostSeries
|
|
||||||
// for rows.Next() {
|
|
||||||
// row := rows.Record()
|
|
||||||
// if ( host == "" || host != row.ValueByKey("hostname").(string) || typeId != row.ValueByKey("type-id").(int) || rows.TableChanged() ) {
|
|
||||||
// if ( host != "" ) {
|
|
||||||
// // Append Series before reset
|
|
||||||
// jobData[field][scope].Series = append(jobData[field][scope].Series, hostSeries)
|
|
||||||
// }
|
|
||||||
// field, host, typeId = row.Measurement(), row.ValueByKey("hostname").(string), row.ValueByKey("type-id").(int)
|
|
||||||
// hostSeries = schema.Series{
|
|
||||||
// Hostname: host,
|
|
||||||
// Id: &typeId,
|
|
||||||
// Statistics: nil,
|
|
||||||
// Data: make([]schema.Float, 0),
|
|
||||||
// }
|
|
||||||
// }
|
|
||||||
// val := row.Value().(float64)
|
|
||||||
// hostSeries.Data = append(hostSeries.Data, schema.Float(val))
|
|
||||||
// }
|
|
||||||
default:
|
|
||||||
continue
|
|
||||||
// return nil, errors.New("the InfluxDB metric data repository does not yet support other scopes than 'node, core'")
|
|
||||||
}
|
|
||||||
// Append last Series
|
|
||||||
jobData[field][scope].Series = append(jobData[field][scope].Series, hostSeries)
|
|
||||||
}
|
|
||||||
|
|
||||||
// Get Stats
|
|
||||||
stats, err := idb.LoadStats(job, metrics, ctx)
|
|
||||||
if err != nil {
|
|
||||||
return nil, err
|
|
||||||
}
|
|
||||||
|
|
||||||
for _, scope := range scopes {
|
|
||||||
if scope == "node" { // No 'socket/core' support yet
|
|
||||||
for metric, nodes := range stats {
|
|
||||||
// log.Println(fmt.Sprintf("<< Add Stats for : Field %s >>", metric))
|
|
||||||
for node, stats := range nodes {
|
|
||||||
// log.Println(fmt.Sprintf("<< Add Stats for : Host %s : Min %.2f, Max %.2f, Avg %.2f >>", node, stats.Min, stats.Max, stats.Avg ))
|
|
||||||
for index, _ := range jobData[metric][scope].Series {
|
|
||||||
// log.Println(fmt.Sprintf("<< Try to add Stats to Series in Position %d >>", index))
|
|
||||||
if jobData[metric][scope].Series[index].Hostname == node {
|
|
||||||
// log.Println(fmt.Sprintf("<< Match for Series in Position %d : Host %s >>", index, jobData[metric][scope].Series[index].Hostname))
|
|
||||||
jobData[metric][scope].Series[index].Statistics = &schema.MetricStatistics{Avg: stats.Avg, Min: stats.Min, Max: stats.Max}
|
|
||||||
// log.Println(fmt.Sprintf("<< Result Inner: Min %.2f, Max %.2f, Avg %.2f >>", jobData[metric][scope].Series[index].Statistics.Min, jobData[metric][scope].Series[index].Statistics.Max, jobData[metric][scope].Series[index].Statistics.Avg))
|
|
||||||
}
|
|
||||||
}
|
|
||||||
}
|
|
||||||
}
|
|
||||||
}
|
|
||||||
}
|
|
||||||
|
|
||||||
// DEBUG:
|
|
||||||
// for _, scope := range scopes {
|
|
||||||
// for _, met := range metrics {
|
|
||||||
// for _, series := range jobData[met][scope].Series {
|
|
||||||
// log.Println(fmt.Sprintf("<< Result: %d data points for metric %s on %s with scope %s, Stats: Min %.2f, Max %.2f, Avg %.2f >>",
|
|
||||||
// len(series.Data), met, series.Hostname, scope,
|
|
||||||
// series.Statistics.Min, series.Statistics.Max, series.Statistics.Avg))
|
|
||||||
// }
|
|
||||||
// }
|
|
||||||
// }
|
|
||||||
|
|
||||||
return jobData, nil
|
|
||||||
}
|
|
||||||
|
|
||||||
func (idb *InfluxDBv2DataRepository) LoadStats(job *schema.Job, metrics []string, ctx context.Context) (map[string]map[string]schema.MetricStatistics, error) {
|
|
||||||
|
|
||||||
stats := map[string]map[string]schema.MetricStatistics{}
|
|
||||||
|
|
||||||
hostsConds := make([]string, 0, len(job.Resources))
|
|
||||||
for _, h := range job.Resources {
|
|
||||||
if h.HWThreads != nil || h.Accelerators != nil {
|
|
||||||
// TODO
|
|
||||||
return nil, errors.New("the InfluxDB metric data repository does not yet support HWThreads or Accelerators")
|
|
||||||
}
|
|
||||||
hostsConds = append(hostsConds, fmt.Sprintf(`r["hostname"] == "%s"`, h.Hostname))
|
|
||||||
}
|
|
||||||
hostsCond := strings.Join(hostsConds, " or ")
|
|
||||||
|
|
||||||
// lenMet := len(metrics)
|
|
||||||
|
|
||||||
for _, metric := range metrics {
|
|
||||||
// log.Println(fmt.Sprintf("<< You are here: %s (Index %d of %d metrics)", metric, index, lenMet))
|
|
||||||
|
|
||||||
query := fmt.Sprintf(`
|
|
||||||
data = from(bucket: "%s")
|
|
||||||
|> range(start: %s, stop: %s)
|
|
||||||
|> filter(fn: (r) => r._measurement == "%s" and r._field == "value" and (%s))
|
|
||||||
union(tables: [data |> mean(column: "_value") |> set(key: "_field", value: "avg"),
|
|
||||||
data |> min(column: "_value") |> set(key: "_field", value: "min"),
|
|
||||||
data |> max(column: "_value") |> set(key: "_field", value: "max")])
|
|
||||||
|> pivot(rowKey: ["hostname"], columnKey: ["_field"], valueColumn: "_value")
|
|
||||||
|> group()`,
|
|
||||||
idb.bucket,
|
|
||||||
idb.formatTime(job.StartTime), idb.formatTime(idb.epochToTime(job.StartTimeUnix + int64(job.Duration) + int64(1) )),
|
|
||||||
metric, hostsCond)
|
|
||||||
|
|
||||||
rows, err := idb.queryClient.Query(ctx, query)
|
|
||||||
if err != nil {
|
|
||||||
return nil, err
|
|
||||||
}
|
|
||||||
|
|
||||||
nodes := map[string]schema.MetricStatistics{}
|
|
||||||
for rows.Next() {
|
|
||||||
row := rows.Record()
|
|
||||||
host := row.ValueByKey("hostname").(string)
|
|
||||||
|
|
||||||
avg, avgok := row.ValueByKey("avg").(float64)
|
|
||||||
if !avgok {
|
|
||||||
// log.Println(fmt.Sprintf(">> Assertion error for metric %s, statistic AVG. Expected 'float64', got %v", metric, avg))
|
|
||||||
avg = 0.0
|
|
||||||
}
|
|
||||||
min, minok := row.ValueByKey("min").(float64)
|
|
||||||
if !minok {
|
|
||||||
// log.Println(fmt.Sprintf(">> Assertion error for metric %s, statistic MIN. Expected 'float64', got %v", metric, min))
|
|
||||||
min = 0.0
|
|
||||||
}
|
|
||||||
max, maxok := row.ValueByKey("max").(float64)
|
|
||||||
if !maxok {
|
|
||||||
// log.Println(fmt.Sprintf(">> Assertion error for metric %s, statistic MAX. Expected 'float64', got %v", metric, max))
|
|
||||||
max = 0.0
|
|
||||||
}
|
|
||||||
|
|
||||||
nodes[host] = schema.MetricStatistics{
|
|
||||||
Avg: avg,
|
|
||||||
Min: min,
|
|
||||||
Max: max,
|
|
||||||
}
|
|
||||||
}
|
|
||||||
stats[metric] = nodes
|
|
||||||
}
|
|
||||||
|
|
||||||
return stats, nil
|
|
||||||
}
|
|
||||||
|
|
||||||
func (idb *InfluxDBv2DataRepository) LoadNodeData(cluster string, metrics, nodes []string, scopes []schema.MetricScope, from, to time.Time, ctx context.Context) (map[string]map[string][]*schema.JobMetric, error) {
|
|
||||||
// TODO : Implement to be used in Analysis- und System/Node-View
|
|
||||||
log.Println(fmt.Sprintf("LoadNodeData unimplemented for InfluxDBv2DataRepository, Args: cluster %s, metrics %v, nodes %v, scopes %v", cluster, metrics, nodes, scopes))
|
|
||||||
|
|
||||||
return nil, errors.New("unimplemented for InfluxDBv2DataRepository")
|
|
||||||
}
|
|
Loading…
Reference in New Issue
Block a user