e4730a16eb
-Scopes socket and core added as switch cases, only barebone |
||
---|---|---|
.github/workflows | ||
api | ||
auth | ||
config | ||
frontend@184a5079bd | ||
graph | ||
log | ||
metricdata | ||
repository | ||
schema | ||
templates | ||
test | ||
utils | ||
.env | ||
.gitignore | ||
.gitmodules | ||
api_test.go | ||
go.mod | ||
go.sum | ||
gqlgen.yml | ||
init-db.go | ||
LICENSE | ||
README.md | ||
routes.go | ||
runtimeSetup.go | ||
server.go |
ClusterCockpit with a Golang backend
Create your job-archive accoring to this specification. At least one cluster with a valid cluster.json
file is required. Having no jobs in the job-archive at all is fine. You may use the sample job-archive available for download in cc-docker/develop.
Run server
# The frontend is a submodule, so use `--recursive`
git clone --recursive git@github.com:ClusterCockpit/cc-backend.git
# Prepare frontend
cd ./cc-backend/frontend
yarn install
yarn build
cd ..
go get
go build
# The job-archive directory must be organised the same way as
# as for the regular ClusterCockpit.
ln -s <your-existing-job-archive> ./var/job-archive
# Create empty job.db (Will be initialized as SQLite3 database)
touch ./var/job.db
# EDIT THE .env FILE BEFORE YOU DEPLOY (Change the secrets)!
# If authentication is disabled, it can be empty.
vim ./.env
# This will first initialize the job.db database by traversing all
# `meta.json` files in the job-archive and add a new user. `--no-server` will cause the
# executable to stop once it has done that instead of starting a server.
./cc-backend --init-db --add-user <your-username>:admin:<your-password> --no-server
# Start a HTTP server (HTTPS can be enabled, the default port is 8080):
./cc-backend
# Show other options:
./cc-backend --help
In order to run this program as a deamon, look at utils/systemd/README.md where a systemd unit file and more explanation is provided.
Configuration
A config file in the JSON format can be provided using --config
to override the defaults. Look at the beginning of server.go
for the defaults and consequently the format of the configuration file.
Update GraphQL schema
This project uses gqlgen for the GraphQL API. The schema can be found in ./graph/schema.graphqls
. After changing it, you need to run go run github.com/99designs/gqlgen
which will update graph/model
. In case new resolvers are needed, they will be inserted into graph/schema.resolvers.go
, where you will need to implement them.
Project Structure
api/
contains the REST API. The routes defined there should be called whenever a job starts/stops. The API is documented in the OpenAPI 3.0 format in ./api/openapi.yaml.auth/
is where the (optional) authentication middleware can be found, which adds the currently authenticated user to the request context. Theuser
table is created and managed here as well.auth/ldap.go
contains everything to do with automatically syncing and authenticating users form an LDAP server.
config
handles thecluster.json
files and the user-specific configurations (changeable via GraphQL) for the Web-UI such as the selected metrics etc.frontend
is a submodule, this is where the Svelte based frontend resides.graph/generated
should not be touched.graph/model
contains all types defined in the GraphQL schema not manually defined inschema/
. Manually defined types have to be listed ingqlgen.yml
.graph/schema.graphqls
contains the GraphQL schema. Whenever you change it, you should callgo run github.com/99designs/gqlgen
.graph/
contains the resolvers and handlers for the GraphQL API. Function signatures ingraph/schema.resolvers.go
are automatically generated.metricdata/
handles getting and archiving the metrics associated with a job.metricdata/metricdata.go
defines the interfaceMetricDataRepository
and provides functions to the GraphQL and REST API for accessing a jobs metrics which automatically take care of selecting the source for the metrics (the archive or one of the metric data repositories).metricdata/archive.go
provides functions for fetching metrics from the job-archive and archiving a job to the job-archive.metricdata/cc-metric-store.go
contains an implementation of theMetricDataRepository
interface which can fetch data from an cc-metric-storemetricdata/influxdb-v2
contains an implementation of theMetricDataRepository
interface which can fetch data from an InfluxDBv2 database. It is currently disabled and out of date and can not be used as of writing.
schema/
contains type definitions used all over this project extracted in this package as Go disallows cyclic dependencies between packages.schema/float.go
contains a customfloat64
type which overwrites JSON and GraphQL Marshaling/Unmarshalling. This is needed because a regular optionalFloat
in GraphQL will map to*float64
types in Go. Wrapping every single metric value in an allocation would be a lot of overhead.schema/job.go
provides the types representing a job and its resources. Those can be used as type for ameta.json
file and/or a row in thejob
table.
templates/
is mostly full of HTML templates and a small helper go module.utils/systemd
describes how to deploy/install this as a systemd serviceutils/
is mostly outdated. Look at the cc-util repo for more up-to-date scripts..env
must be changed before you deploy this. It contains a Base64 encoded Ed25519 key-pair, the secret used for sessions and the password to the LDAP server if LDAP authentication is enabled.gqlgen.yml
configures the behaviour and generation of gqlgen.init-db.go
initializes thejob
(andtag
andjobtag
) table if the--init-db
flag is provided. Not only is the table created in the correct schema, but the job-archive is traversed as well.server.go
contains the main function and starts the actual http server.
TODO
- write (unit) tests
- fix
LoadNodeData
in cc-metric-store MetricDataRepository. Currently does not work for non-node scoped metrics because partition is unkown for a node - make tokens and sessions (currently based on cookies) expire after some configurable time
- when authenticating using a JWT, check if that user still exists
- fix InfluxDB MetricDataRepository (new or old line-protocol format? Support node-level metrics only?)
- documentation, comments in the code base
- write more TODOs
- use more prepared statements and sqrl instead of squirrel