mirror of
https://github.com/ClusterCockpit/cc-backend
synced 2026-03-30 20:47:31 +02:00
Run gofumpt
This commit is contained in:
@@ -1072,10 +1072,12 @@ func (r *Resolver) Query() generated.QueryResolver { return &queryResolver{r} }
|
|||||||
// SubCluster returns generated.SubClusterResolver implementation.
|
// SubCluster returns generated.SubClusterResolver implementation.
|
||||||
func (r *Resolver) SubCluster() generated.SubClusterResolver { return &subClusterResolver{r} }
|
func (r *Resolver) SubCluster() generated.SubClusterResolver { return &subClusterResolver{r} }
|
||||||
|
|
||||||
type clusterResolver struct{ *Resolver }
|
type (
|
||||||
type jobResolver struct{ *Resolver }
|
clusterResolver struct{ *Resolver }
|
||||||
type metricValueResolver struct{ *Resolver }
|
jobResolver struct{ *Resolver }
|
||||||
type mutationResolver struct{ *Resolver }
|
metricValueResolver struct{ *Resolver }
|
||||||
type nodeResolver struct{ *Resolver }
|
mutationResolver struct{ *Resolver }
|
||||||
type queryResolver struct{ *Resolver }
|
nodeResolver struct{ *Resolver }
|
||||||
type subClusterResolver struct{ *Resolver }
|
queryResolver struct{ *Resolver }
|
||||||
|
subClusterResolver struct{ *Resolver }
|
||||||
|
)
|
||||||
|
|||||||
@@ -236,4 +236,3 @@ func (ccms *CCMetricStore) buildNodeQueries(
|
|||||||
|
|
||||||
return queries, assignedScope, nil
|
return queries, assignedScope, nil
|
||||||
}
|
}
|
||||||
|
|
||||||
|
|||||||
@@ -63,10 +63,10 @@ func DefaultConfig() *RepositoryConfig {
|
|||||||
MaxIdleConnections: 4,
|
MaxIdleConnections: 4,
|
||||||
ConnectionMaxLifetime: time.Hour,
|
ConnectionMaxLifetime: time.Hour,
|
||||||
ConnectionMaxIdleTime: 10 * time.Minute,
|
ConnectionMaxIdleTime: 10 * time.Minute,
|
||||||
MinRunningJobDuration: 600, // 10 minutes
|
MinRunningJobDuration: 600, // 10 minutes
|
||||||
DbCacheSizeMB: 2048, // 2GB per connection
|
DbCacheSizeMB: 2048, // 2GB per connection
|
||||||
DbSoftHeapLimitMB: 16384, // 16GB process-wide
|
DbSoftHeapLimitMB: 16384, // 16GB process-wide
|
||||||
BusyTimeoutMs: 60000, // 60 seconds
|
BusyTimeoutMs: 60000, // 60 seconds
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
|
|||||||
@@ -133,10 +133,10 @@ func (pt *prefixedTarget) WriteFile(name string, data []byte) error {
|
|||||||
// ClusterAwareParquetWriter organizes Parquet output by cluster.
|
// ClusterAwareParquetWriter organizes Parquet output by cluster.
|
||||||
// Each cluster gets its own subdirectory with a cluster.json config file.
|
// Each cluster gets its own subdirectory with a cluster.json config file.
|
||||||
type ClusterAwareParquetWriter struct {
|
type ClusterAwareParquetWriter struct {
|
||||||
target ParquetTarget
|
target ParquetTarget
|
||||||
maxSizeMB int
|
maxSizeMB int
|
||||||
writers map[string]*ParquetWriter
|
writers map[string]*ParquetWriter
|
||||||
clusterCfgs map[string]*schema.Cluster
|
clusterCfgs map[string]*schema.Cluster
|
||||||
}
|
}
|
||||||
|
|
||||||
// NewClusterAwareParquetWriter creates a writer that routes jobs to per-cluster ParquetWriters.
|
// NewClusterAwareParquetWriter creates a writer that routes jobs to per-cluster ParquetWriters.
|
||||||
|
|||||||
@@ -100,7 +100,6 @@ func Checkpointing(wg *sync.WaitGroup, ctx context.Context) {
|
|||||||
ms := GetMemoryStore()
|
ms := GetMemoryStore()
|
||||||
|
|
||||||
wg.Go(func() {
|
wg.Go(func() {
|
||||||
|
|
||||||
d := 12 * time.Hour // default checkpoint interval
|
d := 12 * time.Hour // default checkpoint interval
|
||||||
if Keys.CheckpointInterval != "" {
|
if Keys.CheckpointInterval != "" {
|
||||||
parsed, err := time.ParseDuration(Keys.CheckpointInterval)
|
parsed, err := time.ParseDuration(Keys.CheckpointInterval)
|
||||||
|
|||||||
Reference in New Issue
Block a user