509 lines
16 KiB
Go
509 lines
16 KiB
Go
/*
|
|
Copyright 2020 The actions-runner-controller authors.
|
|
|
|
Licensed under the Apache License, Version 2.0 (the "License");
|
|
you may not use this file except in compliance with the License.
|
|
You may obtain a copy of the License at
|
|
|
|
http://www.apache.org/licenses/LICENSE-2.0
|
|
|
|
Unless required by applicable law or agreed to in writing, software
|
|
distributed under the License is distributed on an "AS IS" BASIS,
|
|
WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
|
|
See the License for the specific language governing permissions and
|
|
limitations under the License.
|
|
*/
|
|
|
|
package actionssummerwindnet
|
|
|
|
import (
|
|
"context"
|
|
"fmt"
|
|
"hash/fnv"
|
|
"reflect"
|
|
"sort"
|
|
"time"
|
|
|
|
"k8s.io/apimachinery/pkg/types"
|
|
|
|
"github.com/davecgh/go-spew/spew"
|
|
"github.com/go-logr/logr"
|
|
"k8s.io/apimachinery/pkg/runtime"
|
|
"k8s.io/apimachinery/pkg/util/rand"
|
|
"k8s.io/client-go/tools/record"
|
|
ctrl "sigs.k8s.io/controller-runtime"
|
|
"sigs.k8s.io/controller-runtime/pkg/client"
|
|
|
|
corev1 "k8s.io/api/core/v1"
|
|
metav1 "k8s.io/apimachinery/pkg/apis/meta/v1"
|
|
|
|
"github.com/actions/actions-runner-controller/apis/actions.summerwind.net/v1alpha1"
|
|
"github.com/actions/actions-runner-controller/controllers/actions.summerwind.net/metrics"
|
|
)
|
|
|
|
const (
|
|
LabelKeyRunnerTemplateHash = "runner-template-hash"
|
|
LabelKeyRunnerDeploymentName = "runner-deployment-name"
|
|
|
|
runnerSetOwnerKey = ".metadata.controller"
|
|
)
|
|
|
|
// RunnerDeploymentReconciler reconciles a Runner object
|
|
type RunnerDeploymentReconciler struct {
|
|
client.Client
|
|
Log logr.Logger
|
|
Recorder record.EventRecorder
|
|
Scheme *runtime.Scheme
|
|
CommonRunnerLabels []string
|
|
Name string
|
|
}
|
|
|
|
// +kubebuilder:rbac:groups=actions.summerwind.dev,resources=runnerdeployments,verbs=get;list;watch;create;update;patch;delete
|
|
// +kubebuilder:rbac:groups=actions.summerwind.dev,resources=runnerdeployments/finalizers,verbs=get;list;watch;create;update;patch;delete
|
|
// +kubebuilder:rbac:groups=actions.summerwind.dev,resources=runnerdeployments/status,verbs=get;update;patch
|
|
// +kubebuilder:rbac:groups=actions.summerwind.dev,resources=runnerreplicasets,verbs=get;list;watch;create;update;patch;delete
|
|
// +kubebuilder:rbac:groups=actions.summerwind.dev,resources=runnerreplicasets/status,verbs=get;update;patch
|
|
// +kubebuilder:rbac:groups=core,resources=events,verbs=create;patch
|
|
|
|
func (r *RunnerDeploymentReconciler) Reconcile(ctx context.Context, req ctrl.Request) (ctrl.Result, error) {
|
|
log := r.Log.WithValues("runnerdeployment", req.NamespacedName)
|
|
|
|
var rd v1alpha1.RunnerDeployment
|
|
if err := r.Get(ctx, req.NamespacedName, &rd); err != nil {
|
|
return ctrl.Result{}, client.IgnoreNotFound(err)
|
|
}
|
|
|
|
if !rd.ObjectMeta.DeletionTimestamp.IsZero() {
|
|
return ctrl.Result{}, nil
|
|
}
|
|
|
|
metrics.SetRunnerDeployment(rd)
|
|
|
|
var myRunnerReplicaSetList v1alpha1.RunnerReplicaSetList
|
|
if err := r.List(ctx, &myRunnerReplicaSetList, client.InNamespace(req.Namespace), client.MatchingFields{runnerSetOwnerKey: req.Name}); err != nil {
|
|
return ctrl.Result{}, err
|
|
}
|
|
|
|
myRunnerReplicaSets := myRunnerReplicaSetList.Items
|
|
|
|
sort.Slice(myRunnerReplicaSets, func(i, j int) bool {
|
|
return myRunnerReplicaSets[i].GetCreationTimestamp().After(myRunnerReplicaSets[j].GetCreationTimestamp().Time)
|
|
})
|
|
|
|
var newestSet *v1alpha1.RunnerReplicaSet
|
|
|
|
var oldSets []v1alpha1.RunnerReplicaSet
|
|
|
|
if len(myRunnerReplicaSets) > 0 {
|
|
newestSet = &myRunnerReplicaSets[0]
|
|
}
|
|
|
|
if len(myRunnerReplicaSets) > 1 {
|
|
oldSets = myRunnerReplicaSets[1:]
|
|
}
|
|
|
|
desiredRS, err := r.newRunnerReplicaSet(rd)
|
|
if err != nil {
|
|
r.Recorder.Event(&rd, corev1.EventTypeNormal, "RunnerAutoscalingFailure", err.Error())
|
|
|
|
log.Error(err, "Could not create runnerreplicaset")
|
|
|
|
return ctrl.Result{}, err
|
|
}
|
|
|
|
if newestSet == nil {
|
|
if err := r.Client.Create(ctx, desiredRS); err != nil {
|
|
log.Error(err, "Failed to create runnerreplicaset resource")
|
|
|
|
return ctrl.Result{}, err
|
|
}
|
|
|
|
log.Info("Created runnerreplicaset", "runnerreplicaset", desiredRS.Name)
|
|
|
|
return ctrl.Result{}, nil
|
|
}
|
|
|
|
newestTemplateHash, ok := getTemplateHash(newestSet)
|
|
if !ok {
|
|
log.Info("Failed to get template hash of newest runnerreplicaset resource. It must be in an invalid state. Please manually delete the runnerreplicaset so that it is recreated")
|
|
|
|
return ctrl.Result{}, nil
|
|
}
|
|
|
|
desiredTemplateHash, ok := getTemplateHash(desiredRS)
|
|
if !ok {
|
|
log.Info("Failed to get template hash of desired runnerreplicaset resource. It must be in an invalid state. Please manually delete the runnerreplicaset so that it is recreated")
|
|
|
|
return ctrl.Result{}, nil
|
|
}
|
|
|
|
if newestTemplateHash != desiredTemplateHash {
|
|
if err := r.Client.Create(ctx, desiredRS); err != nil {
|
|
log.Error(err, "Failed to create runnerreplicaset resource")
|
|
|
|
return ctrl.Result{}, err
|
|
}
|
|
|
|
log.Info("Created runnerreplicaset", "runnerreplicaset", desiredRS.Name)
|
|
|
|
// We requeue in order to clean up old runner replica sets later.
|
|
// Otherwise, they aren't cleaned up until the next re-sync interval.
|
|
return ctrl.Result{RequeueAfter: 5 * time.Second}, nil
|
|
}
|
|
|
|
if !reflect.DeepEqual(newestSet.Spec.Selector, desiredRS.Spec.Selector) {
|
|
updateSet := newestSet.DeepCopy()
|
|
updateSet.Spec = *desiredRS.Spec.DeepCopy()
|
|
|
|
// A selector update change doesn't trigger replicaset replacement,
|
|
// but we still need to update the existing replicaset with it.
|
|
// Otherwise selector-based runner query will never work on replicasets created before the controller v0.17.0
|
|
// See https://github.com/actions/actions-runner-controller/pull/355#discussion_r585379259
|
|
if err := r.Client.Update(ctx, updateSet); err != nil {
|
|
log.Error(err, "Failed to update runnerreplicaset resource")
|
|
|
|
return ctrl.Result{}, err
|
|
}
|
|
|
|
log.V(1).Info("Updated runnerreplicaset due to selector change")
|
|
|
|
// At this point, we are already sure that there's no need to create a new replicaset
|
|
// as the runner template hash is not changed.
|
|
//
|
|
// But we still need to requeue for the (possibly rare) cases that there are still old replicasets that needs
|
|
// to be cleaned up.
|
|
return ctrl.Result{RequeueAfter: 5 * time.Second}, nil
|
|
}
|
|
|
|
const defaultReplicas = 1
|
|
|
|
currentDesiredReplicas := getIntOrDefault(newestSet.Spec.Replicas, defaultReplicas)
|
|
newDesiredReplicas := getIntOrDefault(desiredRS.Spec.Replicas, defaultReplicas)
|
|
|
|
// Please add more conditions that we can in-place update the newest runnerreplicaset without disruption
|
|
//
|
|
// If we missed taking the EffectiveTime diff into account, you might end up experiencing scale-ups being delayed scale-down.
|
|
// See https://github.com/actions/actions-runner-controller/pull/1477#issuecomment-1164154496
|
|
var et1, et2 time.Time
|
|
if newestSet.Spec.EffectiveTime != nil {
|
|
et1 = newestSet.Spec.EffectiveTime.Time
|
|
}
|
|
if rd.Spec.EffectiveTime != nil {
|
|
et2 = rd.Spec.EffectiveTime.Time
|
|
}
|
|
if currentDesiredReplicas != newDesiredReplicas || et1 != et2 {
|
|
newestSet.Spec.Replicas = &newDesiredReplicas
|
|
newestSet.Spec.EffectiveTime = rd.Spec.EffectiveTime
|
|
|
|
if err := r.Client.Update(ctx, newestSet); err != nil {
|
|
log.Error(err, "Failed to update runnerreplicaset resource")
|
|
|
|
return ctrl.Result{}, err
|
|
}
|
|
|
|
log.V(1).Info("Updated runnerreplicaset due to spec change",
|
|
"currentDesiredReplicas", currentDesiredReplicas,
|
|
"newDesiredReplicas", newDesiredReplicas,
|
|
"currentEffectiveTime", newestSet.Spec.EffectiveTime,
|
|
"newEffectiveTime", rd.Spec.EffectiveTime,
|
|
)
|
|
|
|
return ctrl.Result{}, err
|
|
}
|
|
|
|
// Do we have old runner replica sets that should eventually deleted?
|
|
if len(oldSets) > 0 {
|
|
var readyReplicas int
|
|
if newestSet.Status.ReadyReplicas != nil {
|
|
readyReplicas = *newestSet.Status.ReadyReplicas
|
|
}
|
|
|
|
oldSetsCount := len(oldSets)
|
|
|
|
logWithDebugInfo := log.WithValues(
|
|
"newest_runnerreplicaset", types.NamespacedName{
|
|
Namespace: newestSet.Namespace,
|
|
Name: newestSet.Name,
|
|
},
|
|
"newest_runnerreplicaset_replicas_ready", readyReplicas,
|
|
"newest_runnerreplicaset_replicas_desired", currentDesiredReplicas,
|
|
"old_runnerreplicasets_count", oldSetsCount,
|
|
)
|
|
|
|
if readyReplicas < currentDesiredReplicas {
|
|
logWithDebugInfo.
|
|
Info("Waiting until the newest runnerreplicaset to be 100% available")
|
|
|
|
return ctrl.Result{}, nil
|
|
}
|
|
|
|
if oldSetsCount > 0 {
|
|
logWithDebugInfo.
|
|
Info("The newest runnerreplicaset is 100% available. Deleting old runnerreplicasets")
|
|
}
|
|
|
|
for i := range oldSets {
|
|
rs := oldSets[i]
|
|
|
|
rslog := log.WithValues("runnerreplicaset", rs.Name)
|
|
|
|
if rs.Status.Replicas != nil && *rs.Status.Replicas > 0 {
|
|
if rs.Spec.Replicas != nil && *rs.Spec.Replicas == 0 {
|
|
rslog.V(2).Info("Waiting for runnerreplicaset to scale to zero")
|
|
|
|
continue
|
|
}
|
|
|
|
updated := rs.DeepCopy()
|
|
zero := 0
|
|
updated.Spec.Replicas = &zero
|
|
if err := r.Client.Update(ctx, updated); err != nil {
|
|
rslog.Error(err, "Failed to scale runnerreplicaset to zero")
|
|
|
|
return ctrl.Result{}, err
|
|
}
|
|
|
|
rslog.Info("Scaled runnerreplicaset to zero")
|
|
|
|
continue
|
|
}
|
|
|
|
if err := r.Client.Delete(ctx, &rs); err != nil {
|
|
rslog.Error(err, "Failed to delete runnerreplicaset resource")
|
|
|
|
return ctrl.Result{}, err
|
|
}
|
|
|
|
r.Recorder.Event(&rd, corev1.EventTypeNormal, "RunnerReplicaSetDeleted", fmt.Sprintf("Deleted runnerreplicaset '%s'", rs.Name))
|
|
|
|
rslog.Info("Deleted runnerreplicaset")
|
|
}
|
|
}
|
|
|
|
var replicaSets []v1alpha1.RunnerReplicaSet
|
|
|
|
replicaSets = append(replicaSets, *newestSet)
|
|
replicaSets = append(replicaSets, oldSets...)
|
|
|
|
var totalCurrentReplicas, totalStatusAvailableReplicas, updatedReplicas int
|
|
|
|
for _, rs := range replicaSets {
|
|
var current, available int
|
|
|
|
if rs.Status.Replicas != nil {
|
|
current = *rs.Status.Replicas
|
|
}
|
|
|
|
if rs.Status.AvailableReplicas != nil {
|
|
available = *rs.Status.AvailableReplicas
|
|
}
|
|
|
|
totalCurrentReplicas += current
|
|
totalStatusAvailableReplicas += available
|
|
}
|
|
|
|
if newestSet.Status.Replicas != nil {
|
|
updatedReplicas = *newestSet.Status.Replicas
|
|
}
|
|
|
|
var status v1alpha1.RunnerDeploymentStatus
|
|
|
|
status.AvailableReplicas = &totalStatusAvailableReplicas
|
|
status.ReadyReplicas = &totalStatusAvailableReplicas
|
|
status.DesiredReplicas = &newDesiredReplicas
|
|
status.Replicas = &totalCurrentReplicas
|
|
status.UpdatedReplicas = &updatedReplicas
|
|
|
|
if !reflect.DeepEqual(rd.Status, status) {
|
|
updated := rd.DeepCopy()
|
|
updated.Status = status
|
|
|
|
if err := r.Status().Patch(ctx, updated, client.MergeFrom(&rd)); err != nil {
|
|
log.Info("Failed to patch runnerdeployment status. Retrying immediately", "error", err.Error())
|
|
return ctrl.Result{
|
|
Requeue: true,
|
|
}, nil
|
|
}
|
|
}
|
|
|
|
return ctrl.Result{}, nil
|
|
}
|
|
|
|
func getIntOrDefault(p *int, d int) int {
|
|
if p == nil {
|
|
return d
|
|
}
|
|
|
|
return *p
|
|
}
|
|
|
|
func getTemplateHash(rs *v1alpha1.RunnerReplicaSet) (string, bool) {
|
|
hash, ok := rs.Labels[LabelKeyRunnerTemplateHash]
|
|
|
|
return hash, ok
|
|
}
|
|
|
|
// ComputeHash returns a hash value calculated from pod template and
|
|
// a collisionCount to avoid hash collision. The hash will be safe encoded to
|
|
// avoid bad words.
|
|
//
|
|
// Proudly modified and adopted from k8s.io/kubernetes/pkg/util/hash.DeepHashObject and
|
|
// k8s.io/kubernetes/pkg/controller.ComputeHash.
|
|
func ComputeHash(template interface{}) string {
|
|
hasher := fnv.New32a()
|
|
|
|
hasher.Reset()
|
|
|
|
printer := spew.ConfigState{
|
|
Indent: " ",
|
|
SortKeys: true,
|
|
DisableMethods: true,
|
|
SpewKeys: true,
|
|
}
|
|
printer.Fprintf(hasher, "%#v", template)
|
|
|
|
return rand.SafeEncodeString(fmt.Sprint(hasher.Sum32()))
|
|
}
|
|
|
|
// Clones the given map and returns a new map with the given key and value added.
|
|
// Returns the given map, if labelKey is empty.
|
|
//
|
|
// Proudly copied from k8s.io/kubernetes/pkg/util/labels.CloneAndAddLabel
|
|
func CloneAndAddLabel(labels map[string]string, labelKey, labelValue string) map[string]string {
|
|
if labelKey == "" {
|
|
// Don't need to add a label.
|
|
return labels
|
|
}
|
|
// Clone.
|
|
newLabels := map[string]string{}
|
|
for key, value := range labels {
|
|
newLabels[key] = value
|
|
}
|
|
newLabels[labelKey] = labelValue
|
|
return newLabels
|
|
}
|
|
|
|
// Clones the given selector and returns a new selector with the given key and value added.
|
|
// Returns the given selector, if labelKey is empty.
|
|
//
|
|
// Proudly copied from k8s.io/kubernetes/pkg/util/labels.CloneSelectorAndAddLabel
|
|
func CloneSelectorAndAddLabel(selector *metav1.LabelSelector, labelKey, labelValue string) *metav1.LabelSelector {
|
|
if labelKey == "" {
|
|
// Don't need to add a label.
|
|
return selector
|
|
}
|
|
|
|
// Clone.
|
|
newSelector := new(metav1.LabelSelector)
|
|
|
|
newSelector.MatchLabels = make(map[string]string)
|
|
if selector.MatchLabels != nil {
|
|
for key, val := range selector.MatchLabels {
|
|
newSelector.MatchLabels[key] = val
|
|
}
|
|
}
|
|
newSelector.MatchLabels[labelKey] = labelValue
|
|
|
|
if selector.MatchExpressions != nil {
|
|
newMExps := make([]metav1.LabelSelectorRequirement, len(selector.MatchExpressions))
|
|
for i, me := range selector.MatchExpressions {
|
|
newMExps[i].Key = me.Key
|
|
newMExps[i].Operator = me.Operator
|
|
if me.Values != nil {
|
|
newMExps[i].Values = make([]string, len(me.Values))
|
|
copy(newMExps[i].Values, me.Values)
|
|
} else {
|
|
newMExps[i].Values = nil
|
|
}
|
|
}
|
|
newSelector.MatchExpressions = newMExps
|
|
} else {
|
|
newSelector.MatchExpressions = nil
|
|
}
|
|
|
|
return newSelector
|
|
}
|
|
|
|
func (r *RunnerDeploymentReconciler) newRunnerReplicaSet(rd v1alpha1.RunnerDeployment) (*v1alpha1.RunnerReplicaSet, error) {
|
|
return newRunnerReplicaSet(&rd, r.CommonRunnerLabels, r.Scheme)
|
|
}
|
|
|
|
func getSelector(rd *v1alpha1.RunnerDeployment) *metav1.LabelSelector {
|
|
selector := rd.Spec.Selector
|
|
if selector == nil {
|
|
selector = &metav1.LabelSelector{MatchLabels: map[string]string{LabelKeyRunnerDeploymentName: rd.Name}}
|
|
}
|
|
|
|
return selector
|
|
}
|
|
|
|
func newRunnerReplicaSet(rd *v1alpha1.RunnerDeployment, commonRunnerLabels []string, scheme *runtime.Scheme) (*v1alpha1.RunnerReplicaSet, error) {
|
|
newRSTemplate := *rd.Spec.Template.DeepCopy()
|
|
|
|
newRSTemplate.Spec.Labels = append(newRSTemplate.Spec.Labels, commonRunnerLabels...)
|
|
|
|
templateHash := ComputeHash(&newRSTemplate)
|
|
|
|
// Add template hash label to selector.
|
|
newRSTemplate.ObjectMeta.Labels = CloneAndAddLabel(newRSTemplate.ObjectMeta.Labels, LabelKeyRunnerTemplateHash, templateHash)
|
|
|
|
// This label selector is used by default when rd.Spec.Selector is empty.
|
|
newRSTemplate.ObjectMeta.Labels = CloneAndAddLabel(newRSTemplate.ObjectMeta.Labels, LabelKeyRunnerDeploymentName, rd.Name)
|
|
|
|
selector := getSelector(rd)
|
|
|
|
newRSSelector := CloneSelectorAndAddLabel(selector, LabelKeyRunnerTemplateHash, templateHash)
|
|
|
|
rs := v1alpha1.RunnerReplicaSet{
|
|
TypeMeta: metav1.TypeMeta{},
|
|
ObjectMeta: metav1.ObjectMeta{
|
|
GenerateName: rd.ObjectMeta.Name + "-",
|
|
Namespace: rd.ObjectMeta.Namespace,
|
|
Labels: newRSTemplate.ObjectMeta.Labels,
|
|
},
|
|
Spec: v1alpha1.RunnerReplicaSetSpec{
|
|
Replicas: rd.Spec.Replicas,
|
|
Selector: newRSSelector,
|
|
Template: newRSTemplate,
|
|
EffectiveTime: rd.Spec.EffectiveTime,
|
|
},
|
|
}
|
|
|
|
if err := ctrl.SetControllerReference(rd, &rs, scheme); err != nil {
|
|
return &rs, err
|
|
}
|
|
|
|
return &rs, nil
|
|
}
|
|
|
|
func (r *RunnerDeploymentReconciler) SetupWithManager(mgr ctrl.Manager) error {
|
|
name := "runnerdeployment-controller"
|
|
if r.Name != "" {
|
|
name = r.Name
|
|
}
|
|
|
|
r.Recorder = mgr.GetEventRecorderFor(name)
|
|
|
|
if err := mgr.GetFieldIndexer().IndexField(context.TODO(), &v1alpha1.RunnerReplicaSet{}, runnerSetOwnerKey, func(rawObj client.Object) []string {
|
|
runnerSet := rawObj.(*v1alpha1.RunnerReplicaSet)
|
|
owner := metav1.GetControllerOf(runnerSet)
|
|
if owner == nil {
|
|
return nil
|
|
}
|
|
|
|
if owner.APIVersion != v1alpha1.GroupVersion.String() || owner.Kind != "RunnerDeployment" {
|
|
return nil
|
|
}
|
|
|
|
return []string{owner.Name}
|
|
}); err != nil {
|
|
return err
|
|
}
|
|
|
|
return ctrl.NewControllerManagedBy(mgr).
|
|
For(&v1alpha1.RunnerDeployment{}).
|
|
Owns(&v1alpha1.RunnerReplicaSet{}).
|
|
Named(name).
|
|
Complete(r)
|
|
}
|