feat: configurable concurrency for `helmfile test` (#442)
`helmfile test --concurency N` to set a concurrency number. It is automatically ceiled at the number of releases just to reduce wasting computing resources. Also, I've refactored the scatter-gather logic scattered across the code-base. Resolves #433
This commit is contained in:
parent
b9a097ed14
commit
571f351a8f
8
main.go
8
main.go
|
|
@ -496,18 +496,24 @@ Do you really want to delete?
|
|||
Value: 300,
|
||||
Usage: "maximum time for tests to run before being considered failed",
|
||||
},
|
||||
cli.IntFlag{
|
||||
Name: "concurrency",
|
||||
Value: 0,
|
||||
Usage: "maximum number of concurrent helm processes to run, 0 is unlimited",
|
||||
},
|
||||
},
|
||||
Action: func(c *cli.Context) error {
|
||||
return findAndIterateOverDesiredStatesUsingFlags(c, func(state *state.HelmState, helm helmexec.Interface, _ context) []error {
|
||||
cleanup := c.Bool("cleanup")
|
||||
timeout := c.Int("timeout")
|
||||
concurrency := c.Int("concurrency")
|
||||
|
||||
args := args.GetArgs(c.String("args"), state)
|
||||
if len(args) > 0 {
|
||||
helm.SetExtraArgs(args...)
|
||||
}
|
||||
|
||||
return state.TestReleases(helm, cleanup, timeout)
|
||||
return state.TestReleases(helm, cleanup, timeout, concurrency)
|
||||
})
|
||||
},
|
||||
},
|
||||
|
|
|
|||
389
state/state.go
389
state/state.go
|
|
@ -3,6 +3,7 @@ package state
|
|||
import (
|
||||
"errors"
|
||||
"fmt"
|
||||
"github.com/roboll/helmfile/helmexec"
|
||||
"io/ioutil"
|
||||
"os"
|
||||
"path"
|
||||
|
|
@ -10,9 +11,6 @@ import (
|
|||
"sort"
|
||||
"strconv"
|
||||
"strings"
|
||||
"sync"
|
||||
|
||||
"github.com/roboll/helmfile/helmexec"
|
||||
|
||||
"regexp"
|
||||
|
||||
|
|
@ -188,16 +186,19 @@ func (st *HelmState) prepareSyncReleases(helm helmexec.Interface, additionalValu
|
|||
jobs := make(chan *ReleaseSpec, numReleases)
|
||||
results := make(chan syncPrepareResult, numReleases)
|
||||
|
||||
if concurrency < 1 {
|
||||
concurrency = numReleases
|
||||
}
|
||||
res := []syncPrepareResult{}
|
||||
errs := []error{}
|
||||
|
||||
// WaitGroup is required to wait until goroutine per job in job queue cleanly stops.
|
||||
var waitGroup sync.WaitGroup
|
||||
waitGroup.Add(concurrency)
|
||||
|
||||
for w := 1; w <= concurrency; w++ {
|
||||
go func() {
|
||||
st.scatterGather(
|
||||
concurrency,
|
||||
numReleases,
|
||||
func() {
|
||||
for i := 0; i < numReleases; i++ {
|
||||
jobs <- &releases[i]
|
||||
}
|
||||
close(jobs)
|
||||
},
|
||||
func(_ int) {
|
||||
for release := range jobs {
|
||||
st.applyDefaultsTo(release)
|
||||
|
||||
|
|
@ -227,29 +228,20 @@ func (st *HelmState) prepareSyncReleases(helm helmexec.Interface, additionalValu
|
|||
|
||||
results <- syncPrepareResult{release: release, flags: flags, errors: []*ReleaseError{}}
|
||||
}
|
||||
waitGroup.Done()
|
||||
}()
|
||||
}
|
||||
|
||||
for i := 0; i < numReleases; i++ {
|
||||
jobs <- &releases[i]
|
||||
}
|
||||
close(jobs)
|
||||
|
||||
res := []syncPrepareResult{}
|
||||
errs := []error{}
|
||||
for i := 0; i < numReleases; {
|
||||
select {
|
||||
case r := <-results:
|
||||
for _, e := range r.errors {
|
||||
errs = append(errs, e)
|
||||
},
|
||||
func() {
|
||||
for i := 0; i < numReleases; {
|
||||
select {
|
||||
case r := <-results:
|
||||
for _, e := range r.errors {
|
||||
errs = append(errs, e)
|
||||
}
|
||||
res = append(res, r)
|
||||
i++
|
||||
}
|
||||
}
|
||||
res = append(res, r)
|
||||
i++
|
||||
}
|
||||
}
|
||||
|
||||
waitGroup.Wait()
|
||||
},
|
||||
)
|
||||
|
||||
return res, errs
|
||||
}
|
||||
|
|
@ -286,21 +278,20 @@ func (st *HelmState) SyncReleases(helm helmexec.Interface, additionalValues []st
|
|||
return prepErrs
|
||||
}
|
||||
|
||||
errs := []error{}
|
||||
jobQueue := make(chan *syncPrepareResult, len(preps))
|
||||
results := make(chan syncResult, len(preps))
|
||||
|
||||
if workerLimit < 1 {
|
||||
workerLimit = len(preps)
|
||||
}
|
||||
|
||||
// WaitGroup is required to wait until goroutine per job in job queue cleanly stops.
|
||||
// Otherwise, cleanup hooks won't run fully.
|
||||
// See #363 for more context.
|
||||
var waitGroup sync.WaitGroup
|
||||
waitGroup.Add(workerLimit)
|
||||
|
||||
for w := 1; w <= workerLimit; w++ {
|
||||
go func() {
|
||||
st.scatterGather(
|
||||
workerLimit,
|
||||
len(preps),
|
||||
func() {
|
||||
for i := 0; i < len(preps); i++ {
|
||||
jobQueue <- &preps[i]
|
||||
}
|
||||
close(jobQueue)
|
||||
},
|
||||
func(_ int) {
|
||||
for prep := range jobQueue {
|
||||
release := prep.release
|
||||
flags := prep.flags
|
||||
|
|
@ -323,29 +314,21 @@ func (st *HelmState) SyncReleases(helm helmexec.Interface, additionalValues []st
|
|||
st.logger.Warnf("warn: %v\n", err)
|
||||
}
|
||||
}
|
||||
waitGroup.Done()
|
||||
}()
|
||||
}
|
||||
|
||||
for i := 0; i < len(preps); i++ {
|
||||
jobQueue <- &preps[i]
|
||||
}
|
||||
close(jobQueue)
|
||||
|
||||
errs := []error{}
|
||||
for i := 0; i < len(preps); {
|
||||
select {
|
||||
case res := <-results:
|
||||
if len(res.errors) > 0 {
|
||||
for _, e := range res.errors {
|
||||
errs = append(errs, e)
|
||||
},
|
||||
func() {
|
||||
for i := 0; i < len(preps); {
|
||||
select {
|
||||
case res := <-results:
|
||||
if len(res.errors) > 0 {
|
||||
for _, e := range res.errors {
|
||||
errs = append(errs, e)
|
||||
}
|
||||
}
|
||||
}
|
||||
i++
|
||||
}
|
||||
}
|
||||
i++
|
||||
}
|
||||
|
||||
waitGroup.Wait()
|
||||
},
|
||||
)
|
||||
|
||||
if len(errs) > 0 {
|
||||
return errs
|
||||
|
|
@ -355,7 +338,7 @@ func (st *HelmState) SyncReleases(helm helmexec.Interface, additionalValues []st
|
|||
}
|
||||
|
||||
// downloadCharts will download and untar charts for Lint and Template
|
||||
func (st *HelmState) downloadCharts(helm helmexec.Interface, dir string, workerLimit int, helmfileCommand string) (map[string]string, []error) {
|
||||
func (st *HelmState) downloadCharts(helm helmexec.Interface, dir string, concurrency int, helmfileCommand string) (map[string]string, []error) {
|
||||
temp := make(map[string]string, len(st.Releases))
|
||||
type downloadResults struct {
|
||||
releaseName string
|
||||
|
|
@ -363,17 +346,19 @@ func (st *HelmState) downloadCharts(helm helmexec.Interface, dir string, workerL
|
|||
}
|
||||
errs := []error{}
|
||||
|
||||
var wgFetch sync.WaitGroup
|
||||
jobQueue := make(chan *ReleaseSpec, len(st.Releases))
|
||||
results := make(chan *downloadResults, len(st.Releases))
|
||||
wgFetch.Add(len(st.Releases))
|
||||
|
||||
if workerLimit < 1 {
|
||||
workerLimit = len(st.Releases)
|
||||
}
|
||||
|
||||
for w := 1; w <= workerLimit; w++ {
|
||||
go func() {
|
||||
st.scatterGather(
|
||||
concurrency,
|
||||
len(st.Releases),
|
||||
func() {
|
||||
for i := 0; i < len(st.Releases); i++ {
|
||||
jobQueue <- &st.Releases[i]
|
||||
}
|
||||
close(jobQueue)
|
||||
},
|
||||
func(_ int) {
|
||||
for release := range jobQueue {
|
||||
chartPath := ""
|
||||
if pathExists(normalizeChart(st.basePath, release.Chart)) {
|
||||
|
|
@ -403,20 +388,14 @@ func (st *HelmState) downloadCharts(helm helmexec.Interface, dir string, workerL
|
|||
|
||||
results <- &downloadResults{release.Name, chartPath}
|
||||
}
|
||||
wgFetch.Done()
|
||||
}()
|
||||
}
|
||||
for i := 0; i < len(st.Releases); i++ {
|
||||
jobQueue <- &st.Releases[i]
|
||||
}
|
||||
close(jobQueue)
|
||||
|
||||
for i := 0; i < len(st.Releases); i++ {
|
||||
downloadRes := <-results
|
||||
temp[downloadRes.releaseName] = downloadRes.chartPath
|
||||
}
|
||||
|
||||
wgFetch.Wait()
|
||||
},
|
||||
func() {
|
||||
for i := 0; i < len(st.Releases); i++ {
|
||||
downloadRes := <-results
|
||||
temp[downloadRes.releaseName] = downloadRes.chartPath
|
||||
}
|
||||
},
|
||||
)
|
||||
|
||||
if len(errs) > 0 {
|
||||
return nil, errs
|
||||
|
|
@ -568,16 +547,19 @@ func (st *HelmState) prepareDiffReleases(helm helmexec.Interface, additionalValu
|
|||
jobs := make(chan *ReleaseSpec, numReleases)
|
||||
results := make(chan diffPrepareResult, numReleases)
|
||||
|
||||
if concurrency < 1 {
|
||||
concurrency = numReleases
|
||||
}
|
||||
rs := []diffPrepareResult{}
|
||||
errs := []error{}
|
||||
|
||||
// WaitGroup is required to wait until goroutine per job in job queue cleanly stops.
|
||||
var waitGroup sync.WaitGroup
|
||||
waitGroup.Add(concurrency)
|
||||
|
||||
for w := 1; w <= concurrency; w++ {
|
||||
go func() {
|
||||
st.scatterGather(
|
||||
concurrency,
|
||||
numReleases,
|
||||
func() {
|
||||
for i := 0; i < numReleases; i++ {
|
||||
jobs <- &releases[i]
|
||||
}
|
||||
close(jobs)
|
||||
},
|
||||
func(_ int) {
|
||||
for release := range jobs {
|
||||
errs := []error{}
|
||||
|
||||
|
|
@ -618,32 +600,20 @@ func (st *HelmState) prepareDiffReleases(helm helmexec.Interface, additionalValu
|
|||
results <- diffPrepareResult{release: release, flags: flags, errors: []*ReleaseError{}}
|
||||
}
|
||||
}
|
||||
waitGroup.Done()
|
||||
}()
|
||||
}
|
||||
|
||||
for i := 0; i < numReleases; i++ {
|
||||
jobs <- &releases[i]
|
||||
}
|
||||
close(jobs)
|
||||
|
||||
rs := []diffPrepareResult{}
|
||||
errs := []error{}
|
||||
for i := 0; i < numReleases; {
|
||||
select {
|
||||
case res := <-results:
|
||||
if res.errors != nil && len(res.errors) > 0 {
|
||||
for _, e := range res.errors {
|
||||
errs = append(errs, e)
|
||||
},
|
||||
func() {
|
||||
for i := 0; i < numReleases; i++ {
|
||||
res := <-results
|
||||
if res.errors != nil && len(res.errors) > 0 {
|
||||
for _, e := range res.errors {
|
||||
errs = append(errs, e)
|
||||
}
|
||||
} else if res.release != nil {
|
||||
rs = append(rs, res)
|
||||
}
|
||||
} else if res.release != nil {
|
||||
rs = append(rs, res)
|
||||
}
|
||||
}
|
||||
i++
|
||||
}
|
||||
|
||||
waitGroup.Wait()
|
||||
},
|
||||
)
|
||||
|
||||
return rs, errs
|
||||
}
|
||||
|
|
@ -659,18 +629,19 @@ func (st *HelmState) DiffReleases(helm helmexec.Interface, additionalValues []st
|
|||
jobQueue := make(chan *diffPrepareResult, len(preps))
|
||||
results := make(chan diffResult, len(preps))
|
||||
|
||||
if workerLimit < 1 {
|
||||
workerLimit = len(preps)
|
||||
}
|
||||
rs := []*ReleaseSpec{}
|
||||
errs := []error{}
|
||||
|
||||
// WaitGroup is required to wait until goroutine per job in job queue cleanly stops.
|
||||
// Otherwise, cleanup hooks won't run fully.
|
||||
// See #363 for more context.
|
||||
var waitGroup sync.WaitGroup
|
||||
waitGroup.Add(workerLimit)
|
||||
|
||||
for w := 1; w <= workerLimit; w++ {
|
||||
go func() {
|
||||
st.scatterGather(
|
||||
workerLimit,
|
||||
len(preps),
|
||||
func() {
|
||||
for i := 0; i < len(preps); i++ {
|
||||
jobQueue <- &preps[i]
|
||||
}
|
||||
close(jobQueue)
|
||||
},
|
||||
func(_ int) {
|
||||
for prep := range jobQueue {
|
||||
flags := prep.flags
|
||||
release := prep.release
|
||||
|
|
@ -694,140 +665,50 @@ func (st *HelmState) DiffReleases(helm helmexec.Interface, additionalValues []st
|
|||
}
|
||||
}
|
||||
}
|
||||
waitGroup.Done()
|
||||
}()
|
||||
}
|
||||
|
||||
for i := 0; i < len(preps); i++ {
|
||||
jobQueue <- &preps[i]
|
||||
}
|
||||
close(jobQueue)
|
||||
|
||||
rs := []*ReleaseSpec{}
|
||||
errs := []error{}
|
||||
for i := 0; i < len(preps); {
|
||||
select {
|
||||
case res := <-results:
|
||||
if res.err != nil {
|
||||
errs = append(errs, res.err)
|
||||
if res.err.Code == 2 {
|
||||
rs = append(rs, res.err.ReleaseSpec)
|
||||
},
|
||||
func() {
|
||||
for i := 0; i < len(preps); i++ {
|
||||
res := <-results
|
||||
if res.err != nil {
|
||||
errs = append(errs, res.err)
|
||||
if res.err.Code == 2 {
|
||||
rs = append(rs, res.err.ReleaseSpec)
|
||||
}
|
||||
}
|
||||
}
|
||||
i++
|
||||
}
|
||||
}
|
||||
close(results)
|
||||
|
||||
waitGroup.Wait()
|
||||
},
|
||||
)
|
||||
|
||||
return rs, errs
|
||||
}
|
||||
|
||||
func (st *HelmState) ReleaseStatuses(helm helmexec.Interface, workerLimit int) []error {
|
||||
var errs []error
|
||||
jobQueue := make(chan ReleaseSpec)
|
||||
doneQueue := make(chan bool)
|
||||
errQueue := make(chan error)
|
||||
|
||||
if workerLimit < 1 {
|
||||
workerLimit = len(st.Releases)
|
||||
}
|
||||
|
||||
// WaitGroup is required to wait until goroutine per job in job queue cleanly stops.
|
||||
var waitGroup sync.WaitGroup
|
||||
waitGroup.Add(workerLimit)
|
||||
|
||||
for w := 1; w <= workerLimit; w++ {
|
||||
go func() {
|
||||
for release := range jobQueue {
|
||||
if err := helm.ReleaseStatus(release.Name); err != nil {
|
||||
errQueue <- err
|
||||
}
|
||||
doneQueue <- true
|
||||
}
|
||||
waitGroup.Done()
|
||||
}()
|
||||
}
|
||||
|
||||
go func() {
|
||||
for _, release := range st.Releases {
|
||||
jobQueue <- release
|
||||
}
|
||||
close(jobQueue)
|
||||
}()
|
||||
|
||||
for i := 0; i < len(st.Releases); {
|
||||
select {
|
||||
case err := <-errQueue:
|
||||
errs = append(errs, err)
|
||||
case <-doneQueue:
|
||||
i++
|
||||
}
|
||||
}
|
||||
|
||||
waitGroup.Wait()
|
||||
|
||||
if len(errs) != 0 {
|
||||
return errs
|
||||
}
|
||||
|
||||
return nil
|
||||
return st.scatterGatherReleases(helm, workerLimit, func(release ReleaseSpec) error {
|
||||
return helm.ReleaseStatus(release.Name)
|
||||
})
|
||||
}
|
||||
|
||||
// DeleteReleases wrapper for executing helm delete on the releases
|
||||
func (st *HelmState) DeleteReleases(helm helmexec.Interface, purge bool) []error {
|
||||
var wg sync.WaitGroup
|
||||
errs := []error{}
|
||||
|
||||
for _, release := range st.Releases {
|
||||
wg.Add(1)
|
||||
go func(wg *sync.WaitGroup, release ReleaseSpec) {
|
||||
flags := []string{}
|
||||
if purge {
|
||||
flags = append(flags, "--purge")
|
||||
}
|
||||
if err := helm.DeleteRelease(release.Name, flags...); err != nil {
|
||||
errs = append(errs, err)
|
||||
}
|
||||
wg.Done()
|
||||
}(&wg, release)
|
||||
}
|
||||
wg.Wait()
|
||||
|
||||
if len(errs) != 0 {
|
||||
return errs
|
||||
}
|
||||
|
||||
return nil
|
||||
return st.scatterGatherReleases(helm, len(st.Releases), func(release ReleaseSpec) error {
|
||||
flags := []string{}
|
||||
if purge {
|
||||
flags = append(flags, "--purge")
|
||||
}
|
||||
return helm.DeleteRelease(release.Name, flags...)
|
||||
})
|
||||
}
|
||||
|
||||
// TestReleases wrapper for executing helm test on the releases
|
||||
func (st *HelmState) TestReleases(helm helmexec.Interface, cleanup bool, timeout int) []error {
|
||||
var wg sync.WaitGroup
|
||||
errs := []error{}
|
||||
|
||||
for _, release := range st.Releases {
|
||||
wg.Add(1)
|
||||
go func(wg *sync.WaitGroup, release ReleaseSpec) {
|
||||
flags := []string{}
|
||||
if cleanup {
|
||||
flags = append(flags, "--cleanup")
|
||||
}
|
||||
flags = append(flags, "--timeout", strconv.Itoa(timeout))
|
||||
if err := helm.TestRelease(release.Name, flags...); err != nil {
|
||||
errs = append(errs, err)
|
||||
}
|
||||
wg.Done()
|
||||
}(&wg, release)
|
||||
}
|
||||
wg.Wait()
|
||||
|
||||
if len(errs) != 0 {
|
||||
return errs
|
||||
}
|
||||
|
||||
return nil
|
||||
func (st *HelmState) TestReleases(helm helmexec.Interface, cleanup bool, timeout int, concurrency int) []error {
|
||||
return st.scatterGatherReleases(helm, concurrency, func(release ReleaseSpec) error {
|
||||
flags := []string{}
|
||||
if cleanup {
|
||||
flags = append(flags, "--cleanup")
|
||||
}
|
||||
flags = append(flags, "--timeout", strconv.Itoa(timeout))
|
||||
return helm.TestRelease(release.Name, flags...)
|
||||
})
|
||||
}
|
||||
|
||||
// Clean will remove any generated secrets
|
||||
|
|
|
|||
|
|
@ -0,0 +1,88 @@
|
|||
package state
|
||||
|
||||
import (
|
||||
"fmt"
|
||||
"github.com/roboll/helmfile/helmexec"
|
||||
"sync"
|
||||
)
|
||||
|
||||
type result struct {
|
||||
release ReleaseSpec
|
||||
err error
|
||||
}
|
||||
|
||||
func (st *HelmState) scatterGather(concurrency int, items int, produceInputs func(), receiveInputsAndProduceIntermediates func(int), aggregateIntermediates func()) {
|
||||
numReleases := len(st.Releases)
|
||||
if concurrency < 1 {
|
||||
concurrency = numReleases
|
||||
} else if concurrency > numReleases {
|
||||
concurrency = numReleases
|
||||
}
|
||||
|
||||
// WaitGroup is required to wait until goroutine per job in job queue cleanly stops.
|
||||
var waitGroup sync.WaitGroup
|
||||
waitGroup.Add(concurrency)
|
||||
|
||||
go produceInputs()
|
||||
|
||||
for w := 1; w <= concurrency; w++ {
|
||||
go func(id int) {
|
||||
st.logger.Debugf("worker %d/%d started", id, concurrency)
|
||||
receiveInputsAndProduceIntermediates(id)
|
||||
st.logger.Debugf("worker %d/%d finished", id, concurrency)
|
||||
waitGroup.Done()
|
||||
}(w)
|
||||
}
|
||||
|
||||
aggregateIntermediates()
|
||||
|
||||
// Wait until all the goroutines to gracefully finish
|
||||
waitGroup.Wait()
|
||||
}
|
||||
|
||||
func (st *HelmState) scatterGatherReleases(helm helmexec.Interface, concurrency int, do func(ReleaseSpec) error) []error {
|
||||
var errs []error
|
||||
|
||||
inputs := st.Releases
|
||||
inputsSize := len(inputs)
|
||||
|
||||
releases := make(chan ReleaseSpec)
|
||||
results := make(chan result)
|
||||
|
||||
st.scatterGather(
|
||||
concurrency,
|
||||
inputsSize,
|
||||
func() {
|
||||
for _, release := range inputs {
|
||||
releases <- release
|
||||
}
|
||||
close(releases)
|
||||
},
|
||||
func(id int) {
|
||||
for release := range releases {
|
||||
err := do(release)
|
||||
st.logger.Debugf("sending result for release: %s\n", release.Name)
|
||||
results <- result{release: release, err: err}
|
||||
st.logger.Debugf("sent result for release: %s\n", release.Name)
|
||||
}
|
||||
},
|
||||
func() {
|
||||
for i := range inputs {
|
||||
st.logger.Debugf("receiving result %d", i)
|
||||
r := <-results
|
||||
if r.err != nil {
|
||||
errs = append(errs, fmt.Errorf("release \"%s\" failed: %v", r.release.Name, r.err))
|
||||
} else {
|
||||
st.logger.Debugf("received result for release \"%s\"", r.release.Name)
|
||||
}
|
||||
st.logger.Debugf("received result for %d", i)
|
||||
}
|
||||
},
|
||||
)
|
||||
|
||||
if len(errs) != 0 {
|
||||
return errs
|
||||
}
|
||||
|
||||
return nil
|
||||
}
|
||||
|
|
@ -738,6 +738,7 @@ func TestHelmState_SyncReleases(t *testing.T) {
|
|||
t.Run(tt.name, func(t *testing.T) {
|
||||
state := &HelmState{
|
||||
Releases: tt.releases,
|
||||
logger: logger,
|
||||
}
|
||||
if _ = state.SyncReleases(tt.helm, []string{}, 1); !reflect.DeepEqual(tt.helm.releases, tt.wantReleases) {
|
||||
t.Errorf("HelmState.SyncReleases() for [%s] = %v, want %v", tt.name, tt.helm.releases, tt.wantReleases)
|
||||
|
|
@ -815,6 +816,7 @@ func TestHelmState_ReleaseStatuses(t *testing.T) {
|
|||
i := func(t *testing.T) {
|
||||
state := &HelmState{
|
||||
Releases: tt.releases,
|
||||
logger: logger,
|
||||
}
|
||||
errs := state.ReleaseStatuses(tt.helm, 1)
|
||||
if (errs != nil) != tt.wantErr {
|
||||
|
|
@ -874,8 +876,9 @@ func TestHelmState_TestReleasesNoCleanUp(t *testing.T) {
|
|||
i := func(t *testing.T) {
|
||||
state := &HelmState{
|
||||
Releases: tt.releases,
|
||||
logger: logger,
|
||||
}
|
||||
errs := state.TestReleases(tt.helm, tt.cleanup, 1)
|
||||
errs := state.TestReleases(tt.helm, tt.cleanup, 1, 1)
|
||||
if (errs != nil) != tt.wantErr {
|
||||
t.Errorf("TestReleases() for %s error = %v, wantErr %v", tt.name, errs, tt.wantErr)
|
||||
return
|
||||
|
|
|
|||
Loading…
Reference in New Issue