From b1a63bb98d5ec41d7e43115b00f3d3f04d4ecdfc Mon Sep 17 00:00:00 2001 From: Gani Georgiev Date: Tue, 13 Dec 2022 09:07:50 +0200 Subject: [PATCH] minor code optimizations --- daos/record.go | 12 ++++++------ daos/record_expand.go | 2 +- 2 files changed, 7 insertions(+), 7 deletions(-) diff --git a/daos/record.go b/daos/record.go index 445cbf1e..439ee352 100644 --- a/daos/record.go +++ b/daos/record.go @@ -423,7 +423,7 @@ func (dao *Dao) cascadeRecordDelete(mainRecord *models.Record, refs map[*models. // trigger cascade for each 1000 rel items until there is none batchSize := 1000 for { - rows := []dbx.NullStringMap{} + rows := make([]dbx.NullStringMap, 0, batchSize) if err := query.Limit(int64(batchSize)).All(&rows); err != nil { return err } @@ -433,8 +433,8 @@ func (dao *Dao) cascadeRecordDelete(mainRecord *models.Record, refs map[*models. break } - perWorkers := 50 - workers := int(math.Ceil(float64(total) / float64(perWorkers))) + perWorker := 50 + workers := int(math.Ceil(float64(total) / float64(perWorker))) batchErr := func() error { ch := make(chan error) @@ -442,12 +442,12 @@ func (dao *Dao) cascadeRecordDelete(mainRecord *models.Record, refs map[*models. for i := 0; i < workers; i++ { var chunks []dbx.NullStringMap - if len(rows) <= perWorkers { + if len(rows) <= perWorker { chunks = rows rows = nil } else { - chunks = rows[:perWorkers] - rows = rows[perWorkers:] + chunks = rows[:perWorker] + rows = rows[perWorker:] } go func() { diff --git a/daos/record_expand.go b/daos/record_expand.go index b7e3c7e4..c796a7a9 100644 --- a/daos/record_expand.go +++ b/daos/record_expand.go @@ -176,7 +176,7 @@ func (dao *Dao) expandRecords(records []*models.Record, expandPath string, fetch } // reindex with the rel id - indexedRels := map[string]*models.Record{} + indexedRels := make(map[string]*models.Record, len(rels)) for _, rel := range rels { indexedRels[rel.GetId()] = rel }