diff --git a/flow/connectors/bigquery/merge_statement_generator.go b/flow/connectors/bigquery/merge_statement_generator.go index 726ea37e27..7e6d94dc9b 100644 --- a/flow/connectors/bigquery/merge_statement_generator.go +++ b/flow/connectors/bigquery/merge_statement_generator.go @@ -185,7 +185,7 @@ func (m *mergeStmtGenerator) generateMergeStmts() []string { // TODO (kaushik): This is so that the statement size for individual merge statements // doesn't exceed the limit. We should make this configurable. const batchSize = 8 - partitions := utils.ArrayChunksGen(m.unchangedToastColumns, batchSize) + partitions := utils.ArrayChunks(m.unchangedToastColumns, batchSize) mergeStmts := make([]string, 0, len(partitions)) for _, partition := range partitions { diff --git a/flow/connectors/utils/array.go b/flow/connectors/utils/array.go index e1a94fbf06..487e8791ad 100644 --- a/flow/connectors/utils/array.go +++ b/flow/connectors/utils/array.go @@ -17,7 +17,7 @@ func ArrayMinus(first []string, second []string) []string { return result } -func ArrayChunksGen[T any](slice []T, size int) [][]T { +func ArrayChunks[T any](slice []T, size int) [][]T { var chunks [][]T for size < len(slice) {