diff --git a/flow/activities/flowable.go b/flow/activities/flowable.go index b34ccf6f48..b077a06f04 100644 --- a/flow/activities/flowable.go +++ b/flow/activities/flowable.go @@ -25,6 +25,7 @@ import ( "github.com/PeerDB-io/peer-flow/connectors/utils" "github.com/PeerDB-io/peer-flow/connectors/utils/monitoring" "github.com/PeerDB-io/peer-flow/generated/protos" + "github.com/PeerDB-io/peer-flow/logger" "github.com/PeerDB-io/peer-flow/model" "github.com/PeerDB-io/peer-flow/peerdbenv" "github.com/PeerDB-io/peer-flow/shared" @@ -45,6 +46,21 @@ type SlotSnapshotSignal struct { type FlowableActivity struct { CatalogPool *pgxpool.Pool Alerter *alerting.Alerter + CdcCacheRw sync.RWMutex + CdcCache map[string]connectors.CDCPullConnector +} + +func (a *FlowableActivity) recordSlotSize( + ctx context.Context, + srcConn connectors.CDCPullConnector, + slotName string, + peerName string, +) { + logger := logger.LoggerFromCtx(ctx) + err := srcConn.HandleSlotInfo(ctx, a.Alerter, a.CatalogPool, slotName, peerName) + if err != nil { + logger.Error("failed to record slot size", "error", err) + } } func (a *FlowableActivity) CheckConnection( @@ -207,7 +223,76 @@ func (a *FlowableActivity) CreateNormalizedTable( }, nil } -func (a *FlowableActivity) StartFlow(ctx context.Context, +func (a *FlowableActivity) MaintainPull( + ctx context.Context, + config *protos.FlowConnectionConfigs, + sessionID string, +) error { + srcConn, err := connectors.GetCDCPullConnector(ctx, config.Source) + if err != nil { + return err + } + defer connectors.CloseConnector(ctx, srcConn) + + if err := srcConn.SetupReplConn(ctx); err != nil { + return err + } + + a.CdcCacheRw.Lock() + a.CdcCache[sessionID] = srcConn + a.CdcCacheRw.Unlock() + + ticker := time.NewTicker(15 * time.Second) + defer ticker.Stop() + + slotNameForMetrics := fmt.Sprintf("peerflow_slot_%s", config.FlowJobName) + if config.ReplicationSlotName != "" { + slotNameForMetrics = config.ReplicationSlotName + } + + go a.recordSlotSize(ctx, srcConn, slotNameForMetrics, config.Source.Name) + slotSizeTicker := time.NewTicker(5 * time.Minute) + defer slotSizeTicker.Stop() + + for { + select { + case <-ticker.C: + activity.RecordHeartbeat(ctx, "keep session alive") + if err := srcConn.ReplPing(ctx); err != nil { + activity.GetLogger(ctx).Error("Failed to send keep alive ping to replication connection", "Error", err) + } + case <-slotSizeTicker.C: + go a.recordSlotSize(ctx, srcConn, slotNameForMetrics, config.Source.Name) + case <-ctx.Done(): + a.CdcCacheRw.Lock() + delete(a.CdcCache, sessionID) + a.CdcCacheRw.Unlock() + return nil + } + } +} + +func (a *FlowableActivity) WaitForSourceConnector( + ctx context.Context, + sessionID string, +) error { + for { + a.CdcCacheRw.RLock() + _, ok := a.CdcCache[sessionID] + a.CdcCacheRw.RUnlock() + if ok { + return nil + } + activity.RecordHeartbeat(ctx, "wait another second for source connector") + if err := ctx.Err(); err != nil { + return err + } + time.Sleep(time.Second) + } +} + +func (a *FlowableActivity) StartFlow( + ctx context.Context, input *protos.StartFlowInput, ) (*model.SyncResponse, error) { ctx = context.WithValue(ctx, shared.FlowNameKey, input.FlowConnectionConfigs.FlowJobName) @@ -226,15 +311,14 @@ func (a *FlowableActivity) StartFlow(ctx context.Context, tblNameMapping[v.SourceTableIdentifier] = model.NewNameAndExclude(v.DestinationTableIdentifier, v.Exclude) } - srcConn, err := connectors.GetCDCPullConnector(ctx, config.Source) - if err != nil { - return nil, fmt.Errorf("failed to get source connector: %w", err) + a.CdcCacheRw.RLock() + srcConn, ok := a.CdcCache[input.SessionId] + a.CdcCacheRw.RUnlock() + if !ok { + return nil, errors.New("source connector missing from CdcCache") } - defer connectors.CloseConnector(ctx, srcConn) - - slotNameForMetrics := fmt.Sprintf("peerflow_slot_%s", input.FlowConnectionConfigs.FlowJobName) - if input.FlowConnectionConfigs.ReplicationSlotName != "" { - slotNameForMetrics = input.FlowConnectionConfigs.ReplicationSlotName + if err := srcConn.ConnectionActive(ctx); err != nil { + return nil, err } shutdown := utils.HeartbeatRoutine(ctx, func() string { @@ -243,23 +327,22 @@ func (a *FlowableActivity) StartFlow(ctx context.Context, }) defer shutdown() - errGroup, errCtx := errgroup.WithContext(ctx) - go a.recordSlotSizePeriodically(errCtx, srcConn, slotNameForMetrics, input.FlowConnectionConfigs.Source.Name) - batchSize := input.SyncFlowOptions.BatchSize if batchSize <= 0 { batchSize = 1_000_000 } - lastOffset, err := dstConn.GetLastOffset(ctx, input.FlowConnectionConfigs.FlowJobName) - if err != nil { - return nil, err - } - // start a goroutine to pull records from the source recordBatch := model.NewCDCRecordStream() startTime := time.Now() flowName := input.FlowConnectionConfigs.FlowJobName + + lastOffset, err := dstConn.GetLastOffset(ctx, flowName) + if err != nil { + return nil, err + } + + errGroup, errCtx := errgroup.WithContext(ctx) errGroup.Go(func() error { return srcConn.PullRecords(errCtx, a.CatalogPool, &model.PullRecordsRequest{ FlowJobName: flowName, @@ -352,11 +435,7 @@ func (a *FlowableActivity) StartFlow(ctx context.Context, logger.Info(fmt.Sprintf("pushed %d records in %d seconds", numRecords, int(syncDuration.Seconds()))) - lastCheckpoint, err := recordBatch.GetLastCheckpoint() - if err != nil { - a.Alerter.LogFlowError(ctx, flowName, err) - return nil, fmt.Errorf("failed to get last checkpoint: %w", err) - } + lastCheckpoint := pglogrepl.LSN(recordBatch.GetLastCheckpoint()) err = monitoring.UpdateNumRowsAndEndLSNForCDCBatch( ctx, @@ -364,7 +443,7 @@ func (a *FlowableActivity) StartFlow(ctx context.Context, input.FlowConnectionConfigs.FlowJobName, res.CurrentSyncBatchID, uint32(numRecords), - pglogrepl.LSN(lastCheckpoint), + lastCheckpoint, ) if err != nil { a.Alerter.LogFlowError(ctx, flowName, err) @@ -375,7 +454,7 @@ func (a *FlowableActivity) StartFlow(ctx context.Context, ctx, a.CatalogPool, input.FlowConnectionConfigs.FlowJobName, - pglogrepl.LSN(lastCheckpoint), + lastCheckpoint, ) if err != nil { a.Alerter.LogFlowError(ctx, flowName, err) diff --git a/flow/activities/slot.go b/flow/activities/slot.go deleted file mode 100644 index 8f6f3325b4..0000000000 --- a/flow/activities/slot.go +++ /dev/null @@ -1,36 +0,0 @@ -package activities - -import ( - "context" - "time" - - "github.com/PeerDB-io/peer-flow/connectors" -) - -func (a *FlowableActivity) recordSlotSizePeriodically( - ctx context.Context, - srcConn connectors.CDCPullConnector, - slotName string, - peerName string, -) { - // ensures slot info is logged at least once per SyncFlow - err := srcConn.HandleSlotInfo(ctx, a.Alerter, a.CatalogPool, slotName, peerName) - if err != nil { - return - } - - ticker := time.NewTicker(5 * time.Minute) - defer ticker.Stop() - - for { - select { - case <-ticker.C: - err := srcConn.HandleSlotInfo(ctx, a.Alerter, a.CatalogPool, slotName, peerName) - if err != nil { - return - } - case <-ctx.Done(): - return - } - } -} diff --git a/flow/cmd/worker.go b/flow/cmd/worker.go index c63511470d..4eac4b8939 100644 --- a/flow/cmd/worker.go +++ b/flow/cmd/worker.go @@ -16,6 +16,7 @@ import ( "go.temporal.io/sdk/worker" "github.com/PeerDB-io/peer-flow/activities" + "github.com/PeerDB-io/peer-flow/connectors" utils "github.com/PeerDB-io/peer-flow/connectors/utils/catalog" "github.com/PeerDB-io/peer-flow/logger" "github.com/PeerDB-io/peer-flow/shared" @@ -127,9 +128,10 @@ func WorkerMain(opts *WorkerOptions) error { return queueErr } - w := worker.New(c, taskQueue, worker.Options{}) + w := worker.New(c, taskQueue, worker.Options{ + EnableSessionWorker: true, + }) w.RegisterWorkflow(peerflow.CDCFlowWorkflowWithConfig) - w.RegisterWorkflow(peerflow.SyncFlowWorkflow) w.RegisterWorkflow(peerflow.SetupFlowWorkflow) w.RegisterWorkflow(peerflow.NormalizeFlowWorkflow) w.RegisterWorkflow(peerflow.QRepFlowWorkflow) @@ -146,6 +148,7 @@ func WorkerMain(opts *WorkerOptions) error { w.RegisterActivity(&activities.FlowableActivity{ CatalogPool: conn, Alerter: alerter, + CdcCache: make(map[string]connectors.CDCPullConnector), }) err = w.Run(worker.InterruptCh()) diff --git a/flow/connectors/bigquery/qrep_avro_sync.go b/flow/connectors/bigquery/qrep_avro_sync.go index 60b6400c75..d69642028c 100644 --- a/flow/connectors/bigquery/qrep_avro_sync.go +++ b/flow/connectors/bigquery/qrep_avro_sync.go @@ -74,10 +74,7 @@ func (s *QRepAvroSyncMethod) SyncRecords( insertStmt := fmt.Sprintf("INSERT INTO `%s` SELECT * FROM `%s`;", rawTableName, stagingTable) - lastCP, err := req.Records.GetLastCheckpoint() - if err != nil { - return nil, fmt.Errorf("failed to get last checkpoint: %w", err) - } + lastCP := req.Records.GetLastCheckpoint() activity.RecordHeartbeat(ctx, fmt.Sprintf("Flow job %s: performing insert and update transaction"+ diff --git a/flow/connectors/clickhouse/cdc.go b/flow/connectors/clickhouse/cdc.go index 24c9a8f0f5..cd96966d7b 100644 --- a/flow/connectors/clickhouse/cdc.go +++ b/flow/connectors/clickhouse/cdc.go @@ -107,13 +107,8 @@ func (c *ClickhouseConnector) syncRecordsViaAvro( return nil, fmt.Errorf("failed to sync schema changes: %w", err) } - lastCheckpoint, err := req.Records.GetLastCheckpoint() - if err != nil { - return nil, err - } - return &model.SyncResponse{ - LastSyncedCheckpointID: lastCheckpoint, + LastSyncedCheckpointID: req.Records.GetLastCheckpoint(), NumRecordsSynced: int64(numRecords), CurrentSyncBatchID: syncBatchID, TableNameRowsMapping: tableNameRowsMapping, @@ -130,12 +125,7 @@ func (c *ClickhouseConnector) SyncRecords(ctx context.Context, req *model.SyncRe return nil, err } - lastCheckpoint, err := req.Records.GetLastCheckpoint() - if err != nil { - return nil, fmt.Errorf("failed to get last checkpoint: %w", err) - } - - err = c.pgMetadata.FinishBatch(ctx, req.FlowJobName, req.SyncBatchID, lastCheckpoint) + err = c.pgMetadata.FinishBatch(ctx, req.FlowJobName, req.SyncBatchID, req.Records.GetLastCheckpoint()) if err != nil { c.logger.Error("failed to increment id", slog.Any("error", err)) return nil, err diff --git a/flow/connectors/core.go b/flow/connectors/core.go index f83a1a7836..0c590de1de 100644 --- a/flow/connectors/core.go +++ b/flow/connectors/core.go @@ -39,6 +39,10 @@ type CDCPullConnector interface { *protos.EnsurePullabilityBatchOutput, error) // Methods related to retrieving and pushing records for this connector as a source and destination. + SetupReplConn(context.Context) error + + // Ping source to keep connection alive. Can be called concurrently with PullRecords; skips ping in that case. + ReplPing(context.Context) error // PullRecords pulls records from the source, and returns a RecordBatch. // This method should be idempotent, and should be able to be called multiple times with the same request. diff --git a/flow/connectors/eventhub/eventhub.go b/flow/connectors/eventhub/eventhub.go index 2641dc5ff2..763964ea81 100644 --- a/flow/connectors/eventhub/eventhub.go +++ b/flow/connectors/eventhub/eventhub.go @@ -208,11 +208,7 @@ func (c *EventHubConnector) SyncRecords(ctx context.Context, req *model.SyncReco return nil, err } - lastCheckpoint, err := req.Records.GetLastCheckpoint() - if err != nil { - c.logger.Error("failed to get last checkpoint", slog.Any("error", err)) - return nil, err - } + lastCheckpoint := req.Records.GetLastCheckpoint() err = c.pgMetadata.FinishBatch(ctx, req.FlowJobName, req.SyncBatchID, lastCheckpoint) if err != nil { diff --git a/flow/connectors/postgres/cdc.go b/flow/connectors/postgres/cdc.go index 76de074bbd..7fb9e31624 100644 --- a/flow/connectors/postgres/cdc.go +++ b/flow/connectors/postgres/cdc.go @@ -27,7 +27,6 @@ import ( type PostgresCDCSource struct { *PostgresConnector - replConn *pgx.Conn SrcTableIDNameMapping map[uint32]string TableNameMapping map[string]model.NameAndExclude slot string @@ -48,6 +47,7 @@ type PostgresCDCConfig struct { Connection *pgx.Conn Slot string Publication string + ChildToParentRelIDMap map[uint32]uint32 SrcTableIDNameMapping map[uint32]string TableNameMapping map[string]model.NameAndExclude RelationMessageMapping model.RelationMessageMapping @@ -62,36 +62,28 @@ type startReplicationOpts struct { } // Create a new PostgresCDCSource -func (c *PostgresConnector) NewPostgresCDCSource(ctx context.Context, cdcConfig *PostgresCDCConfig) (*PostgresCDCSource, error) { - childToParentRelIDMap, err := getChildToParentRelIDMap(ctx, cdcConfig.Connection) - if err != nil { - return nil, fmt.Errorf("error getting child to parent relid map: %w", err) - } - +func (c *PostgresConnector) NewPostgresCDCSource(cdcConfig *PostgresCDCConfig) (*PostgresCDCSource, error) { return &PostgresCDCSource{ PostgresConnector: c, - replConn: cdcConfig.Connection, SrcTableIDNameMapping: cdcConfig.SrcTableIDNameMapping, TableNameMapping: cdcConfig.TableNameMapping, slot: cdcConfig.Slot, publication: cdcConfig.Publication, relationMessageMapping: cdcConfig.RelationMessageMapping, typeMap: pgtype.NewMap(), - childToParentRelIDMapping: childToParentRelIDMap, + childToParentRelIDMapping: cdcConfig.ChildToParentRelIDMap, commitLock: false, catalogPool: cdcConfig.CatalogPool, flowJobName: cdcConfig.FlowJobName, }, nil } -func getChildToParentRelIDMap(ctx context.Context, conn *pgx.Conn) (map[uint32]uint32, error) { +func GetChildToParentRelIDMap(ctx context.Context, conn *pgx.Conn) (map[uint32]uint32, error) { query := ` - SELECT - parent.oid AS parentrelid, - child.oid AS childrelid + SELECT parent.oid AS parentrelid, child.oid AS childrelid FROM pg_inherits - JOIN pg_class parent ON pg_inherits.inhparent = parent.oid - JOIN pg_class child ON pg_inherits.inhrelid = child.oid + JOIN pg_class parent ON pg_inherits.inhparent = parent.oid + JOIN pg_class child ON pg_inherits.inhrelid = child.oid WHERE parent.relkind='p'; ` @@ -99,7 +91,6 @@ func getChildToParentRelIDMap(ctx context.Context, conn *pgx.Conn) (map[uint32]u if err != nil { return nil, fmt.Errorf("error querying for child to parent relid map: %w", err) } - defer rows.Close() childToParentRelIDMap := make(map[uint32]uint32) @@ -118,83 +109,14 @@ func getChildToParentRelIDMap(ctx context.Context, conn *pgx.Conn) (map[uint32]u // PullRecords pulls records from the cdc stream func (p *PostgresCDCSource) PullRecords(ctx context.Context, req *model.PullRecordsRequest) error { - replicationOpts, err := p.replicationOptions() - if err != nil { - return fmt.Errorf("error getting replication options: %w", err) - } - - pgConn := p.replConn.PgConn() - - // start replication - var clientXLogPos, startLSN pglogrepl.LSN - if req.LastOffset > 0 { - p.logger.Info("starting replication from last sync state", slog.Int64("last checkpoint", req.LastOffset)) - clientXLogPos = pglogrepl.LSN(req.LastOffset) - startLSN = clientXLogPos + 1 - } - - opts := startReplicationOpts{ - conn: pgConn, - startLSN: startLSN, - replicationOpts: *replicationOpts, - } - - err = p.startReplication(ctx, opts) - if err != nil { - return fmt.Errorf("error starting replication: %w", err) - } - - p.logger.Info(fmt.Sprintf("started replication on slot %s at startLSN: %d", p.slot, startLSN)) - - return p.consumeStream(ctx, pgConn, req, clientXLogPos, req.RecordStream) -} - -func (p *PostgresCDCSource) startReplication(ctx context.Context, opts startReplicationOpts) error { - err := pglogrepl.StartReplication(ctx, opts.conn, p.slot, opts.startLSN, opts.replicationOpts) - if err != nil { - p.logger.Error("error starting replication", slog.Any("error", err)) - return fmt.Errorf("error starting replication at startLsn - %d: %w", opts.startLSN, err) - } - - p.logger.Info(fmt.Sprintf("started replication on slot %s at startLSN: %d", p.slot, opts.startLSN)) - return nil -} - -func (p *PostgresCDCSource) replicationOptions() (*pglogrepl.StartReplicationOptions, error) { - pluginArguments := []string{ - "proto_version '1'", - } - - if p.publication != "" { - pubOpt := fmt.Sprintf("publication_names '%s'", p.publication) - pluginArguments = append(pluginArguments, pubOpt) - } else { - return nil, fmt.Errorf("publication name is not set") - } - - return &pglogrepl.StartReplicationOptions{PluginArgs: pluginArguments}, nil -} - -// start consuming the cdc stream -func (p *PostgresCDCSource) consumeStream( - ctx context.Context, - conn *pgconn.PgConn, - req *model.PullRecordsRequest, - clientXLogPos pglogrepl.LSN, - records *model.CDCRecordStream, -) error { - defer func() { - err := conn.Close(ctx) - if err != nil { - p.logger.Error("error closing replication connection", slog.Any("error", err)) - } - }() - + conn := p.replConn.PgConn() + records := req.RecordStream // clientXLogPos is the last checkpoint id, we need to ack that we have processed // until clientXLogPos each time we send a standby status update. // consumedXLogPos is the lsn that has been committed on the destination. - consumedXLogPos := pglogrepl.LSN(0) - if clientXLogPos > 0 { + var clientXLogPos, consumedXLogPos pglogrepl.LSN + if req.LastOffset > 0 { + clientXLogPos = pglogrepl.LSN(req.LastOffset) consumedXLogPos = clientXLogPos err := pglogrepl.SendStandbyStatusUpdate(ctx, conn, @@ -307,7 +229,6 @@ func (p *PostgresCDCSource) consumeStream( var receiveCtx context.Context var cancel context.CancelFunc - if cdcRecordsStorage.IsEmpty() { receiveCtx, cancel = context.WithCancel(ctx) } else { diff --git a/flow/connectors/postgres/postgres.go b/flow/connectors/postgres/postgres.go index b8d3d27f20..32296cabf6 100644 --- a/flow/connectors/postgres/postgres.go +++ b/flow/connectors/postgres/postgres.go @@ -6,9 +6,11 @@ import ( "log/slog" "regexp" "strings" + "sync" "time" "github.com/google/uuid" + "github.com/jackc/pglogrepl" "github.com/jackc/pgx/v5" "github.com/jackc/pgx/v5/pgconn" "github.com/jackc/pgx/v5/pgtype" @@ -30,12 +32,21 @@ type PostgresConnector struct { ssh *SSHTunnel conn *pgx.Conn replConfig *pgx.ConnConfig + replConn *pgx.Conn + replState *ReplState + replLock sync.Mutex customTypesMapping map[uint32]string metadataSchema string hushWarnOID map[uint32]struct{} logger log.Logger } +type ReplState struct { + Slot string + Publication string + Offset int64 +} + func NewPostgresConnector(ctx context.Context, pgConfig *protos.PostgresConfig) (*PostgresConnector, error) { connectionString := utils.GetPGConnectionString(pgConfig) @@ -81,6 +92,8 @@ func NewPostgresConnector(ctx context.Context, pgConfig *protos.PostgresConfig) ssh: tunnel, conn: conn, replConfig: replConfig, + replState: nil, + replLock: sync.Mutex{}, customTypesMapping: customTypeMap, metadataSchema: metadataSchema, hushWarnOID: make(map[uint32]struct{}), @@ -94,13 +107,122 @@ func (c *PostgresConnector) CreateReplConn(ctx context.Context) (*pgx.Conn, erro logger.LoggerFromCtx(ctx).Error("failed to create replication connection", "error", err) return nil, fmt.Errorf("failed to create replication connection: %w", err) } - return conn, nil } +func (c *PostgresConnector) SetupReplConn(ctx context.Context) error { + conn, err := c.CreateReplConn(ctx) + if err != nil { + return err + } + c.replConn = conn + return nil +} + +func (c *PostgresConnector) ReplKeepAlive(ctx context.Context) (time.Duration, error) { + var timeout int64 + err := c.replConn.QueryRow(ctx, "show wal_sender_timeout").Scan(&timeout) + if err != nil { + return 0, err + } + return time.Duration(timeout) * time.Millisecond, nil +} + +// To keep connection alive between sync batches. +// By default postgres drops connection after 1 minute of inactivity. +func (c *PostgresConnector) ReplPing(ctx context.Context) error { + if c.replLock.TryLock() { + defer c.replLock.Unlock() + if c.replState != nil { + return pglogrepl.SendStandbyStatusUpdate( + ctx, + c.replConn.PgConn(), + pglogrepl.StandbyStatusUpdate{WALWritePosition: pglogrepl.LSN(c.replState.Offset)}, + ) + } + } + return nil +} + +func (c *PostgresConnector) MaybeStartReplication( + ctx context.Context, + slotName string, + publicationName string, + req *model.PullRecordsRequest, +) error { + if c.replState != nil && (c.replState.Offset != req.LastOffset || + c.replState.Slot != slotName || + c.replState.Publication != publicationName) { + return fmt.Errorf("replState changed, reset connector. slot name: old=%s new=%s, publication: old=%s new=%s, offset: old=%d new=%d", + c.replState.Slot, slotName, c.replState.Publication, publicationName, c.replState.Offset, req.LastOffset, + ) + } + + if c.replState == nil { + replicationOpts, err := c.replicationOptions(publicationName) + if err != nil { + return fmt.Errorf("error getting replication options: %w", err) + } + + var startLSN pglogrepl.LSN + if req.LastOffset > 0 { + c.logger.Info("starting replication from last sync state", slog.Int64("last checkpoint", req.LastOffset)) + startLSN = pglogrepl.LSN(req.LastOffset + 1) + } + + opts := startReplicationOpts{ + conn: c.replConn.PgConn(), + startLSN: startLSN, + replicationOpts: *replicationOpts, + } + + err = c.startReplication(ctx, slotName, opts) + if err != nil { + return fmt.Errorf("error starting replication: %w", err) + } + + c.logger.Info(fmt.Sprintf("started replication on slot %s at startLSN: %d", slotName, startLSN)) + c.replState = &ReplState{ + Slot: slotName, + Publication: publicationName, + Offset: req.LastOffset, + } + } + return nil +} + +func (c *PostgresConnector) startReplication(ctx context.Context, slotName string, opts startReplicationOpts) error { + err := pglogrepl.StartReplication(ctx, opts.conn, slotName, opts.startLSN, opts.replicationOpts) + if err != nil { + c.logger.Error("error starting replication", slog.Any("error", err)) + return fmt.Errorf("error starting replication at startLsn - %d: %w", opts.startLSN, err) + } + + c.logger.Info(fmt.Sprintf("started replication on slot %s at startLSN: %d", slotName, opts.startLSN)) + return nil +} + +func (c *PostgresConnector) replicationOptions(publicationName string) (*pglogrepl.StartReplicationOptions, error) { + pluginArguments := []string{ + "proto_version '1'", + } + + if publicationName != "" { + pubOpt := fmt.Sprintf("publication_names %s", QuoteLiteral(publicationName)) + pluginArguments = append(pluginArguments, pubOpt) + } else { + return nil, fmt.Errorf("publication name is not set") + } + + return &pglogrepl.StartReplicationOptions{PluginArgs: pluginArguments}, nil +} + // Close closes all connections. func (c *PostgresConnector) Close(ctx context.Context) error { if c != nil { + if c.replConn != nil { + c.replConn.Close(ctx) + } c.conn.Close(ctx) c.ssh.Close() } @@ -212,17 +334,24 @@ func (c *PostgresConnector) PullRecords(ctx context.Context, catalogPool *pgxpoo c.logger.Info("PullRecords: performed checks for slot and publication") - replConn, err := c.CreateReplConn(ctx) + childToParentRelIDMap, err := GetChildToParentRelIDMap(ctx, c.conn) + if err != nil { + return fmt.Errorf("error getting child to parent relid map: %w", err) + } + + c.replLock.Lock() + defer c.replLock.Unlock() + + err = c.MaybeStartReplication(ctx, slotName, publicationName, req) if err != nil { return err } - defer replConn.Close(ctx) - cdc, err := c.NewPostgresCDCSource(ctx, &PostgresCDCConfig{ - Connection: replConn, + cdc, err := c.NewPostgresCDCSource(&PostgresCDCConfig{ SrcTableIDNameMapping: req.SrcTableIDNameMapping, Slot: slotName, Publication: publicationName, + ChildToParentRelIDMap: childToParentRelIDMap, TableNameMapping: req.TableNameMapping, RelationMessageMapping: req.RelationMessageMapping, CatalogPool: catalogPool, @@ -236,11 +365,14 @@ func (c *PostgresConnector) PullRecords(ctx context.Context, catalogPool *pgxpoo if err != nil { return err } + req.RecordStream.Close() + c.replState.Offset = req.RecordStream.GetLastCheckpoint() latestLSN, err := c.getCurrentLSN(ctx) if err != nil { return fmt.Errorf("failed to get current LSN: %w", err) } + err = monitoring.UpdateLatestLSNAtSourceForCDCFlow(ctx, catalogPool, req.FlowJobName, latestLSN) if err != nil { return fmt.Errorf("failed to update latest LSN at source for CDC flow: %w", err) @@ -372,10 +504,7 @@ func (c *PostgresConnector) SyncRecords(ctx context.Context, req *model.SyncReco c.logger.Info(fmt.Sprintf("synced %d records to Postgres table %s via COPY", syncedRecordsCount, rawTableIdentifier)) - lastCP, err := req.Records.GetLastCheckpoint() - if err != nil { - return nil, fmt.Errorf("error getting last checkpoint: %w", err) - } + lastCP := req.Records.GetLastCheckpoint() // updating metadata with new offset and syncBatchID err = c.updateSyncMetadata(ctx, req.FlowJobName, lastCP, req.SyncBatchID, syncRecordsTx) @@ -946,7 +1075,6 @@ func (c *PostgresConnector) HandleSlotInfo( return nil } -// GetLastOffset returns the last synced offset for a job. func getOpenConnectionsForUser(ctx context.Context, conn *pgx.Conn, user string) (*protos.GetOpenConnectionsForUserResult, error) { row := conn.QueryRow(ctx, getNumConnectionsForUser, user) diff --git a/flow/connectors/s3/s3.go b/flow/connectors/s3/s3.go index 5628a3c3ff..6142a8d642 100644 --- a/flow/connectors/s3/s3.go +++ b/flow/connectors/s3/s3.go @@ -184,10 +184,7 @@ func (c *S3Connector) SyncRecords(ctx context.Context, req *model.SyncRecordsReq } c.logger.Info(fmt.Sprintf("Synced %d records", numRecords)) - lastCheckpoint, err := req.Records.GetLastCheckpoint() - if err != nil { - return nil, fmt.Errorf("failed to get last checkpoint: %w", err) - } + lastCheckpoint := req.Records.GetLastCheckpoint() err = c.pgMetadata.FinishBatch(ctx, req.FlowJobName, req.SyncBatchID, lastCheckpoint) if err != nil { diff --git a/flow/connectors/snowflake/snowflake.go b/flow/connectors/snowflake/snowflake.go index cb86bc4389..2f6c0210e5 100644 --- a/flow/connectors/snowflake/snowflake.go +++ b/flow/connectors/snowflake/snowflake.go @@ -478,13 +478,8 @@ func (c *SnowflakeConnector) syncRecordsViaAvro( return nil, fmt.Errorf("failed to sync schema changes: %w", err) } - lastCheckpoint, err := req.Records.GetLastCheckpoint() - if err != nil { - return nil, err - } - return &model.SyncResponse{ - LastSyncedCheckpointID: lastCheckpoint, + LastSyncedCheckpointID: req.Records.GetLastCheckpoint(), NumRecordsSynced: int64(numRecords), CurrentSyncBatchID: syncBatchID, TableNameRowsMapping: tableNameRowsMapping, diff --git a/flow/e2e/test_utils.go b/flow/e2e/test_utils.go index efcee75636..a1822f796d 100644 --- a/flow/e2e/test_utils.go +++ b/flow/e2e/test_utils.go @@ -20,8 +20,10 @@ import ( "github.com/stretchr/testify/require" "go.temporal.io/sdk/temporal" "go.temporal.io/sdk/testsuite" + "go.temporal.io/sdk/worker" "github.com/PeerDB-io/peer-flow/activities" + "github.com/PeerDB-io/peer-flow/connectors" connpostgres "github.com/PeerDB-io/peer-flow/connectors/postgres" connsnowflake "github.com/PeerDB-io/peer-flow/connectors/snowflake" "github.com/PeerDB-io/peer-flow/connectors/utils" @@ -58,7 +60,6 @@ func RegisterWorkflowsAndActivities(t *testing.T, env *testsuite.TestWorkflowEnv env.SetTestTimeout(5 * time.Minute) env.RegisterWorkflow(peerflow.CDCFlowWorkflowWithConfig) - env.RegisterWorkflow(peerflow.SyncFlowWorkflow) env.RegisterWorkflow(peerflow.SetupFlowWorkflow) env.RegisterWorkflow(peerflow.SnapshotFlowWorkflow) env.RegisterWorkflow(peerflow.NormalizeFlowWorkflow) @@ -74,6 +75,7 @@ func RegisterWorkflowsAndActivities(t *testing.T, env *testsuite.TestWorkflowEnv env.RegisterActivity(&activities.FlowableActivity{ CatalogPool: conn, Alerter: alerter, + CdcCache: make(map[string]connectors.CDCPullConnector), }) env.RegisterActivity(&activities.SnapshotActivity{ Alerter: alerter, @@ -550,10 +552,10 @@ func NewTemporalTestWorkflowEnvironment(t *testing.T) *testsuite.TestWorkflowEnv &slog.HandlerOptions{Level: slog.LevelWarn}, ), )) - tLogger := TStructuredLogger{logger: logger} + testSuite.SetLogger(&TStructuredLogger{logger: logger}) - testSuite.SetLogger(&tLogger) env := testSuite.NewTestWorkflowEnvironment() + env.SetWorkerOptions(worker.Options{EnableSessionWorker: true}) RegisterWorkflowsAndActivities(t, env) return env } diff --git a/flow/model/cdc_record_stream.go b/flow/model/cdc_record_stream.go index 29833112a9..dcdadfbb67 100644 --- a/flow/model/cdc_record_stream.go +++ b/flow/model/cdc_record_stream.go @@ -1,7 +1,6 @@ package model import ( - "errors" "sync/atomic" "github.com/PeerDB-io/peer-flow/generated/protos" @@ -41,11 +40,11 @@ func (r *CDCRecordStream) UpdateLatestCheckpoint(val int64) { } } -func (r *CDCRecordStream) GetLastCheckpoint() (int64, error) { +func (r *CDCRecordStream) GetLastCheckpoint() int64 { if !r.lastCheckpointSet { - return 0, errors.New("last checkpoint not set, stream is still active") + panic("last checkpoint not set, stream is still active") } - return r.lastCheckpointID.Load(), nil + return r.lastCheckpointID.Load() } func (r *CDCRecordStream) AddRecord(record Record) { @@ -66,9 +65,11 @@ func (r *CDCRecordStream) WaitAndCheckEmpty() bool { } func (r *CDCRecordStream) Close() { - close(r.emptySignal) - close(r.records) - r.lastCheckpointSet = true + if !r.lastCheckpointSet { + close(r.emptySignal) + close(r.records) + r.lastCheckpointSet = true + } } func (r *CDCRecordStream) GetRecords() <-chan Record { diff --git a/flow/workflows/cdc_flow.go b/flow/workflows/cdc_flow.go index afc17927ce..09af1b0c45 100644 --- a/flow/workflows/cdc_flow.go +++ b/flow/workflows/cdc_flow.go @@ -95,12 +95,12 @@ func (s *CDCFlowWorkflowState) TruncateProgress(logger log.Logger) { } if s.SyncFlowErrors != nil { - logger.Warn("SyncFlowErrors: ", s.SyncFlowErrors) + logger.Warn("SyncFlowErrors", "errors", s.SyncFlowErrors) s.SyncFlowErrors = nil } if s.NormalizeFlowErrors != nil { - logger.Warn("NormalizeFlowErrors: ", s.NormalizeFlowErrors) + logger.Warn("NormalizeFlowErrors", "errors", s.NormalizeFlowErrors) s.NormalizeFlowErrors = nil } } @@ -119,21 +119,24 @@ func NewCDCFlowWorkflowExecution(ctx workflow.Context) *CDCFlowWorkflowExecution } } -func GetChildWorkflowID( - ctx workflow.Context, - prefix string, - peerFlowName string, -) (string, error) { - childWorkflowIDSideEffect := workflow.SideEffect(ctx, func(ctx workflow.Context) interface{} { - return fmt.Sprintf("%s-%s-%s", prefix, peerFlowName, uuid.New().String()) +func GetUUID(ctx workflow.Context) (string, error) { + uuidSideEffect := workflow.SideEffect(ctx, func(ctx workflow.Context) interface{} { + return uuid.New().String() }) - var childWorkflowID string - if err := childWorkflowIDSideEffect.Get(&childWorkflowID); err != nil { - return "", fmt.Errorf("failed to get child workflow ID: %w", err) + var uuidString string + if err := uuidSideEffect.Get(&uuidString); err != nil { + return "", fmt.Errorf("failed to generate UUID: %w", err) } + return uuidString, nil +} - return childWorkflowID, nil +func GetChildWorkflowID( + prefix string, + peerFlowName string, + uuid string, +) string { + return fmt.Sprintf("%s-%s-%s", prefix, peerFlowName, uuid) } // CDCFlowWorkflowResult is the result of the PeerFlowWorkflow. @@ -164,16 +167,16 @@ func (w *CDCFlowWorkflowExecution) processCDCFlowConfigUpdates(ctx workflow.Cont return err } - additionalTablesWorkflowCfg := proto.Clone(cfg).(*protos.FlowConnectionConfigs) - additionalTablesWorkflowCfg.DoInitialSnapshot = true - additionalTablesWorkflowCfg.InitialSnapshotOnly = true - additionalTablesWorkflowCfg.TableMappings = flowConfigUpdate.AdditionalTables + additionalTablesCfg := proto.Clone(cfg).(*protos.FlowConnectionConfigs) + additionalTablesCfg.DoInitialSnapshot = true + additionalTablesCfg.InitialSnapshotOnly = true + additionalTablesCfg.TableMappings = flowConfigUpdate.AdditionalTables - childAdditionalTablesCDCFlowID, - err := GetChildWorkflowID(ctx, "additional-cdc-flow", additionalTablesWorkflowCfg.FlowJobName) + additionalTablesUUID, err := GetUUID(ctx) if err != nil { return err } + childAdditionalTablesCDCFlowID := GetChildWorkflowID("additional-cdc-flow", additionalTablesCfg.FlowJobName, additionalTablesUUID) // execute the sync flow as a child workflow childAdditionalTablesCDCFlowOpts := workflow.ChildWorkflowOptions{ @@ -189,7 +192,7 @@ func (w *CDCFlowWorkflowExecution) processCDCFlowConfigUpdates(ctx workflow.Cont childAdditionalTablesCDCFlowFuture := workflow.ExecuteChildWorkflow( childAdditionalTablesCDCFlowCtx, CDCFlowWorkflowWithConfig, - additionalTablesWorkflowCfg, + additionalTablesCfg, nil, ) var res *CDCFlowWorkflowResult @@ -248,6 +251,8 @@ func CDCFlowWorkflowWithConfig( shared.MirrorNameSearchAttribute: cfg.FlowJobName, } + originalRunID := workflow.GetInfo(ctx).OriginalRunID + // we cannot skip SetupFlow if SnapshotFlow did not complete in cases where Resync is enabled // because Resync modifies TableMappings before Setup and also before Snapshot // for safety, rely on the idempotency of SetupFlow instead @@ -268,10 +273,8 @@ func CDCFlowWorkflowWithConfig( // start the SetupFlow workflow as a child workflow, and wait for it to complete // it should return the table schema for the source peer - setupFlowID, err := GetChildWorkflowID(ctx, "setup-flow", cfg.FlowJobName) - if err != nil { - return state, err - } + setupFlowID := GetChildWorkflowID("setup-flow", cfg.FlowJobName, originalRunID) + childSetupFlowOpts := workflow.ChildWorkflowOptions{ WorkflowID: setupFlowID, ParentClosePolicy: enums.PARENT_CLOSE_POLICY_REQUEST_CANCEL, @@ -292,10 +295,7 @@ func CDCFlowWorkflowWithConfig( state.CurrentFlowStatus = protos.FlowStatus_STATUS_SNAPSHOT // next part of the setup is to snapshot-initial-copy and setup replication slots. - snapshotFlowID, err := GetChildWorkflowID(ctx, "snapshot-flow", cfg.FlowJobName) - if err != nil { - return state, err - } + snapshotFlowID := GetChildWorkflowID("snapshot-flow", cfg.FlowJobName, originalRunID) taskQueue, err := shared.GetPeerFlowTaskQueueName(shared.SnapshotFlowTaskQueueID) if err != nil { @@ -361,6 +361,54 @@ func CDCFlowWorkflowWithConfig( } } + sessionOptions := &workflow.SessionOptions{ + CreationTimeout: time.Minute, + ExecutionTimeout: time.Minute, + } + syncSessionCtx, err := workflow.CreateSession(ctx, sessionOptions) + if err != nil { + return nil, err + } + defer workflow.CompleteSession(syncSessionCtx) + sessionInfo := workflow.GetSessionInfo(syncSessionCtx) + + syncCtx := workflow.WithActivityOptions(syncSessionCtx, workflow.ActivityOptions{ + StartToCloseTimeout: 72 * time.Hour, + HeartbeatTimeout: time.Minute, + WaitForCancellation: true, + }) + fMaintain := workflow.ExecuteActivity( + syncCtx, + flowable.MaintainPull, + cfg, + sessionInfo.SessionID, + ) + fSessionSetup := workflow.ExecuteActivity( + syncCtx, + flowable.WaitForSourceConnector, + sessionInfo.SessionID, + ) + + var sessionError error + sessionSelector := workflow.NewNamedSelector(ctx, "Session Setup") + sessionSelector.AddFuture(fMaintain, func(f workflow.Future) { + // MaintainPull should never exit without an error before this point + sessionError = f.Get(syncCtx, nil) + }) + sessionSelector.AddFuture(fSessionSetup, func(f workflow.Future) { + // Happy path is waiting for this to return without error + sessionError = f.Get(syncCtx, nil) + }) + sessionSelector.AddReceive(ctx.Done(), func(_ workflow.ReceiveChannel, _ bool) { + sessionError = ctx.Err() + }) + sessionSelector.Select(ctx) + if sessionError != nil { + state.SyncFlowErrors = append(state.SyncFlowErrors, sessionError.Error()) + state.TruncateProgress(w.logger) + return state, workflow.NewContinueAsNewError(ctx, CDCFlowWorkflowWithConfig, cfg, state) + } + // when we carry forward state, don't remake the options if state.SyncFlowOptions == nil { state.SyncFlowOptions = &protos.SyncFlowOptions{ @@ -376,11 +424,7 @@ func CDCFlowWorkflowWithConfig( currentSyncFlowNum := 0 totalRecordsSynced := int64(0) - normalizeFlowID, err := GetChildWorkflowID(ctx, "normalize-flow", cfg.FlowJobName) - if err != nil { - return state, err - } - + normalizeFlowID := GetChildWorkflowID("normalize-flow", cfg.FlowJobName, originalRunID) childNormalizeFlowOpts := workflow.ChildWorkflowOptions{ WorkflowID: normalizeFlowID, ParentClosePolicy: enums.PARENT_CLOSE_POLICY_REQUEST_CANCEL, @@ -391,6 +435,7 @@ func CDCFlowWorkflowWithConfig( WaitForCancellation: true, } normCtx := workflow.WithChildOptions(ctx, childNormalizeFlowOpts) + childNormalizeFlowFuture := workflow.ExecuteChildWorkflow( normCtx, NormalizeFlowWorkflow, @@ -422,12 +467,12 @@ func CDCFlowWorkflowWithConfig( } var canceled bool - signalChan := workflow.GetSignalChannel(ctx, shared.FlowSignalName) - mainLoopSelector := workflow.NewSelector(ctx) + flowSignalChan := workflow.GetSignalChannel(ctx, shared.FlowSignalName) + mainLoopSelector := workflow.NewNamedSelector(ctx, "Main Loop") mainLoopSelector.AddReceive(ctx.Done(), func(_ workflow.ReceiveChannel, _ bool) { canceled = true }) - mainLoopSelector.AddReceive(signalChan, func(c workflow.ReceiveChannel, _ bool) { + mainLoopSelector.AddReceive(flowSignalChan, func(c workflow.ReceiveChannel, _ bool) { var signalVal shared.CDCFlowSignal c.ReceiveAsync(&signalVal) state.ActiveSignal = shared.FlowSignalHandler(state.ActiveSignal, signalVal, w.logger) @@ -491,43 +536,31 @@ func CDCFlowWorkflowWithConfig( " limit on the number of syncflows to be executed: ", currentSyncFlowNum) break } - currentSyncFlowNum++ - - syncFlowID, err := GetChildWorkflowID(ctx, "sync-flow", cfg.FlowJobName) - if err != nil { - finishNormalize() - return state, err - } + currentSyncFlowNum += 1 + w.logger.Info("executing sync flow", slog.Int("count", currentSyncFlowNum), slog.String("flowName", cfg.FlowJobName)) - // execute the sync flow as a child workflow - childSyncFlowOpts := workflow.ChildWorkflowOptions{ - WorkflowID: syncFlowID, - ParentClosePolicy: enums.PARENT_CLOSE_POLICY_REQUEST_CANCEL, - RetryPolicy: &temporal.RetryPolicy{ - MaximumAttempts: 20, - }, - SearchAttributes: mirrorNameSearch, - WaitForCancellation: true, + startFlowInput := &protos.StartFlowInput{ + FlowConnectionConfigs: cfg, + SyncFlowOptions: state.SyncFlowOptions, + RelationMessageMapping: state.RelationMessageMapping, + SrcTableIdNameMapping: state.SyncFlowOptions.SrcTableIdNameMapping, + TableNameSchemaMapping: state.SyncFlowOptions.TableNameSchemaMapping, + SessionId: sessionInfo.SessionID, } - syncCtx := workflow.WithChildOptions(ctx, childSyncFlowOpts) + fStartFlow := workflow.ExecuteActivity(syncCtx, flowable.StartFlow, startFlowInput) state.SyncFlowOptions.RelationMessageMapping = state.RelationMessageMapping - childSyncFlowFuture := workflow.ExecuteChildWorkflow( - syncCtx, - SyncFlowWorkflow, - cfg, - state.SyncFlowOptions, - ) - var syncDone bool + var syncDone, syncErr bool var normalizeSignalError error normDone := normWaitChan == nil - mainLoopSelector.AddFuture(childSyncFlowFuture, func(f workflow.Future) { + mainLoopSelector.AddFuture(fStartFlow, func(f workflow.Future) { syncDone = true var childSyncFlowRes *model.SyncResponse if err := f.Get(syncCtx, &childSyncFlowRes); err != nil { - w.logger.Error("failed to execute sync flow: ", err) + w.logger.Error("failed to execute sync flow", "error", err) state.SyncFlowErrors = append(state.SyncFlowErrors, err.Error()) + syncErr = true } else if childSyncFlowRes != nil { state.SyncFlowStatuses = append(state.SyncFlowStatuses, childSyncFlowRes) state.RelationMessageMapping = childSyncFlowRes.RelationMessageMapping @@ -587,8 +620,14 @@ func CDCFlowWorkflowWithConfig( if canceled { break } + if syncErr { + state.TruncateProgress(w.logger) + return state, workflow.NewContinueAsNewError(ctx, CDCFlowWorkflowWithConfig, cfg, state) + } if normalizeSignalError != nil { - return state, normalizeSignalError + state.NormalizeFlowErrors = append(state.NormalizeFlowErrors, normalizeSignalError.Error()) + state.TruncateProgress(w.logger) + return state, workflow.NewContinueAsNewError(ctx, CDCFlowWorkflowWithConfig, cfg, state) } if !normDone { normWaitChan.Receive(ctx, nil) diff --git a/flow/workflows/normalize_flow.go b/flow/workflows/normalize_flow.go index f37544b14a..16c53ba7a5 100644 --- a/flow/workflows/normalize_flow.go +++ b/flow/workflows/normalize_flow.go @@ -12,7 +12,8 @@ import ( "github.com/PeerDB-io/peer-flow/shared" ) -func NormalizeFlowWorkflow(ctx workflow.Context, +func NormalizeFlowWorkflow( + ctx workflow.Context, config *protos.FlowConnectionConfigs, ) (*model.NormalizeFlowResponse, error) { logger := workflow.GetLogger(ctx) @@ -25,10 +26,10 @@ func NormalizeFlowWorkflow(ctx workflow.Context, results := make([]model.NormalizeResponse, 0, 4) errors := make([]string, 0) syncChan := workflow.GetSignalChannel(ctx, shared.NormalizeSyncSignalName) - var tableNameSchemaMapping map[string]*protos.TableSchema var stopLoop, canceled bool var lastSyncBatchID, syncBatchID int64 + var tableNameSchemaMapping map[string]*protos.TableSchema lastSyncBatchID = -1 syncBatchID = -1 selector := workflow.NewNamedSelector(ctx, config.FlowJobName+"-normalize") diff --git a/flow/workflows/sync_flow.go b/flow/workflows/sync_flow.go deleted file mode 100644 index e3a53c9250..0000000000 --- a/flow/workflows/sync_flow.go +++ /dev/null @@ -1,79 +0,0 @@ -package peerflow - -import ( - "fmt" - "log/slog" - "time" - - "go.temporal.io/sdk/log" - "go.temporal.io/sdk/workflow" - - "github.com/PeerDB-io/peer-flow/generated/protos" - "github.com/PeerDB-io/peer-flow/model" -) - -type SyncFlowState struct { - CDCFlowName string - Progress []string -} - -type SyncFlowExecution struct { - SyncFlowState - executionID string - logger log.Logger -} - -// NewSyncFlowExecution creates a new instance of SyncFlowExecution. -func NewSyncFlowExecution(ctx workflow.Context, state *SyncFlowState) *SyncFlowExecution { - return &SyncFlowExecution{ - SyncFlowState: *state, - executionID: workflow.GetInfo(ctx).WorkflowExecution.ID, - logger: workflow.GetLogger(ctx), - } -} - -// executeSyncFlow executes the sync flow. -func (s *SyncFlowExecution) executeSyncFlow( - ctx workflow.Context, - config *protos.FlowConnectionConfigs, - opts *protos.SyncFlowOptions, - relationMessageMapping model.RelationMessageMapping, -) (*model.SyncResponse, error) { - s.logger.Info("executing sync flow", slog.String("flowName", s.CDCFlowName)) - - startFlowCtx := workflow.WithActivityOptions(ctx, workflow.ActivityOptions{ - StartToCloseTimeout: 72 * time.Hour, - HeartbeatTimeout: time.Minute, - WaitForCancellation: true, - }) - - // execute StartFlow on the peers to start the flow - startFlowInput := &protos.StartFlowInput{ - FlowConnectionConfigs: config, - SyncFlowOptions: opts, - RelationMessageMapping: relationMessageMapping, - SrcTableIdNameMapping: opts.SrcTableIdNameMapping, - TableNameSchemaMapping: opts.TableNameSchemaMapping, - } - fStartFlow := workflow.ExecuteActivity(startFlowCtx, flowable.StartFlow, startFlowInput) - - var syncRes *model.SyncResponse - if err := fStartFlow.Get(startFlowCtx, &syncRes); err != nil { - return nil, fmt.Errorf("failed to flow: %w", err) - } - return syncRes, nil -} - -// SyncFlowWorkflow is the synchronization workflow for a peer flow. -// This workflow assumes that the metadata tables have already been setup, -// and the checkpoint for the source peer is known. -func SyncFlowWorkflow(ctx workflow.Context, - config *protos.FlowConnectionConfigs, - options *protos.SyncFlowOptions, -) (*model.SyncResponse, error) { - s := NewSyncFlowExecution(ctx, &SyncFlowState{ - CDCFlowName: config.FlowJobName, - Progress: []string{}, - }) - return s.executeSyncFlow(ctx, config, options, options.RelationMessageMapping) -} diff --git a/protos/flow.proto b/protos/flow.proto index 2147ee2fc8..3bbebdc9a4 100644 --- a/protos/flow.proto +++ b/protos/flow.proto @@ -106,19 +106,13 @@ message SyncFlowOptions { repeated TableMapping table_mappings = 6; } -// deprecated, unused -message LastSyncState { - int64 checkpoint = 1; -} - message StartFlowInput { - // deprecated, unused - LastSyncState last_sync_state = 1; FlowConnectionConfigs flow_connection_configs = 2; SyncFlowOptions sync_flow_options = 3; map relation_message_mapping = 4; map src_table_id_name_mapping = 5; map table_name_schema_mapping = 6; + string session_id = 7; } message StartNormalizeInput {