package main import ( "context" "database/sql" "sync" "time" "github.com/jackc/pgx/v5/pgxpool" _ "github.com/microsoft/go-mssqldb" log "github.com/sirupsen/logrus" ) func processMigrationJob(sourceDb *sql.DB, targetDb *pgxpool.Pool, job MigrationJob) { jobStartTime := time.Now() log.Infof("Starting migration job: %s.%s [PK: %s]", job.Schema, job.Table, job.PrimaryKey) sourceColTypes, targetColTypes, err := GetColumnTypes(sourceDb, targetDb, job) if err != nil { log.Fatal("Unexpected error: ", err) } logColumnTypes(sourceColTypes, "Source col types") logColumnTypes(targetColTypes, "Target col types") ctx, cancel := context.WithCancel(context.Background()) defer cancel() batches, err := batchGeneratorMssql(ctx, sourceDb, job) if err != nil { log.Error("Unexpected error calculating batch ranges: ", err) } chJobErrors := make(chan JobError, 50) chBatches := make(chan Batch, QueueSize) chExtractorErrors := make(chan ExtractorError, QueueSize) chChunksRaw := make(chan Chunk, QueueSize) chChunksTransformed := make(chan Chunk, QueueSize) chLoadersErrors := make(chan LoaderError, QueueSize) var wgActiveBatches sync.WaitGroup var wgActiveChunks sync.WaitGroup var wgExtractors sync.WaitGroup var wgTransformers sync.WaitGroup var wgLoaders sync.WaitGroup go func() { if err := jobErrorHandler(ctx, chJobErrors); err != nil { cancel() } }() go extractorErrorHandler(ctx, chExtractorErrors, chBatches, chJobErrors, &wgActiveBatches) go loaderErrorHandler(ctx, chLoadersErrors, chChunksTransformed, chJobErrors, &wgActiveChunks) maxExtractors := min(NumExtractors, len(batches)) log.Infof("Starting %d extractors...", maxExtractors) extractStartTime := time.Now() for range maxExtractors { wgExtractors.Go(func() { extractFromMssql(ctx, sourceDb, job, sourceColTypes, ChunkSize, chBatches, chChunksRaw, chExtractorErrors, chJobErrors, &wgActiveBatches) }) } wgActiveBatches.Add(len(batches)) go func() { for _, batch := range batches { chBatches <- batch } }() log.Infof("Starting %d transformers...", maxExtractors) transformStartTime := time.Now() for range maxExtractors { wgTransformers.Go(func() { transformRowsMssql(ctx, sourceColTypes, chChunksRaw, chChunksTransformed, chJobErrors, &wgActiveChunks) }) } log.Infof("Starting %d PostgreSQL loader(s)...", NumLoaders) loadStartTime := time.Now() for range NumLoaders { wgLoaders.Go(func() { loadRowsPostgres(ctx, targetDb, job, targetColTypes, chChunksTransformed, chLoadersErrors, chJobErrors, &wgActiveChunks) }) } go func() { wgActiveBatches.Wait() close(chBatches) close(chExtractorErrors) wgExtractors.Wait() log.Infof("Extraction completed in %v", time.Since(extractStartTime)) close(chChunksRaw) wgTransformers.Wait() log.Infof("Transformation completed in %v", time.Since(transformStartTime)) wgActiveChunks.Wait() close(chChunksTransformed) close(chLoadersErrors) wgLoaders.Wait() log.Infof("Loading completed in %v", time.Since(loadStartTime)) cancel() }() <-ctx.Done() log.Infof("Migration job completed. Total time: %v", time.Since(jobStartTime)) } func logColumnTypes(columnTypes []ColumnType, label string) { log.Debug(label) for _, col := range columnTypes { log.Debugf("%+v", col) } } func logSampleRow(job MigrationJob, columns []ColumnType, rowValues UnknownRowValues, tag string) { log.Infof("[%s.%s] Sample row: (%s)", job.Schema, job.Table, tag) for i, col := range columns { log.Infof("%s (%T): %v", col.Name(), rowValues[i], rowValues[i]) } }