feat(writer): add continue-on-error option for SQL writers

* Introduce ContinueOnError option to WriterOptions
* Update writer functions to support continue-on-error behavior
* Modify migration and database writing to handle continue-on-error
This commit is contained in:
Hein
2026-05-07 14:44:36 +02:00
parent 30ef1db010
commit b91d6b33b5
9 changed files with 128 additions and 64 deletions

View File

@@ -43,16 +43,17 @@ import (
)
var (
convertSourceType string
convertSourcePath string
convertSourceConn string
convertFromList []string
convertTargetType string
convertTargetPath string
convertPackageName string
convertSchemaFilter string
convertFlattenSchema bool
convertNullableTypes string
convertSourceType string
convertSourcePath string
convertSourceConn string
convertFromList []string
convertTargetType string
convertTargetPath string
convertPackageName string
convertSchemaFilter string
convertFlattenSchema bool
convertNullableTypes string
convertContinueOnError bool
)
var convertCmd = &cobra.Command{
@@ -177,6 +178,7 @@ func init() {
convertCmd.Flags().StringVar(&convertSchemaFilter, "schema", "", "Filter to a specific schema by name (required for formats like dctx that only support single schemas)")
convertCmd.Flags().BoolVar(&convertFlattenSchema, "flatten-schema", false, "Flatten schema.table names to schema_table (useful for databases like SQLite that do not support schemas)")
convertCmd.Flags().StringVar(&convertNullableTypes, "types", "", "Nullable type package for code-gen writers (bun/gorm): 'resolvespec' (default) or 'stdlib' (database/sql)")
convertCmd.Flags().BoolVar(&convertContinueOnError, "continue-on-error", false, "Prepend \\set ON_ERROR_STOP off to generated SQL so psql continues past errors (pgsql output only)")
err := convertCmd.MarkFlagRequired("from")
if err != nil {
@@ -243,7 +245,7 @@ func runConvert(cmd *cobra.Command, args []string) error {
fmt.Fprintf(os.Stderr, " Schema: %s\n", convertSchemaFilter)
}
if err := writeDatabase(db, convertTargetType, convertTargetPath, convertPackageName, convertSchemaFilter, convertFlattenSchema, convertNullableTypes); err != nil {
if err := writeDatabase(db, convertTargetType, convertTargetPath, convertPackageName, convertSchemaFilter, convertFlattenSchema, convertNullableTypes, convertContinueOnError); err != nil {
return fmt.Errorf("failed to write target: %w", err)
}
@@ -383,10 +385,10 @@ func readDatabaseForConvert(dbType, filePath, connString string) (*models.Databa
return db, nil
}
func writeDatabase(db *models.Database, dbType, outputPath, packageName, schemaFilter string, flattenSchema bool, nullableTypes string) error {
func writeDatabase(db *models.Database, dbType, outputPath, packageName, schemaFilter string, flattenSchema bool, nullableTypes string, continueOnError bool) error {
var writer writers.Writer
writerOpts := newWriterOptions(outputPath, packageName, flattenSchema, nullableTypes)
writerOpts := newWriterOptions(outputPath, packageName, flattenSchema, nullableTypes, continueOnError)
switch strings.ToLower(dbType) {
case "dbml":

View File

@@ -323,31 +323,31 @@ func writeDatabaseForEdit(dbType, filePath, connString string, db *models.Databa
switch strings.ToLower(dbType) {
case "dbml":
writer = wdbml.NewWriter(newWriterOptions(filePath, "", false, ""))
writer = wdbml.NewWriter(newWriterOptions(filePath, "", false, "", false))
case "dctx":
writer = wdctx.NewWriter(newWriterOptions(filePath, "", false, ""))
writer = wdctx.NewWriter(newWriterOptions(filePath, "", false, "", false))
case "drawdb":
writer = wdrawdb.NewWriter(newWriterOptions(filePath, "", false, ""))
writer = wdrawdb.NewWriter(newWriterOptions(filePath, "", false, "", false))
case "graphql":
writer = wgraphql.NewWriter(newWriterOptions(filePath, "", false, ""))
writer = wgraphql.NewWriter(newWriterOptions(filePath, "", false, "", false))
case "json":
writer = wjson.NewWriter(newWriterOptions(filePath, "", false, ""))
writer = wjson.NewWriter(newWriterOptions(filePath, "", false, "", false))
case "yaml":
writer = wyaml.NewWriter(newWriterOptions(filePath, "", false, ""))
writer = wyaml.NewWriter(newWriterOptions(filePath, "", false, "", false))
case "gorm":
writer = wgorm.NewWriter(newWriterOptions(filePath, "", false, ""))
writer = wgorm.NewWriter(newWriterOptions(filePath, "", false, "", false))
case "bun":
writer = wbun.NewWriter(newWriterOptions(filePath, "", false, ""))
writer = wbun.NewWriter(newWriterOptions(filePath, "", false, "", false))
case "drizzle":
writer = wdrizzle.NewWriter(newWriterOptions(filePath, "", false, ""))
writer = wdrizzle.NewWriter(newWriterOptions(filePath, "", false, "", false))
case "prisma":
writer = wprisma.NewWriter(newWriterOptions(filePath, "", false, ""))
writer = wprisma.NewWriter(newWriterOptions(filePath, "", false, "", false))
case "typeorm":
writer = wtypeorm.NewWriter(newWriterOptions(filePath, "", false, ""))
writer = wtypeorm.NewWriter(newWriterOptions(filePath, "", false, "", false))
case "sqlite", "sqlite3":
writer = wsqlite.NewWriter(newWriterOptions(filePath, "", false, ""))
writer = wsqlite.NewWriter(newWriterOptions(filePath, "", false, "", false))
case "pgsql":
writer = wpgsql.NewWriter(newWriterOptions(filePath, "", false, ""))
writer = wpgsql.NewWriter(newWriterOptions(filePath, "", false, "", false))
default:
return fmt.Errorf("%s: unsupported format: %s", label, dbType)
}

View File

@@ -375,61 +375,61 @@ func writeDatabaseForMerge(dbType, filePath, connString string, db *models.Datab
if filePath == "" {
return fmt.Errorf("%s: file path is required for DBML format", label)
}
writer = wdbml.NewWriter(newWriterOptions(filePath, "", flattenSchema, ""))
writer = wdbml.NewWriter(newWriterOptions(filePath, "", flattenSchema, "", false))
case "dctx":
if filePath == "" {
return fmt.Errorf("%s: file path is required for DCTX format", label)
}
writer = wdctx.NewWriter(newWriterOptions(filePath, "", flattenSchema, ""))
writer = wdctx.NewWriter(newWriterOptions(filePath, "", flattenSchema, "", false))
case "drawdb":
if filePath == "" {
return fmt.Errorf("%s: file path is required for DrawDB format", label)
}
writer = wdrawdb.NewWriter(newWriterOptions(filePath, "", flattenSchema, ""))
writer = wdrawdb.NewWriter(newWriterOptions(filePath, "", flattenSchema, "", false))
case "graphql":
if filePath == "" {
return fmt.Errorf("%s: file path is required for GraphQL format", label)
}
writer = wgraphql.NewWriter(newWriterOptions(filePath, "", flattenSchema, ""))
writer = wgraphql.NewWriter(newWriterOptions(filePath, "", flattenSchema, "", false))
case "json":
if filePath == "" {
return fmt.Errorf("%s: file path is required for JSON format", label)
}
writer = wjson.NewWriter(newWriterOptions(filePath, "", flattenSchema, ""))
writer = wjson.NewWriter(newWriterOptions(filePath, "", flattenSchema, "", false))
case "yaml":
if filePath == "" {
return fmt.Errorf("%s: file path is required for YAML format", label)
}
writer = wyaml.NewWriter(newWriterOptions(filePath, "", flattenSchema, ""))
writer = wyaml.NewWriter(newWriterOptions(filePath, "", flattenSchema, "", false))
case "gorm":
if filePath == "" {
return fmt.Errorf("%s: file path is required for GORM format", label)
}
writer = wgorm.NewWriter(newWriterOptions(filePath, "", flattenSchema, ""))
writer = wgorm.NewWriter(newWriterOptions(filePath, "", flattenSchema, "", false))
case "bun":
if filePath == "" {
return fmt.Errorf("%s: file path is required for Bun format", label)
}
writer = wbun.NewWriter(newWriterOptions(filePath, "", flattenSchema, ""))
writer = wbun.NewWriter(newWriterOptions(filePath, "", flattenSchema, "", false))
case "drizzle":
if filePath == "" {
return fmt.Errorf("%s: file path is required for Drizzle format", label)
}
writer = wdrizzle.NewWriter(newWriterOptions(filePath, "", flattenSchema, ""))
writer = wdrizzle.NewWriter(newWriterOptions(filePath, "", flattenSchema, "", false))
case "prisma":
if filePath == "" {
return fmt.Errorf("%s: file path is required for Prisma format", label)
}
writer = wprisma.NewWriter(newWriterOptions(filePath, "", flattenSchema, ""))
writer = wprisma.NewWriter(newWriterOptions(filePath, "", flattenSchema, "", false))
case "typeorm":
if filePath == "" {
return fmt.Errorf("%s: file path is required for TypeORM format", label)
}
writer = wtypeorm.NewWriter(newWriterOptions(filePath, "", flattenSchema, ""))
writer = wtypeorm.NewWriter(newWriterOptions(filePath, "", flattenSchema, "", false))
case "sqlite", "sqlite3":
writer = wsqlite.NewWriter(newWriterOptions(filePath, "", flattenSchema, ""))
writer = wsqlite.NewWriter(newWriterOptions(filePath, "", flattenSchema, "", false))
case "pgsql":
writerOpts := newWriterOptions(filePath, "", flattenSchema, "")
writerOpts := newWriterOptions(filePath, "", flattenSchema, "", false)
if connString != "" {
writerOpts.Metadata = map[string]interface{}{
"connection_string": connString,

View File

@@ -13,12 +13,13 @@ func newReaderOptions(filePath, connString string) *readers.ReaderOptions {
}
}
func newWriterOptions(outputPath, packageName string, flattenSchema bool, nullableTypes string) *writers.WriterOptions {
func newWriterOptions(outputPath, packageName string, flattenSchema bool, nullableTypes string, continueOnError bool) *writers.WriterOptions {
return &writers.WriterOptions{
OutputPath: outputPath,
PackageName: packageName,
FlattenSchema: flattenSchema,
NullableTypes: nullableTypes,
Prisma7: prisma7,
OutputPath: outputPath,
PackageName: packageName,
FlattenSchema: flattenSchema,
NullableTypes: nullableTypes,
Prisma7: prisma7,
ContinueOnError: continueOnError,
}
}

View File

@@ -188,6 +188,7 @@ func runSplit(cmd *cobra.Command, args []string) error {
"", // no schema filter for split
false, // no flatten-schema for split
splitNullableTypes,
false, // no continue-on-error for split
)
if err != nil {
return fmt.Errorf("failed to write output: %w", err)