April 2024

Upsolver new features, enhancements, and bug fixes for April 2024.

Release Notes Blog

For more detailed information on these updates, check out the Upsolver May 2024 Feature Summary blog.

2024.04.25-12.36

⬆️ Enhancements

  • Iceberg:

    • Added support for writing to hidden partitions

    • Enabled changing the partition specification of existing tables even while they are actively being written to by a job

    • Support writing to External Iceberg tables

    • Support altering Iceberg table properties via SQL

🔧 Bug Fixes

  • Worksheet tree - Show replication jobs under tables that were created dynamically

  • MongoDB CDC:

    • Corrected the parsing of Decimal types to Double

    • Resolved errors encountered when replicating collections containing fields with types Regex, Min Key, and Max Key


2024.04.16-12.06

⬆️ Enhancements

  • Introduced the PARSE_DEBEZIUM_JSON_TYPE property to the Avro Schema Registry content format for dynamic parsing of JSON columns from Debezium sources into Upsolver records or keeping as JSON strings. For Snowflake outputs with schema evolution, fields are written to columns of type Variant.

  • Upgraded the Snowflake driver to 3.15.0

  • UI: ClickHouse wizard cosmetic changes

🔧 Bug Fixes

  • Fixed a bug preventing the pausing of ingestion jobs to Snowflake

  • Iceberg schema evolution:

    • Nested fields were added without the field docs, which are later used to understand which field evolved from which. Affected tables may need to be recreated if jobs writing to them are causing errors

    • Was not handling cases where a field can have multiple types (e.g., a field can be a record and can also be an array of strings)


2024.04.04-09.33

New Features

⬆️ Enhancements

  • For new entities, you can now use the updated Parquet list structure (parquet.avro.write-old-list-structure = false) when writing Parquet files to S3 and Upsolver tables

  • Support casting strings to JSON in jobs writing to Iceberg tables

  • Previewing Classic Data Sources is now supported (SELECT * FROM "classic data source name")

  • Cost reduction:

    • Reduced S3 API costs of replication jobs and single entity jobs

    • Reduced S3 API costs of Iceberg tables

    • Reduced S3 API costs of Hive tables

🔧 Bug Fixes

  • Fixed a bug that could skip data when reading from CDC sources

  • Fixed a bug where events Written graph wouldn't show for single entity jobs that contains a lot of sub jobs or where the job list page contains a lot of jobs

  • CDC Event log is now deleted right after parsing the log events

  • Increased performance of VPC integration experience

  • Fixed a bug where a single entity job that reads data from a table that is partitioned by time wouldn't read from the start of the table

Last updated