-
Notifications
You must be signed in to change notification settings - Fork 8
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Source versioning: Postgres, MySQL and Load generator #647
base: main
Are you sure you want to change the base?
Conversation
c4a61c8
to
e202851
Compare
docs/resources/source_table.md
Outdated
@@ -29,7 +29,7 @@ description: |- | |||
- `ownership_role` (String) The owernship role of the object. | |||
- `region` (String) The region to use for the resource connection. If not set, the default region is used. | |||
- `schema_name` (String) The identifier for the table schema in Materialize. Defaults to `public`. | |||
- `text_columns` (List of String) Columns to be decoded as text. | |||
- `text_columns` (List of String) Columns to be decoded as text. Not supported for the load generator sources, if the source is a load generator, the attribute will be ignored. |
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
Similar to sources, we might want a source table resource per source type? The existing source-level options will basically shift to source table-level.
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
Yes indeed, I was just thinking about this. With the MySQL and Postgres sources, it is probably fine, but as soon as we add Kafka and Webhook sources, the logic will get out of hand.
Will refactor this to have a separate source table resource per source!
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
nice work!
docs/resources/source_kafka.md
Outdated
- `start_offset` (List of Number, Deprecated) Read partitions from the specified offset. Deprecated: Use the new materialize_source_table_kafka resource instead. | ||
- `start_timestamp` (Number, Deprecated) Use the specified value to set `START OFFSET` based on the Kafka timestamp. Deprecated: Use the new materialize_source_table_kafka resource instead. |
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
these two options are still currently only possible on the top-level CREATE SOURCE
statement for kafka sources -- not yet on a per-table basis. It will require a non-trivial amount more refactoring to allow them on a per-table basis so I'm unsure if we will do that work until it's requested by a customer
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
Ah yes! Good catch! Thank you!
docs/resources/source_mysql.md
Outdated
@@ -53,12 +53,12 @@ resource "materialize_source_mysql" "test" { | |||
- `comment` (String) **Public Preview** Comment on an object in the database. | |||
- `database_name` (String) The identifier for the source database in Materialize. Defaults to `MZ_DATABASE` environment variable if set or `materialize` if environment variable is not set. | |||
- `expose_progress` (Block List, Max: 1) The name of the progress collection for the source. If this is not specified, the collection will be named `<src_name>_progress`. (see [below for nested schema](#nestedblock--expose_progress)) | |||
- `ignore_columns` (List of String, Deprecated) Ignore specific columns when reading data from MySQL. Can only be updated in place when also updating a corresponding `table` attribute. Deprecated: Use the new materialize_source_table resource instead. | |||
- `ignore_columns` (List of String, Deprecated) Ignore specific columns when reading data from MySQL. Can only be updated in place when also updating a corresponding `table` attribute. Deprecated: Use the new materialize_source_table_mysql resource instead. |
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
fyi this option is also being renamed MaterializeInc/materialize#29438 but the old name will be aliased to the new one, so this shouldn't break
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
Sounds good! I will go ahead and use the exclude columns for the new table source resource!
docs/resources/source_table_kafka.md
Outdated
- `start_offset` (List of Number) Read partitions from the specified offset. | ||
- `start_timestamp` (Number) Use the specified value to set `START OFFSET` based on the Kafka timestamp. |
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
these aren't currently available on a per-table basis for kafka sources
docs/resources/source_table_kafka.md
Outdated
- `schema_name` (String) The identifier for the source schema in Materialize. Defaults to `public`. | ||
- `start_offset` (List of Number) Read partitions from the specified offset. | ||
- `start_timestamp` (Number) Use the specified value to set `START OFFSET` based on the Kafka timestamp. | ||
- `upstream_schema_name` (String) The schema of the table in the upstream database. |
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
what does this refer to for kafka sources? we might just want to omit it since the upstream reference should just be the kafka topic name
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
Good catch, this was an overlook on my end in the schema for the Kafka source table resource.
startOffset []int | ||
startTimestamp int |
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
these two aren't used below and also aren't possible on the statement
|
||
This guide will walk you through the process of migrating your existing source table definitions to the new `materialize_source_table_{source}` resource. | ||
|
||
For each source type (e.g., MySQL, Postgres, etc.), you will need to create a new `materialize_source_table_{source}` resource for each table that was previously defined within the source resource. This ensures that the tables are preserved during the migration process. |
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
For each source type (e.g., MySQL, Postgres, etc.), you will need to create a new `materialize_source_table_{source}` resource for each table that was previously defined within the source resource. This ensures that the tables are preserved during the migration process. | |
For each source type (e.g., MySQL, Postgres, etc.), you will need to create a new `materialize_source_table_{source}` resource for each table that was previously defined within the source resource. This ensures that the tables are preserved during the migration process. For Kafka sources, you will need to create at least one `materialize_source_table_kafka` table to hold data for the kafka topic. |
@morsapaes might have better wording for this but I think we should be clear that this migration needs to happen for sources that previously didn't have subsources too (e.g. kafka)
|
||
The same approach can be used for other source types such as Postgres, eg. `materialize_source_table_postgres`. | ||
|
||
## Automated Migration Process (TBD) |
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
nice - this is great! We will probably want to figure out how to tell them that they will be able to coordinate the 'automated' migration process with their field-engineer representative if they go down this path
0b63a29
to
7870d48
Compare
@morsapaes @bobbyiliev let's discuss this PR at the sources & sinks meeting this week - we should decide when it makes sense to merge this - my thinking is we should do so whenever we move into private preview for the source versioning feature. But if we want to merge sooner and just have a disclaimer that the things mentioned as 'deprecated' here are not actually yet deprecated, that could work too |
61cd55c
to
9b96939
Compare
One thing that we can consider here as per this old tracking issue: #391 is take the chance and decide if we still want to rename some of the attributes in the new source table resources: Lists For attributes that use list, we have more cases of singular than plural.
Blocks We had decided should be singular. There are some blocks that use plural so this could be a good chance to rename those attributes in the new source table load gen resource:
|
a467a3e
to
b0ee9ba
Compare
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
Don't let the ~51 comments scare you off -- they're mostly nits. Great job on this!!!
options = append(options, fmt.Sprintf(`FORMAT AVRO USING CONFLUENT SCHEMA REGISTRY CONNECTION %s`, QualifiedName(b.format.Avro.SchemaRegistryConnection.DatabaseName, b.format.Avro.SchemaRegistryConnection.SchemaName, b.format.Avro.SchemaRegistryConnection.Name))) | ||
} | ||
if b.format.Avro.KeyStrategy != "" { | ||
options = append(options, fmt.Sprintf(`KEY STRATEGY %s`, b.format.Avro.KeyStrategy)) |
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
Should we be quoting/escaping this?
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
Here we would not need to escape this is a keyword. If we were to escape it it would result in the following error on the Materialize side:
Expected one of ID or LATEST or INLINE, found string literal "INLINE"
We do however have validation in place on the terraform schema side already, so a user will only be able to specify one of the following values:
terraform-provider-materialize/pkg/resources/config.go
Lines 58 to 62 in 6831919
var strategy = []string{ | |
"INLINE", | |
"ID", | |
"LATEST", | |
} |
options = append(options, fmt.Sprintf(`KEY STRATEGY %s`, b.format.Avro.KeyStrategy)) | ||
} | ||
if b.format.Avro.ValueStrategy != "" { | ||
options = append(options, fmt.Sprintf(`VALUE STRATEGY %s`, b.format.Avro.ValueStrategy)) |
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
Should we be quoting/escaping this?
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
Same as above:
Here we would not need to escape this is a keyword. If we were to escape it it would result in the following error on the Materialize side:
Expected one of ID or LATEST or INLINE, found string literal "INLINE"
We do however have validation in place on the terraform schema side already, so a user will only be able to specify one of the following values:
terraform-provider-materialize/pkg/resources/config.go
Lines 58 to 62 in 6831919
var strategy = []string{ | |
"INLINE", | |
"ID", | |
"LATEST", | |
} |
|
||
if b.format.Protobuf != nil { | ||
if b.format.Protobuf.SchemaRegistryConnection.Name != "" && b.format.Protobuf.MessageName != "" { | ||
options = append(options, fmt.Sprintf(`FORMAT PROTOBUF MESSAGE '%s' USING CONFLUENT SCHEMA REGISTRY CONNECTION %s`, b.format.Protobuf.MessageName, QualifiedName(b.format.Protobuf.SchemaRegistryConnection.DatabaseName, b.format.Protobuf.SchemaRegistryConnection.SchemaName, b.format.Protobuf.SchemaRegistryConnection.Name))) |
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
We should probably quote the MessageName
options = append(options, fmt.Sprintf(`FORMAT CSV WITH %d COLUMNS`, b.format.Csv.Columns)) | ||
} | ||
if b.format.Csv.Header != nil { | ||
options = append(options, fmt.Sprintf(`FORMAT CSV WITH HEADER ( %s )`, strings.Join(b.format.Csv.Header, ", "))) | ||
} | ||
if b.format.Csv.DelimitedBy != "" { | ||
options = append(options, fmt.Sprintf(`DELIMITER '%s'`, b.format.Csv.DelimitedBy)) |
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
Should we be quoting/escaping these?
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
the format.Csv.Columns
is of TypeInt so no need to quote it, but good catch for the other ones! 🙇
|
||
if b.keyFormat.Protobuf != nil { | ||
if b.keyFormat.Protobuf.SchemaRegistryConnection.Name != "" && b.keyFormat.Protobuf.MessageName != "" { | ||
options = append(options, fmt.Sprintf(`KEY FORMAT PROTOBUF MESSAGE '%s' USING CONFLUENT SCHEMA REGISTRY CONNECTION %s`, b.keyFormat.Protobuf.MessageName, QualifiedName(b.keyFormat.Protobuf.SchemaRegistryConnection.DatabaseName, b.keyFormat.Protobuf.SchemaRegistryConnection.SchemaName, b.keyFormat.Protobuf.SchemaRegistryConnection.Name))) |
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
We should quote the MessageName
Description: "Specify the tables to be included in the source. Deprecated: Use the new materialize_source_table_mysql resource instead.", | ||
Deprecated: "Use the new materialize_source_table_mysql resource instead.", |
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
Description: "Specify the tables to be included in the source. Deprecated: Use the new materialize_source_table_mysql resource instead.", | |
Deprecated: "Use the new materialize_source_table_mysql resource instead.", | |
Description: "Specify the tables to be included in the source. Deprecated: Use the new `materialize_source_table_mysql` resource instead.", | |
Deprecated: "Use the new `materialize_source_table_mysql` resource instead.", |
@@ -76,6 +79,13 @@ var sourceMySQLSchema = map[string]*schema.Schema{ | |||
}, | |||
}, | |||
}, | |||
"all_tables": { | |||
Description: "Include all tables in the source. If `table` is specified, this will be ignored.", | |||
Deprecated: "Use the new materialize_source_table_mysql resource instead.", |
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
Deprecated: "Use the new materialize_source_table_mysql resource instead.", | |
Deprecated: "Use the new `materialize_source_table_mysql` resource instead.", |
Description: "Decode data as text for specific columns that contain PostgreSQL types that are unsupported in Materialize. Can only be updated in place when also updating a corresponding `table` attribute. Deprecated: Use the new materialize_source_table_postgres resource instead.", | ||
Deprecated: "Use the new materialize_source_table_postgres resource instead.", |
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
Description: "Decode data as text for specific columns that contain PostgreSQL types that are unsupported in Materialize. Can only be updated in place when also updating a corresponding `table` attribute. Deprecated: Use the new materialize_source_table_postgres resource instead.", | |
Deprecated: "Use the new materialize_source_table_postgres resource instead.", | |
Description: "Decode data as text for specific columns that contain PostgreSQL types that are unsupported in Materialize. Can only be updated in place when also updating a corresponding `table` attribute. Deprecated: Use the new `materialize_source_table_postgres` resource instead.", | |
Deprecated: "Use the new `materialize_source_table_postgres` resource instead.", |
Description: "Creates subsources for specific tables in the Postgres connection. Deprecated: Use the new materialize_source_table_postgres resource instead.", | ||
Deprecated: "Use the new materialize_source_table_postgres resource instead.", |
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
Description: "Creates subsources for specific tables in the Postgres connection. Deprecated: Use the new materialize_source_table_postgres resource instead.", | |
Deprecated: "Use the new materialize_source_table_postgres resource instead.", | |
Description: "Creates subsources for specific tables in the Postgres connection. Deprecated: Use the new `materialize_source_table_postgres` resource instead.", | |
Deprecated: "Use the new `materialize_source_table_postgres` resource instead.", |
ForceNew: true, | ||
}, | ||
"exclude_columns": { | ||
Description: "Exclude specific columns when reading data from MySQL. The option used to be called `ignore_columns`.", |
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
Description: "Exclude specific columns when reading data from MySQL. The option used to be called `ignore_columns`.", | |
Description: "Exclude specific columns when reading data from MySQL. This option used to be called `ignore_columns`.", |
af8d738
to
232787a
Compare
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
looks solid! Just a few minor comments on some things that have been updated, and one about the migration guide
# generated by https://github.com/hashicorp/terraform-plugin-docs | ||
page_title: "materialize_source_reference Data Source - terraform-provider-materialize" | ||
subcategory: "" | ||
description: |- |
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
Can we populate this to explain that these are 'available' source references? Such that these expose all the possible upstream references that this source can create a table for, not necessarily all the references it is already ingesting
|
||
This guide will walk you through the process of migrating your existing source table definitions to the new `materialize_source_table_{source_type}` resource. | ||
|
||
For each source type (e.g., MySQL, Postgres, etc.), you will need to create a new `materialize_source_table_{source_type}` resource for each table that was previously defined within the source resource. This ensures that the tables are preserved during the migration process. For Kafka sources, you will need to create at least one `materialize_source_table_kafka` table to hold data for the kafka topic. |
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
For each source type (e.g., MySQL, Postgres, etc.), you will need to create a new `materialize_source_table_{source_type}` resource for each table that was previously defined within the source resource. This ensures that the tables are preserved during the migration process. For Kafka sources, you will need to create at least one `materialize_source_table_kafka` table to hold data for the kafka topic. | |
For each MySQL and Postgres source, you will need to create a new `materialize_source_table_{source_type}` resource for each table that was previously defined within the source resource. This ensures that the tables are preserved during the migration process. For Kafka sources, you will need to create a `materialize_source_table_kafka` table with the same name as the kafka source to contain the data for the kafka topic. |
|
||
In previous versions of the Materialize Terraform provider, source tables were defined within the source resource itself and were considered subsources of the source rather than separate entities. | ||
|
||
This guide will walk you through the process of migrating your existing source table definitions to the new `materialize_source_table_{source_type}` resource. |
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
I'm confused on whether this guide is meant to explain how to migrate using just the manual process (which is outlined below and looks correct), or also to explain how to reconcile your terraform configuration with the results of the 'automatic' migration process that we would do for a customer with a catalog migration. Specifically, the paragraph below sounds like it is explaining what to do to handle the automatic migration process, but it doesn't explain that the manual process would then be unnecessary.
It might be simpler if this guide just explained the manual process and we had a separate one to explain the 'automatic' process. Or we could just assume any terraform user would only do the manual process. @morsapaes thoughts?
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
Good point! I've removed some of the confusing text so that this should focus on the manual migration!
Happy to work on a follow up migration guide for the automated migration if we decide to do that!
|
||
## Future Improvements | ||
|
||
The Kafka and Webhooks sources are currently being implemented. Once these changes, the migration process will be updated to include them. |
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
The Kafka and Webhooks sources are currently being implemented. Once these changes, the migration process will be updated to include them. | |
Webhooks sources have not yet been migrated to the new model. Once this changes, the migration process will be updated to include them. |
docs/resources/source_table_kafka.md
Outdated
|
||
- `name` (String) The identifier for the source table. | ||
- `source` (Block List, Min: 1, Max: 1) The source this table is created from. (see [below for nested schema](#nestedblock--source)) | ||
- `topic` (String) The name of the Kafka topic in the Kafka cluster. |
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
this is actually no longer necessary to include for Kafka sources -- if the (REFERENCE ..)
option in the statement is omitted it will still work since there is only one possible kafka topic that the table can reference
p := map[string]string{ | ||
"sr.source_id": sourceId, | ||
} | ||
q := sourceReferenceQuery.QueryPredicate(p) |
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
this would be a bonus, but there is a PR to implement a REFRESH SOURCE REFERENCES <source>
statement here MaterializeInc/materialize#29923 and we could automatically run that before listing the references which would mean we always get the most 'up to date' view of them
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
Just implemented this!
terraform-provider-materialize/pkg/materialize/source_reference.go
Lines 65 to 69 in 147606a
func refreshSourceReferences(conn *sqlx.DB, sourceName, schemaName, databaseName string) error { | |
query := fmt.Sprintf(`ALTER SOURCE %s REFRESH REFERENCES`, QualifiedName(databaseName, schemaName, sourceName)) | |
_, err := conn.Exec(query) | |
return err | |
} |
pkg/materialize/source_table.go
Outdated
q.WriteString(` (REFERENCE `) | ||
|
||
if b.upstreamSchemaName != "" { | ||
q.WriteString(fmt.Sprintf(`%s.`, QuoteIdentifier(b.upstreamSchemaName))) | ||
} | ||
q.WriteString(QuoteIdentifier(b.upstreamName)) | ||
|
||
q.WriteString(")") |
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
mentioned above that this can be made optional for kafka and single-output load generator sources
mz_sources.name AS source_name, | ||
source_schemas.name AS source_schema_name, | ||
source_databases.name AS source_database_name, | ||
mz_kafka_source_tables.topic AS upstream_table_name, |
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
I have a PR open now to add the envelope
, key_format
and value_format
columns to this table MaterializeInc/materialize#30076
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
Neat! I created a tracking issue for this for the moment: #665
Should be ok to handle in a follow up PR later on!
36cd7d1
to
147606a
Compare
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
Really just skimmed the entire PR but I think it all looks good! Biggest feedback would be following up on quoting identifiers that Aru commented on, but if there aren't any concerns there I would say we're good to go!
resource.TestCheckResourceAttr("materialize_source_table_kafka.test_kafka", "name", nameSpace+"_table_kafka"), | ||
resource.TestCheckResourceAttr("materialize_source_table_kafka.test_kafka", "database_name", "materialize"), | ||
resource.TestCheckResourceAttr("materialize_source_table_kafka.test_kafka", "schema_name", "public"), | ||
// resource.TestCheckResourceAttr("materialize_source_table_kafka.test_kafka", "qualified_sql_name", fmt.Sprintf(`"materialize"."public"."%s_table_kafka"`, nameSpace)), |
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
left behind?
8d433e5
to
b411b09
Compare
99d66f5
to
d5a738b
Compare
b73611f
to
e3bea80
Compare
Initial implementation for the source versioning refactor as per #646
The main changes to consider:
table
attribute as optional and deprecated for both the MySQL and Postgres sourcesall_tables
bool attribute for the MySQL and the Loadgen sources, as in the past this was defaulting always usingFOR ALL TABLES
in the load gen sources (auction, marketing, tpch) and in the MySQL case whenever notable
blocks were defined, we defaulted toFOR ALL TABLES
. Thisall_tables
bool attribute allows us to create sources without any tables defined as per the source versioning workmaterialize_source_table_{mysql|postgres|load_generator}
resource which allows us to doCREATE TABLE ... FROM SOURCE ...
Things that are still pending: #646