Replies: 2 comments 2 replies
-
My idea is a little different but I'm also solving a different problem: In
I've made a test for it. |
Beta Was this translation helpful? Give feedback.
-
@amiorin Lets not look at the creation process right now. We should use our Authorizer interface on who can create/read/update/delete not have a per entity logic.
If a user has both Redshift and Glue sources connected and want to extract the information we need to make sure it's not duplicated. |
Beta Was this translation helpful? Give feedback.
Uh oh!
There was an error while loading. Please reload this page.
Uh oh!
There was an error while loading. Please reload this page.
-
Flow
Once the crawler runs it creates a partition 2019 (as indicated in step 1) and sets up the table, with the columns. Customer edits/defines the schema to be accurate (update column names if not picked up correctly).
Customer goes to openmetadata-test s3 bucket to configure lifecycle rules on the customer_info prefix
Strategy
For the above table, the location is "s3://openmetadata-test/customer_info/" which is of location type "Prefix"
(not Bucket)
@harshach @amiorin Here's what I have in mind so far. I'm mostly focussing on step 3 of the strategy.
Looking for feedback and suggestions.
PS: The above example is limited to using a bucket per database. I've seen an example which have a handful of buckets and each bucket has several databases and tables, which again are namespaced by prefix. s3://bucket/database/table/partition/data.csv
Beta Was this translation helpful? Give feedback.
All reactions