-
Is it currently possible to configure sling to write to parquet in different encodings. It seems like the default encoding used is DELTA_LENGTH_BYTE_ARRAY, which is not supported when reading into Redshift Spectrum. |
Beta Was this translation helpful? Give feedback.
Answered by
flarco
May 4, 2024
Replies: 1 comment 3 replies
-
Hi, no it's not currently possible. |
Beta Was this translation helpful? Give feedback.
3 replies
Sign up for free
to join this conversation on GitHub.
Already have an account?
Sign in to comment
Yea, CSV is still the most compatible format. Parquet can have compatibility issues depending on the engine reading/writing it. BigQuery, for example, cannot export a JSON column into a Parquet file. You'd just have to manually cast it to a string. But it exports a JSON column to CSV no problem. So that's why Sling uses CSV as the temporary file format for bulk loading into Redshift/BigQuery (via COPY).