Parquet files
Options
- Mark as New
- Bookmark
- Subscribe
- Mute
- Subscribe to RSS Feed
- Permalink
- Report Inappropriate Content
โ02-28-2022 08:00 PM
We have a usecase where we need to take data and reformat into parquet files to load to S3. Does anyone have a sample recipe or methodology where you've done this?
Thanks!
3 REPLIES 3
Options
- Mark as New
- Bookmark
- Subscribe
- Mute
- Subscribe to RSS Feed
- Permalink
- Report Inappropriate Content
โ02-28-2022 08:08 PM
Options
- Mark as New
- Bookmark
- Subscribe
- Mute
- Subscribe to RSS Feed
- Permalink
- Report Inappropriate Content
โ02-28-2022 08:10 PM
I saw this post online too - but I'm not sure how to import packages into the Workato python action and want to keep this off of the OPA.
Options
- Mark as New
- Bookmark
- Subscribe
- Mute
- Subscribe to RSS Feed
- Permalink
- Report Inappropriate Content
โ02-28-2022 08:16 PM
Hi Amy,
It's possible to do this with our JavaScript connector. See sample recipe here: https://app.workato.com/recipes/2226332?st=9a56d4
All you need to do is edit the json schema which defines the schema in the parquet file, then map the input data from there. No need to edit the code (FYI - we're using parquetjs-lite library for this).
That being said, your request is extremely timely as we are gathering feedback from users to support reading/writing Parquet.
Can you share the requirements for creating Parquet files? How large is the content for each file? Is there a need to control encoding/compression/partitions/row group size etc? Do you need to be able to append to an existing parquet file?
If there's anyone else in the community who needs Parquet support as well, please share with us your requirements.
Your feedback will help us prioritize this feature. Thanks!
--

