<?xml version="1.0" encoding="UTF-8"?>
<rss xmlns:content="http://purl.org/rss/1.0/modules/content/" xmlns:dc="http://purl.org/dc/elements/1.1/" xmlns:rdf="http://www.w3.org/1999/02/22-rdf-syntax-ns#" xmlns:taxo="http://purl.org/rss/1.0/modules/taxonomy/" version="2.0">
  <channel>
    <title>topic Re: Cache millions of records in Workato Pros Discussion Board</title>
    <link>https://systematic.workato.com/t5/workato-pros-discussion-board/cache-millions-of-records/m-p/2575#M1157</link>
    <description>&lt;P&gt;Those are some very interesting options, Sanjay - thanks.&lt;/P&gt;&lt;P&gt;I am thinking more of having a data set where you have random access to any record or set of records in the entire set, so the batch method would not apply here, although in some cases it would work really well.&lt;/P&gt;&lt;P&gt;The AWS or third party storage is an interesting idea, for sure.&lt;/P&gt;</description>
    <pubDate>Fri, 18 Mar 2022 02:59:04 GMT</pubDate>
    <dc:creator>kolson</dc:creator>
    <dc:date>2022-03-18T02:59:04Z</dc:date>
    <item>
      <title>Cache millions of records</title>
      <link>https://systematic.workato.com/t5/workato-pros-discussion-board/cache-millions-of-records/m-p/2568#M1150</link>
      <description>&lt;P&gt;If you need to cache a large amount of data for subsequent retrieval in a recipe, how would you do it?  Would you put it in a lookup table (limit 100K rows) or in a list of hashes?   &lt;/P&gt;&lt;P&gt;Is there a limit on the number of rows in a list?  &lt;/P&gt;&lt;P&gt;I am wondering how to cache a million or more records.&lt;/P&gt;</description>
      <pubDate>Thu, 17 Mar 2022 22:23:34 GMT</pubDate>
      <guid>https://systematic.workato.com/t5/workato-pros-discussion-board/cache-millions-of-records/m-p/2568#M1150</guid>
      <dc:creator>kolson</dc:creator>
      <dc:date>2022-03-17T22:23:34Z</dc:date>
    </item>
    <item>
      <title>Re: Cache millions of records</title>
      <link>https://systematic.workato.com/t5/workato-pros-discussion-board/cache-millions-of-records/m-p/2569#M1151</link>
      <description>&lt;P&gt;I see that this is not possible in Workato.&lt;/P&gt;</description>
      <pubDate>Thu, 17 Mar 2022 22:51:27 GMT</pubDate>
      <guid>https://systematic.workato.com/t5/workato-pros-discussion-board/cache-millions-of-records/m-p/2569#M1151</guid>
      <dc:creator>kolson</dc:creator>
      <dc:date>2022-03-17T22:51:27Z</dc:date>
    </item>
    <item>
      <title>Re: Cache millions of records</title>
      <link>https://systematic.workato.com/t5/workato-pros-discussion-board/cache-millions-of-records/m-p/2570#M1152</link>
      <description>&lt;DIV dir="ltr"&gt;I'm not sure what the row limit is on collections, but you might want to look at those,&amp;nbsp;they also allow you to use SQL to query against them and have more columns than lookup&amp;nbsp;tables.&lt;BR /&gt;They do not persist outside of the running recipe though, whereas lookup tables do.&lt;BR /&gt;&lt;BR /&gt;I know this isn't helpful but I'd probably just look at redesigning something so you don't have to cache 1M+ rows in your middleware.&lt;BR /&gt;&lt;BR /&gt;&lt;/DIV&gt;&lt;BR /&gt;&lt;SPAN&gt;--&lt;/SPAN&gt;</description>
      <pubDate>Thu, 17 Mar 2022 22:56:58 GMT</pubDate>
      <guid>https://systematic.workato.com/t5/workato-pros-discussion-board/cache-millions-of-records/m-p/2570#M1152</guid>
      <dc:creator>anthony-oconnor</dc:creator>
      <dc:date>2022-03-17T22:56:58Z</dc:date>
    </item>
    <item>
      <title>Re: Cache millions of records</title>
      <link>https://systematic.workato.com/t5/workato-pros-discussion-board/cache-millions-of-records/m-p/2571#M1153</link>
      <description>&lt;P&gt;Agreed it is not optimal, but some use cases would require it.   &lt;/P&gt;&lt;P&gt;There are some cases where you would might need to have a large dataset of records, say from Workday, to compare to what currently is in a legacy DB, and doing repeated calls across the network to the database is slow, or maybe there are requirements for a snapshot of a continually changing source to compare to another large dataset.   So it may not be the common case, but you might need to do an ETL solution requiring temporary storage of a large dataset.   It is not *always* indicative of bad design.&lt;/P&gt;&lt;P&gt;The limit on the size of a collection is 50K records, BTW.&lt;/P&gt;</description>
      <pubDate>Thu, 17 Mar 2022 23:37:33 GMT</pubDate>
      <guid>https://systematic.workato.com/t5/workato-pros-discussion-board/cache-millions-of-records/m-p/2571#M1153</guid>
      <dc:creator>kolson</dc:creator>
      <dc:date>2022-03-17T23:37:33Z</dc:date>
    </item>
    <item>
      <title>Re: Cache millions of records</title>
      <link>https://systematic.workato.com/t5/workato-pros-discussion-board/cache-millions-of-records/m-p/2572#M1154</link>
      <description>&lt;P&gt;How about storing these records onto one or multiple Pub/Sub topic(s)?&lt;/P&gt;&lt;P&gt;And then have your secondary recipe(s) read from those to process them.&lt;/P&gt;&lt;BR /&gt;&lt;P&gt;Along with that we also use a SQL server to deal with these larger quantities.&lt;/P&gt;</description>
      <pubDate>Thu, 17 Mar 2022 23:43:46 GMT</pubDate>
      <guid>https://systematic.workato.com/t5/workato-pros-discussion-board/cache-millions-of-records/m-p/2572#M1154</guid>
      <dc:creator>steven-marissen</dc:creator>
      <dc:date>2022-03-17T23:43:46Z</dc:date>
    </item>
    <item>
      <title>Re: Cache millions of records</title>
      <link>https://systematic.workato.com/t5/workato-pros-discussion-board/cache-millions-of-records/m-p/2573#M1155</link>
      <description>&lt;P&gt;That is a good suggestion, Steven, for some use cases.   However, I am referring to cases where the entire dataset is being used as a whole - like for joins or comparisons with another related dataset all at once.&lt;/P&gt;&lt;P&gt;So it is more of an ETL type process vs a streaming or record-by-record process.&lt;/P&gt;</description>
      <pubDate>Thu, 17 Mar 2022 23:46:51 GMT</pubDate>
      <guid>https://systematic.workato.com/t5/workato-pros-discussion-board/cache-millions-of-records/m-p/2573#M1155</guid>
      <dc:creator>kolson</dc:creator>
      <dc:date>2022-03-17T23:46:51Z</dc:date>
    </item>
    <item>
      <title>Re: Cache millions of records</title>
      <link>https://systematic.workato.com/t5/workato-pros-discussion-board/cache-millions-of-records/m-p/2574#M1156</link>
      <description>&lt;P&gt;1. If you would retain data across different Jobs.. you need to use external storage(e.g. AWS S3 / MySql etc.) &lt;/P&gt;&lt;P&gt;2. In case of cache and use large data within a single Job executinon...&lt;/P&gt;&lt;P&gt;- You can use Collection (First, prepare the list of raw data and then create the collection from that list). Collection is faster to retrieve the records and easy to understand. I have used this approach for the 100K rows.&lt;/P&gt;&lt;P&gt;Yes, it's limit of 50k records in single query... but by using LIMIT Offset we could get next bunch of records easily. &lt;/P&gt;</description>
      <pubDate>Fri, 18 Mar 2022 01:25:03 GMT</pubDate>
      <guid>https://systematic.workato.com/t5/workato-pros-discussion-board/cache-millions-of-records/m-p/2574#M1156</guid>
      <dc:creator>sanjay-rathod</dc:creator>
      <dc:date>2022-03-18T01:25:03Z</dc:date>
    </item>
    <item>
      <title>Re: Cache millions of records</title>
      <link>https://systematic.workato.com/t5/workato-pros-discussion-board/cache-millions-of-records/m-p/2575#M1157</link>
      <description>&lt;P&gt;Those are some very interesting options, Sanjay - thanks.&lt;/P&gt;&lt;P&gt;I am thinking more of having a data set where you have random access to any record or set of records in the entire set, so the batch method would not apply here, although in some cases it would work really well.&lt;/P&gt;&lt;P&gt;The AWS or third party storage is an interesting idea, for sure.&lt;/P&gt;</description>
      <pubDate>Fri, 18 Mar 2022 02:59:04 GMT</pubDate>
      <guid>https://systematic.workato.com/t5/workato-pros-discussion-board/cache-millions-of-records/m-p/2575#M1157</guid>
      <dc:creator>kolson</dc:creator>
      <dc:date>2022-03-18T02:59:04Z</dc:date>
    </item>
    <item>
      <title>Re: Cache millions of records</title>
      <link>https://systematic.workato.com/t5/workato-pros-discussion-board/cache-millions-of-records/m-p/2576#M1158</link>
      <description>&lt;P&gt;Workato is working on a Cloud DB persistence layer that may be useful for this use case.  Not sure of the details when they will roll this new feature out nor the details of how it will work and any limitations.  I would think it certainly could handle millions of records.  Perhaps someone from the Workato Product team can comment (Konstantin perhaps).&lt;/P&gt;</description>
      <pubDate>Tue, 22 Mar 2022 19:23:32 GMT</pubDate>
      <guid>https://systematic.workato.com/t5/workato-pros-discussion-board/cache-millions-of-records/m-p/2576#M1158</guid>
      <dc:creator>gwilkinson</dc:creator>
      <dc:date>2022-03-22T19:23:32Z</dc:date>
    </item>
    <item>
      <title>Re: Cache millions of records</title>
      <link>https://systematic.workato.com/t5/workato-pros-discussion-board/cache-millions-of-records/m-p/4074#M1876</link>
      <description>&lt;P&gt;I know this is very late to this discussion, but just for future people looking for such a solution.&amp;nbsp; It seems to me that this is something huge that Workato is missing.&amp;nbsp; The ability to save off "state" of one kind of another in some type of cache and then re-use that again with some type of timeout.&lt;/P&gt;&lt;P&gt;So if it cannot be done in Workato, it certainly could be done in a number of different ways via current cloud technologies:&lt;/P&gt;&lt;P&gt;Basically use something like ElasticCache, Redis, AWS DynamoDB, MongoDB to package up anything you want to cache and then save it in the cache.&amp;nbsp; Then the next time you run the recipe, check the cache to see if that data is populated (and not expired).&amp;nbsp; This could save a lot of steps in the workflow, but of course would require at least one new step to check the cache (and to save off your data in the cache).&lt;/P&gt;&lt;P&gt;If anyone is interested in such a system I would love to architect an API that would make this super easy for any Workato recipe.&amp;nbsp; It could even incorporate encryption so that no one could see your data in the cache, or of course be setup specifically for your company to be segregated properly.&lt;/P&gt;&lt;P&gt;Would be a great AWS project so that it is super fast, secure and scalable.&lt;/P&gt;&lt;P&gt;Patrick&lt;/P&gt;</description>
      <pubDate>Thu, 16 Mar 2023 19:12:46 GMT</pubDate>
      <guid>https://systematic.workato.com/t5/workato-pros-discussion-board/cache-millions-of-records/m-p/4074#M1876</guid>
      <dc:creator>patrick-steil</dc:creator>
      <dc:date>2023-03-16T19:12:46Z</dc:date>
    </item>
    <item>
      <title>Re: Cache millions of records</title>
      <link>https://systematic.workato.com/t5/workato-pros-discussion-board/cache-millions-of-records/m-p/4091#M1884</link>
      <description>&lt;P&gt;Hey &lt;a href="https://systematic.workato.com/t5/user/viewprofilepage/user-id/5314"&gt;@patrick-steil&lt;/a&gt;,&lt;/P&gt;&lt;P&gt;&lt;SPAN&gt;We at Workato have recently released two features that looks like a good fit with this specific ETL use case.&lt;/SPAN&gt;&lt;/P&gt;&lt;P&gt;&lt;SPAN&gt;Workato FileStorage is a persistent storage system within Workato that can be used to store large volume data (limitless number of rows in the order of millions) and this data can be fetched and used across recipes.&lt;/SPAN&gt;&lt;/P&gt;&lt;P&gt;&lt;SPAN&gt;Second, we have the new SQL Transformations utility connector that can operate on CSV data stored within FileStorage and perform transformations (compare with other data sets, live application data extract, or other CSV files), etc.&lt;/SPAN&gt;&lt;/P&gt;&lt;P&gt;&lt;SPAN&gt;By using both these features, you can store any volume data as CSV within Workato and run queries on them. More details on the two feature are available here -&amp;nbsp;&lt;/SPAN&gt;&lt;A class="" href="https://docs.workato.com/features/workato-files.html#workato-filestorage" target="_blank" rel="noopener noreferrer"&gt;FileStorage&lt;/A&gt;&lt;SPAN&gt;,&amp;nbsp;&lt;/SPAN&gt;&lt;A class="" href="https://docs.workato.com/features/sql-transformations.html#sql-transformations" target="_blank" rel="noopener noreferrer"&gt;SQL Transformations&lt;/A&gt;&lt;SPAN&gt;.&lt;/SPAN&gt;&lt;/P&gt;&lt;P&gt;The features are premium and are in beta stage, so reach out to Workato's customer support to get more details.&lt;/P&gt;&lt;P&gt;Cheers!&lt;BR /&gt;Meghan&lt;/P&gt;</description>
      <pubDate>Tue, 21 Mar 2023 19:13:30 GMT</pubDate>
      <guid>https://systematic.workato.com/t5/workato-pros-discussion-board/cache-millions-of-records/m-p/4091#M1884</guid>
      <dc:creator>meghan-legaspi</dc:creator>
      <dc:date>2023-03-21T19:13:30Z</dc:date>
    </item>
  </channel>
</rss>

