<?xml version="1.0" encoding="UTF-8"?>
<rss xmlns:content="http://purl.org/rss/1.0/modules/content/" xmlns:dc="http://purl.org/dc/elements/1.1/" xmlns:rdf="http://www.w3.org/1999/02/22-rdf-syntax-ns#" xmlns:taxo="http://purl.org/rss/1.0/modules/taxonomy/" version="2.0">
  <channel>
    <title>topic Re: Upper limit on datasets for a table ( and how are the processed ) ? in Administrative Discussions</title>
    <link>https://community.incorta.com/t5/administrative-discussions/upper-limit-on-datasets-for-a-table-and-how-are-the-processed/m-p/4568#M168</link>
    <description>&lt;P&gt;The multi-source table is a good solution in this case, but is there a reason you must extract them in separate Incorta tables then glue them using a multi-source table through SQLi? I believe it will be better if you create a multi-source table where each dataset extracts from source directly, this way, you reduce one step (re-extraction through SQLi) which will save cluster resources.&lt;/P&gt;&lt;P&gt;As for the limit of number of data sets, I believe there's no limit. I have seen customer use cases where there were 25+ data sets in a table and I've heard of 60. Technically there's no limit. But you have to make sure the loader service has enough resources to execute.&lt;/P&gt;&lt;P&gt;When you load a table with multiple data sets, the loader will try to extract the data set in parallel. This is governed by a thread pool for all data sets being extracted across the load job, which may contain other tables. So, expect parallel execution of the queries against your source.&lt;/P&gt;&lt;P&gt;In the CMC, you can set the number of parallel load jobs, but this is for the whole load jobs (the ones you set in the scheduler), which are usually schemas. The capacity of the thread-pool used to extract tables and data sets depends on the CPU Utilization percentage you set for the loader service. The higher the percentage, the more tables (and data sets) can be extracted in parallel.&lt;/P&gt;</description>
    <pubDate>Sun, 25 Jun 2023 07:41:41 GMT</pubDate>
    <dc:creator>AhmedM</dc:creator>
    <dc:date>2023-06-25T07:41:41Z</dc:date>
    <item>
      <title>Upper limit on datasets for a table ( and how are the processed ) ?</title>
      <link>https://community.incorta.com/t5/administrative-discussions/upper-limit-on-datasets-for-a-table-and-how-are-the-processed/m-p/4529#M163</link>
      <description>&lt;P&gt;I have a scenario in front of me where I have a bunch of identical ( columns ) Incorta tables -- i.e. I have a table for each month of the year for the past 2+ years which I want to consolidate into one big table w/ all the data inside.&amp;nbsp;&lt;/P&gt;&lt;P&gt;Monthly tables:&amp;nbsp; &amp;nbsp;Pulled from original source.&amp;nbsp; Broken into monthly tables because they return exponentially faster that way.&lt;/P&gt;&lt;P&gt;Consolidated table:&amp;nbsp; IncortaSQL&amp;nbsp; &amp;nbsp;apparently does not support doing a big UNION or UNION ALL&amp;nbsp; and we were getting memory errors using SQLi with the big UNION statement.&amp;nbsp; &amp;nbsp;&lt;/P&gt;&lt;P&gt;What *IS* working is having all the month tables as separate datasets via SQLi - and it's pretty fast!&amp;nbsp; &amp;nbsp;&lt;/P&gt;&lt;P&gt;That said, we must remain optimistic and assume the calendar will not stop.&amp;nbsp;&lt;/P&gt;&lt;P&gt;So we will need to keep adding month tables as we go ( we have 2023 covered, w/ July forward obvs not having any data )&amp;nbsp;&lt;/P&gt;&lt;P&gt;So - finally - the question:&amp;nbsp; Is there a limit to the number of datasets which Incorta will support?&amp;nbsp; &amp;nbsp;&lt;/P&gt;&lt;P&gt;Corollary question - how are these processed?&amp;nbsp; Does Incorta send 36 requests ( to itself in this case ) at the same time?&amp;nbsp; Is that governed by the parallel job tasks config in the CMC?&amp;nbsp; &amp;nbsp;I can't imagine they are all done in series, although there must be some staging somewhere w/in the mix because I am having Incorta enforce primary key constraints.&amp;nbsp;&lt;/P&gt;&lt;P&gt;I'm very excited about this - give me some good news Community members!&amp;nbsp; &lt;span class="lia-unicode-emoji" title=":winking_face:"&gt;😉&lt;/span&gt;&amp;nbsp; &amp;nbsp;&lt;/P&gt;&lt;P&gt;Screen cap - 3 of 30! datasets&amp;nbsp; ( 2021 combines Jan-Jun as they were relatively low volume months )&amp;nbsp;&amp;nbsp;&lt;/P&gt;&lt;P&gt;&lt;span class="lia-inline-image-display-wrapper lia-image-align-inline" image-alt="RADSr_0-1687458262087.png" style="width: 400px;"&gt;&lt;img src="https://community.incorta.com/t5/image/serverpage/image-id/2272i53EC373B3ED60E3D/image-size/medium?v=v2&amp;amp;px=400" role="button" title="RADSr_0-1687458262087.png" alt="RADSr_0-1687458262087.png" /&gt;&lt;/span&gt;&lt;/P&gt;&lt;P&gt;&amp;nbsp;&lt;/P&gt;&lt;P&gt;&amp;nbsp;&amp;nbsp;&lt;/P&gt;</description>
      <pubDate>Thu, 22 Jun 2023 18:27:57 GMT</pubDate>
      <guid>https://community.incorta.com/t5/administrative-discussions/upper-limit-on-datasets-for-a-table-and-how-are-the-processed/m-p/4529#M163</guid>
      <dc:creator>RADSr</dc:creator>
      <dc:date>2023-06-22T18:27:57Z</dc:date>
    </item>
    <item>
      <title>Re: Upper limit on datasets for a table ( and how are the processed ) ?</title>
      <link>https://community.incorta.com/t5/administrative-discussions/upper-limit-on-datasets-for-a-table-and-how-are-the-processed/m-p/4555#M164</link>
      <description>&lt;P&gt;I'll answer the corrolary question.&amp;nbsp;&lt;BR /&gt;&lt;BR /&gt;The "staging" or primary key enforcement happens during the load/post-load phase so there should be parallelization across the different datasets and the configuration of the loader node (# of CPUs dedicated to loader, etc.) and parallelization in CMC should drive that.&lt;BR /&gt;&lt;BR /&gt;However, because the queries themselves are hitting the Incorta Analytics service....the SQL interface/Analytics node could decide to queue up query execution on that end even though the client aka its own loader service may have issued a set of queries (8, 12, 16, whatever) in parallel.&lt;BR /&gt;&lt;BR /&gt;I don't know of any limit on datasets per table (probably a theoretical one?) so maybe someone from Engineering can weigh in.&amp;nbsp; Fun fact: the first Proof of Concept I ever did&amp;nbsp;@ Incorta (2017?) was 50 different manufacturing instances (1 per factory) that we poured into a single schema/table structure so it's definitely built since the beginning to handle this type of use case.&lt;/P&gt;</description>
      <pubDate>Fri, 23 Jun 2023 15:55:16 GMT</pubDate>
      <guid>https://community.incorta.com/t5/administrative-discussions/upper-limit-on-datasets-for-a-table-and-how-are-the-processed/m-p/4555#M164</guid>
      <dc:creator>PatrickRafferty</dc:creator>
      <dc:date>2023-06-23T15:55:16Z</dc:date>
    </item>
    <item>
      <title>Re: Upper limit on datasets for a table ( and how are the processed ) ?</title>
      <link>https://community.incorta.com/t5/administrative-discussions/upper-limit-on-datasets-for-a-table-and-how-are-the-processed/m-p/4559#M165</link>
      <description>&lt;P&gt;And here's the parallelization in action with OnlineStore (quickest example I could pull up):&lt;BR /&gt;&lt;BR /&gt;&lt;/P&gt;&lt;P&gt;So, query the largest table (salesorderdetail) 3 different ways making the 9 million row table into a 27 million row table.&lt;/P&gt;&lt;P&gt;&lt;span class="lia-inline-image-display-wrapper lia-image-align-inline" image-alt="PatrickRafferty_0-1687535841918.png" style="width: 400px;"&gt;&lt;img src="https://community.incorta.com/t5/image/serverpage/image-id/2280i7D3BC1A0A81E560E/image-size/medium?v=v2&amp;amp;px=400" role="button" title="PatrickRafferty_0-1687535841918.png" alt="PatrickRafferty_0-1687535841918.png" /&gt;&lt;/span&gt;&lt;/P&gt;&lt;P&gt;And that extracts in 41 seconds:&lt;/P&gt;&lt;P&gt;&lt;span class="lia-inline-image-display-wrapper lia-image-align-inline" image-alt="PatrickRafferty_1-1687535912898.png" style="width: 400px;"&gt;&lt;img src="https://community.incorta.com/t5/image/serverpage/image-id/2281i59DBD9A86C8FF98C/image-size/medium?v=v2&amp;amp;px=400" role="button" title="PatrickRafferty_1-1687535912898.png" alt="PatrickRafferty_1-1687535912898.png" /&gt;&lt;/span&gt;&lt;/P&gt;&lt;P&gt;And a single query to that table extracts in 30 seconds:&lt;/P&gt;&lt;P&gt;&lt;span class="lia-inline-image-display-wrapper lia-image-align-inline" image-alt="PatrickRafferty_2-1687535990684.png" style="width: 400px;"&gt;&lt;img src="https://community.incorta.com/t5/image/serverpage/image-id/2282i3F2ABF3387E3E272/image-size/medium?v=v2&amp;amp;px=400" role="button" title="PatrickRafferty_2-1687535990684.png" alt="PatrickRafferty_2-1687535990684.png" /&gt;&lt;/span&gt;&lt;/P&gt;&lt;P&gt;&amp;nbsp;&lt;/P&gt;</description>
      <pubDate>Fri, 23 Jun 2023 16:00:06 GMT</pubDate>
      <guid>https://community.incorta.com/t5/administrative-discussions/upper-limit-on-datasets-for-a-table-and-how-are-the-processed/m-p/4559#M165</guid>
      <dc:creator>PatrickRafferty</dc:creator>
      <dc:date>2023-06-23T16:00:06Z</dc:date>
    </item>
    <item>
      <title>Re: Upper limit on datasets for a table ( and how are the processed ) ?</title>
      <link>https://community.incorta.com/t5/administrative-discussions/upper-limit-on-datasets-for-a-table-and-how-are-the-processed/m-p/4561#M166</link>
      <description>&lt;P&gt;&lt;a href="https://community.incorta.com/t5/user/viewprofilepage/user-id/24"&gt;@RADSr&lt;/a&gt;&amp;nbsp; I am assuming your historical data (data for previous months) is also getting updated that's why you are doing full load where you load data for all the months using the approach you described.&lt;/P&gt;&lt;P&gt;If that is not the case and data for previous months do not change, then there are other ways to manage this load.&amp;nbsp;&lt;/P&gt;</description>
      <pubDate>Fri, 23 Jun 2023 20:18:32 GMT</pubDate>
      <guid>https://community.incorta.com/t5/administrative-discussions/upper-limit-on-datasets-for-a-table-and-how-are-the-processed/m-p/4561#M166</guid>
      <dc:creator>anurag</dc:creator>
      <dc:date>2023-06-23T20:18:32Z</dc:date>
    </item>
    <item>
      <title>Re: Upper limit on datasets for a table ( and how are the processed ) ?</title>
      <link>https://community.incorta.com/t5/administrative-discussions/upper-limit-on-datasets-for-a-table-and-how-are-the-processed/m-p/4568#M168</link>
      <description>&lt;P&gt;The multi-source table is a good solution in this case, but is there a reason you must extract them in separate Incorta tables then glue them using a multi-source table through SQLi? I believe it will be better if you create a multi-source table where each dataset extracts from source directly, this way, you reduce one step (re-extraction through SQLi) which will save cluster resources.&lt;/P&gt;&lt;P&gt;As for the limit of number of data sets, I believe there's no limit. I have seen customer use cases where there were 25+ data sets in a table and I've heard of 60. Technically there's no limit. But you have to make sure the loader service has enough resources to execute.&lt;/P&gt;&lt;P&gt;When you load a table with multiple data sets, the loader will try to extract the data set in parallel. This is governed by a thread pool for all data sets being extracted across the load job, which may contain other tables. So, expect parallel execution of the queries against your source.&lt;/P&gt;&lt;P&gt;In the CMC, you can set the number of parallel load jobs, but this is for the whole load jobs (the ones you set in the scheduler), which are usually schemas. The capacity of the thread-pool used to extract tables and data sets depends on the CPU Utilization percentage you set for the loader service. The higher the percentage, the more tables (and data sets) can be extracted in parallel.&lt;/P&gt;</description>
      <pubDate>Sun, 25 Jun 2023 07:41:41 GMT</pubDate>
      <guid>https://community.incorta.com/t5/administrative-discussions/upper-limit-on-datasets-for-a-table-and-how-are-the-processed/m-p/4568#M168</guid>
      <dc:creator>AhmedM</dc:creator>
      <dc:date>2023-06-25T07:41:41Z</dc:date>
    </item>
    <item>
      <title>Re: Upper limit on datasets for a table ( and how are the processed ) ?</title>
      <link>https://community.incorta.com/t5/administrative-discussions/upper-limit-on-datasets-for-a-table-and-how-are-the-processed/m-p/4585#M171</link>
      <description>&lt;P&gt;TL/DR - making these modular will give me more flexibility and increase my ability to move corrective changes into the most used dashboards in the shortest time possible.&amp;nbsp;&amp;nbsp;&lt;/P&gt;&lt;P&gt;------------&lt;/P&gt;&lt;P&gt;That said, the primary reason for staging into Incorta first is to meter/limit the activity against the source DB .&amp;nbsp; &amp;nbsp; The source is a view on a production box.&amp;nbsp; The view is ( I gather ) pretty heavy and is completely "black box" to me, so I can't do any tuning or even recreate it w/in Incorta.&amp;nbsp; &amp;nbsp;The short is, because it's production I don't want to be a burden on the system which may affect operations.&lt;/P&gt;&lt;P&gt;Currently I have three schemae ( 2021, 2022, 2023 with 7, 12, and12 tables respectively&amp;nbsp; ) which I can run in series or independently and which can be further broken down if need be.&lt;/P&gt;&lt;P&gt;So - use case:&amp;nbsp; A change is made to move a product to a different category and that change needs to be reflected ASAP for the WTD/MTD/QTD/YTD reporting I can choose to run only the tables necessary for the immediate need *now* and be running correct dashboards very quickly while correcting further history at my schedule.&amp;nbsp;&lt;/P&gt;&lt;P&gt;&amp;nbsp;&lt;/P&gt;&lt;P&gt;&amp;nbsp;&lt;/P&gt;</description>
      <pubDate>Tue, 27 Jun 2023 23:41:21 GMT</pubDate>
      <guid>https://community.incorta.com/t5/administrative-discussions/upper-limit-on-datasets-for-a-table-and-how-are-the-processed/m-p/4585#M171</guid>
      <dc:creator>RADSr</dc:creator>
      <dc:date>2023-06-27T23:41:21Z</dc:date>
    </item>
  </channel>
</rss>

