Uploaded image for project: 'Apache AsterixDB'
  1. Apache AsterixDB
  2. ASTERIXDB-1433

Multiple cores with huge memory slow down in the big fact table aggregation.

VotersWatch issueWatchersLinkCloneUpdate Comment AuthorReplace String in CommentUpdate Comment VisibilityDelete Comments
    XMLWordPrintableJSON

Details

    • Improvement
    • Status: Closed
    • Major
    • Resolution: Fixed
    • None
    • None
    • None
    • 10 nodes X Linux ubuntu/6 cpu X 4 cores/per cpu, 128 GB memory/per node.

    Description

      This is a classic hardware platform that shoes up the TB scale of dataset in total. AsterixDB does extremely well for the complex query that includes multiple join operators over a high-selectivity select operator. However, the running trace results demonstrate that, as compared to the big memory configurations, the original tables is always re-loaded from the disk to the actual memory even they have been handled in the latest query. To this end, why not provide the strategy to keep the intermediate data of the last completed query into the memory and free them in case the memory is not enough for the newly query. In some case, the user will always trigger the query with the different parameters on the same tables, for example, the variant-parameter aggregation on the single big fact table.

      Attachments

        Activity

          This comment will be Viewable by All Users Viewable by All Users
          Cancel

          People

            mhubail Murtadha Makki Al Hubail
            lwhay Wenhai Li
            Votes:
            0 Vote for this issue
            Watchers:
            2 Start watching this issue

            Dates

              Created:
              Updated:
              Resolved:

              Slack

                Issue deployment