Uploaded image for project: 'Spark'
  1. Spark
  2. SPARK-18946

treeAggregate will be low effficiency when aggregate high dimension vectors in ML algorithm

    XMLWordPrintableJSON

Details

    • Improvement
    • Status: Resolved
    • Major
    • Resolution: Incomplete
    • None
    • None
    • ML, MLlib

    Description

      In many machine learning algorithms, we have to treeAggregate large vectors/arrays due to the large number of features. Unfortunately, the treeAggregate operation of RDD will be low efficiency when the dimension of vectors/arrays is bigger than million. Because high dimension of vector/array always occupy more than 100MB Memory, transferring a 100MB element among executors is pretty low efficiency in Spark.

      Attachments

        Activity

          People

            Unassigned Unassigned
            zunwenyou zunwen you
            Votes:
            0 Vote for this issue
            Watchers:
            3 Start watching this issue

            Dates

              Created:
              Updated:
              Resolved: