Uploaded image for project: 'Spark'
  1. Spark
  2. SPARK-12892

Support plugging in Spark scheduler

Log workAgile BoardRank to TopRank to BottomAttach filesAttach ScreenshotBulk Copy AttachmentsBulk Move AttachmentsVotersWatch issueWatchersCreate sub-taskConvert to sub-taskMoveLinkCloneLabelsUpdate Comment AuthorReplace String in CommentUpdate Comment VisibilityDelete CommentsDelete
    XMLWordPrintableJSON

Details

    • Improvement
    • Status: Resolved
    • Major
    • Resolution: Duplicate
    • None
    • None
    • Spark Core
    • None

    Description

      Currently the only supported cluster schedulers are standalone, Mesos, Yarn and Simr. However if users like to build a new one it must be merged back into main, and might not be desirable for Spark and hard to iterate.
      Instead, we should make a plugin architecture possible so that when users like to integrate with new scheduler it can plugged in via configuration and runtime loading instead.

      Attachments

        Issue Links

        Activity

          This comment will be Viewable by All Users Viewable by All Users
          Cancel

          People

            Unassigned Unassigned Assign to me
            tnachen Timothy Chen
            Votes:
            0 Vote for this issue
            Watchers:
            1 Start watching this issue

            Dates

              Created:
              Updated:
              Resolved:

              Slack

                Issue deployment