Details
-
Improvement
-
Status: Closed
-
Major
-
Resolution: Fixed
-
None
-
None
-
None
Description
More likely than not, a user's job may depend on multiple jars.
Right now, when submitting a job through bin/hadoop, there is no way for the user to specify that.
A walk around for that is to re-package all the dependent jars into a new jar or put the dependent jar files in the lib dir of the new jar.
This walk around causes unnecessary inconvenience to the user. Furthermore, if the user does not own the main function
(like the case when the user uses Aggregate, or datajoin, streaming), the user has to re-package those system jar files too.
It is much desired that hadoop provides a clean and simple way for the user to specify a list of dependent jar files at the time
of job submission. Someting like:
bin/hadoop .... --depending_jars j1.jar:j2.jar
Attachments
Attachments
Issue Links
- is duplicated by
-
HADOOP-366 Should be able to specify more than one jar into a JobConf file
- Closed
- is related to
-
MAPREDUCE-574 Fix -file option in Streaming to use Distributed Cache
- Resolved