As a pyspark user, I would like to read/write hudi datasets using pyspark.
There are several components to achieve this goal.
- Create a hudi-pyspark package that users can import and start reading/writing hudi datasets.
- Explain how to read/write hudi datasets using pyspark in a blog post/documentation.
- Add the hudi-pyspark module to the hudi demo docker along with the instructions.
- Make the package available as part of the spark packages index and python package index
hudi-pyspark packages should implement HUDI data source API for Apache Spark using which HUDI files can be read as DataFrame and write to any Hadoop supported file system.
Usage pattern after we launch this feature should be something like this:
Install the package using:
Include hudi-pyspark package in your Spark Applications using:
spark-shell, pyspark, or spark-submit