Details
-
Bug
-
Status: Resolved
-
Major
-
Resolution: Won't Fix
-
2.1.0, 2.1.1
-
None
-
None
Description
Spark returning empty result from when querying remote hadoop cluster.
All firewall settings removed.
Querying using JDBC working properly using hive-jdbc driver from version 1.1.1
Code snippet is:
val spark = SparkSession.builder .appName("RemoteSparkTest") .master("local") .getOrCreate() val df = spark.read .option("url", "jdbc:hive2://remote.hive.local:10000/default") .option("user", "user") .option("password", "pass") .option("dbtable", "test_table") .option("driver", "org.apache.hive.jdbc.HiveDriver") .format("jdbc") .load() df.show()
Result:
+-------------------+ |test_table.test_col| +-------------------+ +-------------------+
All manipulations like:
df.select(*).show()
returns empty result too.
Attachments
Issue Links
- is duplicated by
-
SPARK-31457 spark jdbc read hive created the wrong PreparedStatement
- Resolved
- relates to
-
SPARK-22016 Add HiveDialect for JDBC connection to Hive
- Resolved