Details
Description
Reading from an ORC table which contains char/varchar columns can fail if the table has been created using Spark. This is caused by the fact that spark internally replaces char and varchar columns with a string column, this causes the ORC reader to use the wrong reader, and that eventually causes a ClassCastException.
Attachments
Issue Links
- is duplicated by
-
SPARK-19430 Cannot read external tables with VARCHAR columns if they're backed by ORC files written by Hive 1.2.1
-
- Resolved
-
-
SPARK-20515 Issue with reading Hive ORC tables having char/varchar columns in Spark SQL
-
- Resolved
-
- is related to
-
SPARK-21997 Spark shows different results on char/varchar columns on Parquet
-
- Reopened
-
-
SPARK-23774 `Cast` to CHAR/VARCHAR should truncate the values
-
- Resolved
-
- relates to
-
SPARK-20901 Feature parity for ORC with Parquet
-
- Open
-
- links to