Details
-
Bug
-
Status: Resolved
-
Major
-
Resolution: Fixed
-
2.3.1, 2.4.3
Description
I have two DataFrames, A and B. From B, I have derived two additional DataFrames, B1 and B2. When joining A to B1 and B2, I'm getting a very confusing error:
Join condition is missing or trivial.
Either: use the CROSS JOIN syntax to allow cartesian products between these
relations, or: enable implicit cartesian products by setting the configuration
variable spark.sql.crossJoin.enabled=true;
Then, when IĀ configure "spark.sql.crossJoin.enabled=true" as instructed, Spark appears to give me incorrect answers.
I am not sure if I am missing something obvious, or if there is some kind of bug here. The "join condition is missing" error is confusing and doesn't make sense to me, and the seemingly incorrect output is concerning.
I've attached a reproduction, along with the output I'm seeing with and without the implicit cross join enabled.
I realize the join I've written is not "correct" in the sense that it should be left outer join instead of an inner join (since some of the aggregates are not available for all states), but that doesn't explain Spark's behavior.
Attachments
Attachments
Issue Links
- is duplicated by
-
SPARK-26231 Dataframes inner join on double datatype columns resulting in Cartesian product
- Resolved
- relates to
-
SPARK-20804 Join with null safe equality fails with AnalysisException
- Closed
-
SPARK-6459 Warn when Column API is constructing trivially true equality
- Resolved
- links to