Details
-
Bug
-
Status: Closed
-
Major
-
Resolution: Fixed
-
3.0.0
Description
Currently we only check that the max number of entries in the hashmap for a MapJoin surpasses a certain threshold to decide whether to execute a dynamically partitioned hash join.
We would like to factor the size of the large input that we will shuffle for the dynamically partitioned hash join into the cost model too.
Attachments
Attachments
Issue Links
- breaks
-
HIVE-17464 Fix to be able to disable max shuffle size DHJ config
- Closed
- relates to
-
HIVE-24308 FIX conditions used for DPHJ conversion
- Open
-
HIVE-12492 MapJoin: 4 million unique integers seems to be a probe plateau
- Closed