Details
-
Improvement
-
Status: Closed
-
Major
-
Resolution: Fixed
-
None
-
None
-
None
Description
Currently, during IPA we collect all variables (scalars & matrices) eligible for propagation across blocks (i.e. not updated in block), and then propagate the only the matrix sizes across the blocks. It seems plausible that we could also replace all eligible scalar transient reads with literals based on the variables that have already been collected. The benefit is that many ops will be able to determine their respective output sizes during regular compilation, instead of having to wait until dynamic recompilation, and thus we can reduce the pressure on dynamic recompilation.
Are there drawbacks to this approach? The use case is that I was seeing a large number of memory warnings while training a convolutional net due to the sizes being unknown during regular compilation, yet the engine only having CP versions of the ops. Additionally, I was running into actual heap space OOM errors for situations that should not run out of memory, and thus I started exploring.
I've attached an example script and the explain plan (hops & runtime) w/ and w/o the IPA scalar replacement.
Attachments
Attachments
Issue Links
- is blocked by
-
SYSTEMDS-1575 DataType Change Test Failure
- Closed
- is related to
-
SYSTEMDS-1555 Decouple literal replacement from in-place recompilation
- Closed
-
SYSTEMDS-1566 Possible regression from 0.13 -> 0.14 for MNIST LeNet script
- Closed
-
SYSTEMDS-1466 Update `convnet.dml` to use distributed SGD.
- In Progress
-
SYSTEMDS-1561 Improve constant folding during compilation
- Closed
- relates to
-
SYSTEMDS-540 Deep Learning
- In Progress
-
SYSTEMDS-1185 SystemML Breast Cancer Project
- Resolved
-
SYSTEMDS-427 Extended inter-procedure analysis (constant propagation)
- Open