I have a question as to what is stored within the power BI service storage space and what is processed in the memory but not stored.
I'm in the process of creating several large sequencial dataflows,
The first dataflow beind a upload from a source (lets call this input).
After this there can be muliple dataflows feeing from the input dataflow and i want to understand just want data is the storage space used for.
for example lets say "input" has 3 entities each of 10gb
transform 1 is fed from input using 2 of the entities (entity_1+entity_2), these are merged into another entity "tranform_1_output" which is 7gb in size
transform 2 if fead from transform 1 output and input entity 3, in this dataflow input_entity_3 is referenced and grouped into a a new entity (trans_2_groupings) which is 3gb in size this is merged as new with the transform 1 output to create a final entity (Desired_output) which is 5gb in size.
My questions is how much storage space is used in the service.
I am thinking there are 2 options.
the first being all entities have enable load ticked:
I was unable to find out any information about which elements of a dataflow increase storage size and i do not have access to the azure storage explorer to conduct any testing if unticking enable load just processes those entities in memory or still stores the data within the azure storeage enviroment.