Skip to main content
cancel
Showing results for 
Search instead for 
Did you mean: 

Register now to learn Fabric in free live sessions led by the best Microsoft experts. From Apr 16 to May 9, in English and Spanish.

Reply
jabate
New Member

Azure Data Bricks Data Refresh

I have a report utilizing data import from a persisted table in Databricks. Once the dataset size increased I received the following error:

Total size of serialized results of 17 tasks (4.1 GB) is bigger than spark.driver.maxResultSize

 

Looking up the error I found alot of spark specific posts explaining that spark.driver.maxResultSize is a variable which exists to prevent out of memory exceptions. The reason I'm posting in a Power BI forum is I haven't had any issue interacting with the data (either in munging the data or writing it to hive) on the Databricks side.

 

Does anybody know some details about how the refresh interacts with spark/Databricks and why it could be causing the issue in this particular situation? I would prefer having some understanding of why it's occurring in this situation before I adjust the maxResultSize variable (possibly several times).

5 REPLIES 5
v-shex-msft
Community Support
Community Support

Hi @jabate ,

 

I think this issue should more related to database settings. it sounds like response data amount is greater than default cache size so refresh requests has been blocked/canceled.

 

Maybe you can take a look at following link to know more about this issue:

Total size of serialized results of 16 tasks (1048.5 MB) is bigger than spark.driver.maxResultSize (...

Spark Configuration

 

For Power BI Architecture, you can refer to below link:

Power BI Security

 

Regards,

Xiaoxin Sheng

Community Support Team _ Xiaoxin
If this post helps, please consider accept as solution to help other members find it more quickly.

Hi Xiaoxin,

 

Thanks for looking at this issue. As an update we increased teh variable size to 35gb on both the clusters we are running, but still encounter the same 4gb error when attempting a refresh. We have a ticket in with the dev team to ascertain whether the error is being thrown by our Databricks instance (meaning we missed something in the adjustment of the variable) or whether it's occurring in the attempt to write to our premium capacity storage.

 

Julius

Hi @jabate ,

I'd like to suggest you open a support ticket to get better support from dev team, I think this issue is more related to spark itself.

submit a support ticketsubmit a support ticket

 

Regards,
Xiaoxin Sheng

Community Support Team _ Xiaoxin
If this post helps, please consider accept as solution to help other members find it more quickly.

Thanks Xiaoxin , a ticket is currently in but I have not heard back and need to follow up on it. The changes have been made in spark so I need to confirm with Microsoft support that the issue is not related to the Hive metastore which holds the uploaded PBIX files. 

 

I'll make sure to post their resolution/recommendation once I'm able to get that back on a call and get it sorted out.

Did you manage to figure this out? I am getting the same error.

Helpful resources

Announcements
Microsoft Fabric Learn Together

Microsoft Fabric Learn Together

Covering the world! 9:00-10:30 AM Sydney, 4:00-5:30 PM CET (Paris/Berlin), 7:00-8:30 PM Mexico City

PBI_APRIL_CAROUSEL1

Power BI Monthly Update - April 2024

Check out the April 2024 Power BI update to learn about new features.

April Fabric Community Update

Fabric Community Update - April 2024

Find out what's new and trending in the Fabric Community.

Top Solution Authors
Top Kudoed Authors