Power BI Semantic Model Memory Errors, Part 1: Model Size

You probably know that semantic models in Power BI can use a fixed amount of memory. This is true of all types of semantic model – Import, Direct Lake and DirectQuery – but it’s not something you usually need to worry about for DirectQuery mode. The amount of memory they can use depends on whether … Continue reading Power BI Semantic Model Memory Errors, Part 1: Model Size
Published on:
Learn moreRelated posts
Power BI Update - September 2025
4 hours ago
Power BI September 2025 Feature Summary
The Power BI September 2025 Feature Summary introduces updates for users and coincides with FabCon Vienna! This release introduces several key...
11 hours ago
How To Get The Details Of Power BI Operations Seen In The Capacity Metrics App
It’s the week of Fabcon Europe and you’re about to be overwhelmed with new Fabric feature announcements. However there is a new bl...
2 days ago
Deprecation of Power BI Integration within SharePoint Lists and Libraries
Today, we are announcing the retirement of the Power BI integration within SharePoint lists and libraries. This change impacts customers who c...
5 days ago
AI and Agentic Development for Power BI
7 days ago
New Fabric Tenant Setting: “Set alert” Button Visibility for Power BI Users
A new Microsoft Fabric tenant setting will make the “Set alert” button visible to all Power BI web users, enabling them to create Fabric Activ...
7 days ago
What Happens When Power BI Direct Lake Semantic Models Hit Guardrails?
Direct Lake mode in Power BI allows you to build semantic models on very large volumes of data, but because it is still an in-memory database ...
9 days ago
Power BI Beginner's Tutorial (2025)
12 days ago