Model the Data

Design and implement data models, create DAX calculations, and optimize performance.

Covers designing data models including relationships and table properties, creating DAX measures and calculated columns, implementing time intelligence, and optimizing model performance using best practices.
5 minutes 5 Questions

Modeling the data in Power BI is a crucial step that involves organizing, structuring, and defining relationships between different data tables to create a coherent analytical foundation. This process transforms raw data into a meaningful structure that supports accurate reporting and analysis. Th…

Concepts covered: Configure table and column properties, Implement role-playing dimensions, Define relationship cardinality and cross-filter direction, Create a common date table, Calculated columns vs calculated tables, Create single aggregation measures, Use the CALCULATE function, Implement time intelligence measures, Use basic statistical functions, Create semi-additive measures, Create measures using quick measures, Create calculated tables and columns with DAX, Create calculation groups, Remove unnecessary rows and columns, Identify poorly performing measures and visuals, Use Performance Analyzer and DAX query view, Improve performance by reducing granularity

Test mode:
PL-300 - Model the Data Example Questions

Test your knowledge of Model the Data

Question 1

A credit card fraud detection system uses Power BI to visualize transaction monitoring data. The dataset captures every card swipe, tap, and online purchase with millisecond-precision timestamps, merchant category codes, geographic coordinates, and real-time risk scores calculated at each authorization checkpoint. Over 36 months, the system has accumulated 890 million individual transaction authorization records across consumer, business, and premium card portfolios. The fraud investigation team primarily analyzes daily fraud incident counts by merchant category, weekly chargeback ratios aggregated at the regional processor level, and monthly false positive rates per card product line. The semantic model currently consumes 42GB and interactive filtering between card portfolios causes visuals to hang for 65+ seconds. The data engineering manager has requested optimization recommendations. Given the analytical focus on daily, weekly, and monthly aggregated fraud metrics rather than individual millisecond authorization events, which data transformation strategy would most effectively address the performance and memory constraints?

Question 2

Scenario: You are a Power BI analyst at a national grocery chain with 200 stores across the country. The data engineering team has provided you with three fact tables: Daily Sales (sale_date), Inventory Movements (movement_date), and Promotional Events (promo_start_date). You created a date table using the following DAX expression: Dates = CALENDAR(DATE(2021,1,1), DATE(2026,12,31)) After establishing many-to-one relationships from each fact table to the Dates table, you added calculated columns for Year, MonthNumber, MonthName, QuarterLabel, and WeekOfYear. The regional managers need to analyze store performance using PREVIOUSYEAR and DATEADD functions to compare current metrics against historical periods. When testing these measures, you receive an error indicating that the time intelligence function cannot find a valid date table. Problem: The Date column contains 2,192 unique date values with no gaps or duplicates, and all relationships are properly configured. The data types are correct across all tables. What critical configuration step is missing that prevents the time intelligence functions from recognizing this as a valid date table?

Question 3

A multinational corporation uses Power BI to consolidate headcount data from 30 regional offices worldwide. The Workforce_Snapshots table contains RegionID, ReportDate, DepartmentCode, and EmployeeCount columns, with each region submitting headcount figures at varying intervals - some weekly, others bi-weekly, and a few monthly. The HR director discovers that the existing measure TotalHeadcount = CALCULATE(SUM(Workforce_Snapshots[EmployeeCount]), LASTDATE('Date'[Date])) produces incorrect results when viewing annual data: regions that submitted their final 2024 report on December 15th appear as zero in the regional breakdown, while only regions with December 31st entries contribute to the total. The business requirement specifies that annual headcount should reflect the most recent reported value from each region within the selected year, with these regional figures then combined for a corporate total. The Date dimension has standard calendar hierarchies and maintains an active relationship to ReportDate. The analyst must also ensure the measure calculates appropriately when users drill down to quarterly or monthly views. Which DAX measure construction addresses this multi-frequency reporting challenge for the semi-additive headcount calculation?

More Model the Data questions
675 questions (total)