Exam Questions Flashcards
You plan to get data from flat files for a Power BI dataset.
You need to select a location to store the files.
Which location requires an on-premises data gateway?
a) OneDrive for Business
b) Personal OneDrive account
c) SharePoint Online Team Sites
d) Shared folder on a local network
a) OneDrive for Business
b) Personal OneDrive account
c) SharePoint Online Team Sites
d) Shared folder on a local network
The shared folder on a local network will require a gateway as it is not available to the internet. SharePoint Online, OneDrive, and OneDrive for Business all can be refreshed as cloud data sources without a gateway.
You have multiple Excel files stored in a folder synced with OneDrive for Business.
You need to import data from the files. The solution must NOT use a data gateway.
Which type of connector should you use?
Select only one answer.
A) Excel workbook
B) Folder
C) SharePoint folder
D) SharePoint list
A) Excel workbook
B) Folder
C) SharePoint folder
D) SharePoint list
A SharePoint folder is the only connector that will allow the import of multiple Excel (or CSV) files stored in a OneDrive for Business folder, without using a data gateway. While you can connect to the folder on an on-premises device, it would require a data gateway to refresh in the service. An Excel workbook would only connect to a single Excel file and would also require a data gateway. A SharePoint list connector only connects to SharePoint lists and cannot connect to Excel files.
From Power BI Desktop, you create a data source by importing an Excel file that contains 10,000 rows.
You plan to identify data anomalies within the data source.
You need to ensure that column distribution considers all rows in the Excel file.
What should you do?
A) In the Data Source Settings, modify the Advanced settings.
B) In the Data Source Settings, modify the Edit Permissions settings.
C) In the Power Query Editor window, enable the Column Profile view.
D) In the Power Query Editor window, modify the Query Settings.
A) In the Data Source Settings, modify the Advanced settings.
B) In the Data Source Settings, modify the Edit Permissions settings.
C) In the Power Query Editor window, enable the Column Profile view.
D) In the Power Query Editor window, modify the Query Settings.
By default, Power BI uses the top 1,000 rows for profiling. To ensure that column distribution considers all rows in the Excel file, you need to modify the Power Query Editor profiling status setting. The Power Query Editor settings, Advanced settings, and Permissions settings have no bearing on the profiling characteristics.
From Power BI Desktop, you create a data source by importing an Excel file.
You plan to identify data anomalies within the data source.
Which interface should you use to display the column distribution and column quality graphs?
A) Advanced Editor in Power Query Editor
B) Power Query Editor
C) the Data view
D) the Model view
Power Query Editor provides the ability to display the column distribution and column quality graphs. Data view provides access to data within a dataset. Model view provides access to the data model of a dataset. Advanced Editor of Power Query Editor provides the ability to work with the M code used for shaping data in Power Query Editor.
You create a report with Power BI Desktop. You are tasked with distributing the report to both internal and external recipients in your organization. The report has the following requirements:
• To be accessible from IOS and Android mobile devices
• To maximize branding and other look and feel aspects
You need to select a distribution option for the report.
Which distribution option should you use?
Choose the correct answer
A) Share the report from Power BI Service.
B) Package content and publish an app.
C) Share a workspace that includes the report.
D) Publish to web and use embed code.
A) Share the report from Power BI Service.
B) Package content and publish an app.
C) Share a workspace that includes the report.
D) Publish to web and use embed code.
You have a query that retrieves data from a table that contains more than 8,000 rows of data.
In Power Query Editor, you notice that the column statistics for each column shows a count of exactly 1,000.
You need to ensure that the column statistics for each column shows the statistics based on all rows that are returned by the query.
What should you do?
A) Add a Table.Buffer applied step to the query.
B) Change the query load type from Import to DirectQuery.
C) Create a Top N row count parameter for the query.
D) From the query window, select Column profiling on the entire dataset.
A) Add a Table.Buffer applied step to the query.
B) Change the query load type from Import to DirectQuery.
C) Create a Top N row count parameter for the query.
D) From the query window, select Column profiling on the entire dataset.
Selecting column profiling for the entire dataset will change the column profiler to analyze the entire query dataset. Adding a Table.Buffer applied step will only cache the entire query in memory during refresh. The query load type in the model will not impact how many rows are used by the column profiler. Using a Top N row count applied step or applying a parameter will not change how the column statistics feature calculates.
You are analyzing query data by using Power Query Editor.
You need to ensure that the Column statistics are based on an analysis of the entire dataset.
What should you do?
Select only one answer.
A) Change the query connection type from Import to DirectQuery.
B) From the status bar, change profiling status to entire dataset.
C) In Power Query Editor, enable column profiling from the View ribbon.
D) Load the data into the data model.
A) Change the query connection type from Import to DirectQuery.
B) From the status bar, change profiling status to entire dataset.
C) In Power Query Editor, enable column profiling from the View ribbon.
D) Load the data into the data model.
By default, column profiling is only based on the first 1000 rows of the preview. Changing the column profile status to entire dataset will query the entire dataset for the column profiling information.
“prompt”
You have an Excel spreadsheet that contains three columns labeled Year, 2021, and 2012. The entries in rows for the first column consists of names of the individual months in the year while the other two columns contain the sales amount for each month for the corresponding year.
You import data from the Excel spreadsheet into Power BI Desktop.
You need to transform the data so it will consist of three columns, with the first one containing month, the second containing year, and the third containing the sales amount for that month and year.
Which transformation should you use first?
Select only one answer.
A) Pivot
B) Remove Columns
C) Transpose Table
D) Unpivot
Clarifier
Footnote
A) Pivot
B) Remove Columns
C) Transpose Table
D) Unpivot
Selecting Unpivot will allow you to shape the current table into the one with the year, month, and sales amount columns, which will need to be renamed afterwards. Pivot would be applicable in the opposite scenario, in which flat data needs to be reorganized into one containing aggregate values for each unique value in each column. Transposing would switch rows and columns. Removing columns would result in a table with insufficient data to perform unpivot.
Answer - Clarifier
Answer - Footnote
You have a query named FactSales that retrieves data from a SQL Server table.
You need to update the query to prevent new columns that may be added to the table in the future from getting imported during dataset refreshes.
You select the existing columns in the query.
Which two applied steps can you use in Power Query Editor to achieve the goal? Each correct answer presents a complete solution.
Select all answers that apply.
A) Choose Columns
B) Remove Columns
C) Remove Other Columns
D) Transpose
A) Choose Columns
B) Remove Columns
C) Remove Other Columns
D) Transpose
Using the Choose Columns and Remove Other Columns options will allow you to explicitly select the columns that you want to keep. This enforces the requirement that new columns will not be automatically added in the future. Using Remove Columns will delete the existing columns in the table, and new columns that get added in the future will still be imported automatically. Transpose treats rows as columns, and columns as rows. It will not limit the number of columns being imported.
You have a query that has a column named Country and columns for each month of sales data. The names of the sales data columns use a format of Month Year, for example January 2022. The month columns contain the Sales Amount values for each month.
You need to reshape the table to have all sales amount values in a single column.
A new column should also be added and contain the relevant month year for each row of data.
What should you do from the Power Query Editor?
``` Select only one answer.
A) Select the Country column and then select Pivot Column.
B) Select the Country column and then select Unpivot Other Columns.
C) Select the first sales data column and then select Unpivot Columns.
D) Select Transpose.
~~~
A) Select the Country column and then select Pivot Column.
B) Select the Country column and then select Unpivot Other Columns.
C) Select the first sales data column and then select Unpivot Columns.
D) Select Transpose.
``` Selecting Unpivot Other Columns will unpivot all current and future month/year columns into a new sales data column. Transposing treats rows as columns, and columns as rows. It will not create a single sales data column. Selecting Unpivot Columns in the first sales data column will successfully unpivot these columns into a new single column for sales data, but the other sales month/year columns will not be included in the new sales column. Pivoting the data takes values on rows in a column and turns them into new columns.
You have a Power BI data source that contains the following tables:
ProductCategory: Lists the product category ID and category name
ProductSubcategory: Lists the product subcategory ID, subcategory name, and the related category ID
ProductList: Lists the product ID, name, and subcategory ID
You need to optimize the data to create a dimension for use in a star schema data model.
How should the tables be transformed before loading into the model?
Select only one answer.
A) Import all three tables into the data model and connect them using relationships.
B) Import only the ProductName table into the model.
C) Merge the queries to create a single loaded table for Product.
D) Use the append command to create a single loaded table for product.
A) Import all three tables into the data model and connect them using relationships.
B) Import only the ProductName table into the model.
C) Merge the queries to create a single loaded table for Product.
This answer is correct.
D) Use the append command to create a single loaded table for product.
A star schema should have a single table for each dimension or product, so using the combine command is what is required to create a single product table and aim towards a star schema design.
Combine multiple tables into a single table - Training | Microsoft Learn
You have an Excel spreadsheet that contains two columns as follows:
Category: Contain names of categories
Subcategory: Contain names of subcategories for each category
You import the Excel spreadsheet into Power BI Desktop.
You need to transform the data to meet the following requirements:
Have multiple columns, where one column represents one category
Have a single row that includes the total count of subcategories for each category
Which transformation should you use?
A) Pivot Columns
B) Rename Columns
C) Transpose Table
D) Unpivot Columns
A) Pivot Columns
B) Rename Columns
C) Transpose Table
D) Unpivot Columns
The pivot column operation converts data into a table by aggregating values in a column. In this case, you can pivot the columns to calculate the count of product subcategories in each product category. The unpivot column operation performs the opposite task, converting columns into rows. Renaming columns would result in a table with different column names but the same data content. Transposing would simply switch rows and columns, without the required aggregation.
You have a fact table that contains sales data.
The fact table includes a SalesDate column formatted as a Date data type. Auto date/time setting is disabled in both global and current file options.
You load the fact table into Power BI Desktop.
You need to ensure that you are able to analyze data on a yearly, quarterly, monthly, weekly, and daily basis. Your solution must minimize the model size and administrative effort.
What should you do?
A) Add a separate date dimension table.
B) Add a year, month, and week columns to the fact table.
C) Enable the Auto date/time current file option.
D) Enable the Auto date/time global option.
A) Add a separate date dimension table.
B) Add a year, month, and week columns to the fact table.
C) Enable the Auto date/time current file option.
D) Enable the Auto date/time global option.
Adding a separate date dimension table that includes year, month, and week information is the optimal approach which provides the required functionality, while minimizing the model size and administrative effort. Adding a year, month, and week columns to the fact table would increase the amount of administrative effort. Enabling the Auto date/time global or current file option would increase the model size. In addition, it would not provide the ability to describe weekly time periods (only year, quarter, month, and day).
You have a Power BI model with the following fact tables and storage modes.
```FactStoreSales (Import mode)
FactOnlineSales (DirectQuery mode)
You have a dimension table named DimCalendar that has a relationship to both fact tables.
~~~
You need to assign a storage mode for DimCalendar. The solution must minimize the time to execute queries that combine data from the dimension table and the fact tables.
Which storage mode should you use?
A) DirectQuery
B) Dual
C) Import
D) none
A) DirectQuery
B) Dual
C) Import
D) none
Using Dual mode means that either an import query can be run when accessing data from FactStoreSales, or a DirectQuery query can be run when accessing data from FactOnlineSales. Using Import mode means the queries are only optimized for FactStoreSales (Import). Using DirectQuery mode means the queries are only optimized for FactOnlineSales (DirectQuery). A storage mode of Import, DirectQuery, or Dual must be assigned.
You connect Power Query Editor to a database table.
You need to remove the Row ID column. Your solution must ensure that new columns do NOT display in the table model during a scheduled refresh in the future.
What transformation should you use?
A) Select Row ID, then use the Remove Other Columns command.
B) Use the Remove Column command on the Row ID column.
C) Use the Select Columns command and chose the columns to keep.
D) Use the Transpose command, then filter the rows to remove Row ID.
A) Select Row ID, then use the Remove Other Columns command.
B) Use the Remove Column command on the Row ID column.
C) Use the Select Columns command and chose the columns to keep.
D) Use the Transpose command, then filter the rows to remove Row ID.
Only the Select Columns command will let you choose columns to keep, delete the columns you do not want, and prevent new columns from showing up in the table in the future.
You have designed a star schema to simplify your data.
You need to understand the relationship between the tables in the star schema.
What is the relationship between the fact table and dimension tables?
A) many-to-many
B) many-to-one
C) one-to-many
D) one-to-one
A) many-to-many
B) many-to-one
C) one-to-many
D) one-to-one
There is a many-to-one relationship between a fact table and dimension table. For each row in a dimension table there may be multiple matching rows in the fact table.
You have the following Power Query M formula that generates a range of dates.
= List.Dates(#date(2020,05,31), 365, #duration(1,0,0,0))
What is the resulting date range?
A) a list of days ending on May 31, 2020 and starting 365 days earlier
B) a list of days starting on May 31, 2020 and ending 365 days later
C) a list of months ending in May 2020 and starting 12 months earlier
D) a list of months starting in May 2020 and ending 12 months later
A) a list of days ending on May 31, 2020 and starting 365 days earlier
B) a list of days starting on May 31, 2020 and ending 365 days later
C) a list of months ending in May 2020 and starting 12 months earlier
D) a list of months starting in May 2020 and ending 12 months later
The #date element designates the starting date, 365 designates the duration counter, and #duration(1,0,0,0) designates the duration interval in days, so this formula results in a listing of days starting on May 31, 2020 and ending 365 days later.
You plan to use the calculated table functionality to add a duplicate table in Power BI Desktop.
Which characteristics of the original table will be duplicated?
A) data and column visibility only
B) data and hierarchies only
C) data, hierarchies, and column visibility
D) data only
A) data and column visibility only
B) data and hierarchies only
C) data, hierarchies, and column visibility
D) data only
A calculated table only duplicates data. Any model configurations such as column visibility or hierarchies must be recreated if needed.
You create a data model in Power BI Desktop that contains DAX calculated columns and measures. You now need to create a report.
In which two places can a DAX calculated column be used, but a DAX calculated measure cannot be used? Each correct answer presents a complete solution.
A) as a filter in the “Filters on this page” well of the Filters pane
B) as a filter in the “Filters on this visual” well of the Filters pane
C) as an item in the “Add drill-through fields here” well of the Visualizations pane
D) as an item in the Fields well of a slicer
A) as a filter in the “Filters on this page” well of the Filters pane
B) as a filter in the “Filters on this visual” well of the Filters pane
C) as an item in the “Add drill-through fields here” well of the Visualizations pane
D) as an item in the Fields well of a slicer
```Unlike a measure, a calculated column can be used in a slicer to place filter options on the report page. DAX measures cannot be placed in the “Filters on this page” well. They can only be placed per visual, in the “Filters on this visual” well of the Filters Pane. Both DAX columns and measures may be used as a visual-level filter. Both DAX columns and measures can be used in the drillthrough well.
~~~
You need to develop a quick measure in Power BI Desktop.
Which two elements can you use? Each correct answer presents a complete solution.
Select all answers that apply.
A) Calculations
B) Conditional columns
C) Data Analysis Expression (DAX) queries
D) Fields
E) Power Query M functions
A) Calculations
B) Conditional columns
C) Data Analysis Expression (DAX) queries
D) Fields
E) Power Query M functions
When creating a quick measure in Power BI Desktop, you apply calculations to fields. You do not explicitly create a DAX query, but you choose calculations and fields, which result in automatic generation of a DAX query. Conditional columns are separate from quick measures. Unlike quick measures, they create a value for each row in a table and are stored in the .pbix file. Power Query M functions are not directly accessible from the Quick Measure interface.
You have a Power BI Desktop dataset that includes a table named Warehouse. The Warehouse table includes a column named Inventory Count, which contains the current number of items for each row of a particular type on a given day.
You have the following Data Analysis Expression (DAX) query that calculates the sum of all values in the Inventory Count column of the Warehouse table:
Current Inventory Count =
CALCULATE (
SUM ( ‘Warehouse’[Inventory Count] ))
You need to ensure that Current Inventory Count returns only the current total number of inventory items, rather than the sum of all inventory items that includes item counts from previous days.
What DAX function should you include in the query?
Select only one answer.
A) CALENDAR
B) CALENDARAUTO
C) DISTINCTCOUNT
D) LASTDATE
A) CALENDAR
B) CALENDARAUTO
C) DISTINCTCOUNT
D) LASTDATE
```The LASTDATE function will ensure that the SUM function applies only to the last date of the time period, resulting in a semi-additive behavior. The DISTINCTCOUNT function counts the number of distinct values in a column, which results in additive behavior. The CALENDAR function returns a table with a column named Date that contains a contiguous set of dates based on the start date and end date that you specify. The CALENDARAUTO function returns a table with a column named Date that contains a contiguous set of dates based on data in the model.
~~~
You need to use DAX quick measures to generate results to use in a report.
Which type of DAX quick measure calculations will NOT work against a DirectQuery table?
Select only one answer.
A) aggregate per category
B) mathematical operations
C) time intelligence
D) X-functions
A) aggregate per category
B) mathematical operations
C) time intelligence
D) X-functions
Time intelligence functions have performance implications and are disabled for quick measures against DirectQuery tables. Mathematical operations, aggregate per category, and X-functions are all supported against DirectQuery.
You need to reduce the cardinality of a column in a Power BI dataset.
What should you do?
Select only one answer.
A) Add a prefix to each value in the column.
B) Add a suffix to each value in the column.
C) Replace null values with a value of 0.
D) Reduce the number of distinct values
A) Add a prefix to each value in the column.
B) Add a suffix to each value in the column.
C) Replace null values with a value of 0.
D) Reduce the number of distinct values
```A column that has a lot of repeated values and a low number of distinct values will have a low cardinality. Reducing the number of distinct values reduces cardinality. Replacing nulls does not change the number of distinct values. A static prefix or suffix does not change cardinality.
~~~
You plan to run Power BI Desktop Performance Analyzer.
You need to ensure that the data engine cache will NOT impact the test results without restarting Power BI Desktop.
What should you do first?
Select only one answer.
A) Add a blank page to the .pbix file and select it.
B) Connect DAX Studio to the data model.
C) Invoke the Clear Cache function.
D) Invoke the Refresh Metadata function.
A) Add a blank page to the .pbix file and select it.
B) Connect DAX Studio to the data model.
C) Invoke the Clear Cache function.
D) Invoke the Refresh Metadata
DAX Studio, once connected to the data model, can be used to clear the data engine cache. The Clear Cache function can be invoked from DAX Studio, once you connect it to the data model. The Refresh Metadata function can be invoked from DAX Studio to update the metadata of the currently selected model. Adding a blank page to the .pbix file and selecting it is the first step in clearing the visual cache, not the data engine cache.