A client wants to migrate their Tableau Server to Tableau Cloud. The Tableau Server is configured with three sites: Finance, Strategy, and Marketing. A
consultant must provide a solution that minimizes user impact and costs.
Which configuration should the consultant recommend for Tableau Cloud to meet the client's requirements?
Answer : A
To minimize user impact and costs while migrating from Tableau Server to Tableau Cloud with multiple sites, the best solution is:
Single Tableau Cloud Instance with Multiple Projects: Instead of multiple sites which could imply higher management overhead and possibly higher costs, configuring one Tableau Cloud instance with different project folders for each former site (Finance, Strategy, Marketing) is most efficient.
Benefits: This setup maintains organizational separation of data and access similar to having different sites but leverages the unified management and simplicity of a single cloud instance. It reduces complexity in user access management and integration points.
Implementation: Each project folder acts like a mini-site within the larger instance, where specific permissions and content can be managed independently, akin to the original server setup but within a single cloud-based environment.
Reference This recommendation is in line with best practices for cloud migration focusing on consolidation and cost efficiency, as suggested in Tableau's official documentation for cloud migration strategies.
A client requests a published Tableau data source that is connected to SQL Server. The client needs to leverage the multiple tables option to create an
extract. The extract will include partial data from the SQL Server data source.
Which action will reduce the amount of data in the extract?
Answer : A
Using an extract filter is an effective way to reduce the amount of data in a Tableau extract. Extract filters allow you to specify a subset of the data to include, which can significantly decrease the size of the extract by excluding unnecessary data. This is particularly useful when you only need partial data from a larger SQL Server data source.
When dealing with large datasets in SQL Server and needing to create a manageable extract in Tableau, using an extract filter is the most direct and effective method to limit the data included:
Extract Filter: This involves setting filters that apply directly when the data is extracted from the source. This means that only the data meeting the specified criteria will be extracted and loaded into Tableau, significantly reducing the size of the extract.
To apply an extract filter, in the Data Source page in Tableau, drag the fields you want to filter by to the Filters shelf. Then, configure the desired filter criteria. When you create the extract, choose the option to 'Add Filters to Extract' and select the configured filters. This ensures that only the data that meets these conditions is extracted from the SQL Server.
This approach not only minimizes the data volume but also speeds up performance in Tableau because it processes a smaller subset of the full dataset.
Reference This procedure is described in detail in Tableau's help documentation on managing extracts and optimizing performance by using extract filters, which is recommended for scenarios involving large datasets or when specific subsets of data are required for analysis.
A client wants to provide sales users with the ability to perform the following tasks:
* Access published visualizations and published data sources outside the company network.
* Edit existing visualizations.
* Create new visualizations based on published data sources.
. Minimize licensing costs.
Which site role should the client assign to the sales users?
Answer : A
The Explorer (can publish) site role in Tableau is designed for users who need to access, edit, and create visualizations based on published data sources, even when they are outside the company network. This role allows users to perform web editing and save their work, making it suitable for sales users who need these capabilities. It is also a cost-effective option as it does not require the full capabilities and associated costs of the Creator license.
A client wants to count all the distinct orders placed in 2010. They have written the following calculation, but the result is incorrect.
IF YEAR([Date])=2010 THEN COUNTD ([OrderID]) END
Which calculation will produce the correct result?
Answer : C
The correct calculation to count all distinct orders placed in 2010 involves placing the conditional inside the aggregation function, not the other way around. Here's how to correct the client's calculation:
Original Calculation Issue: The client's original calculation attempts to apply the COUNTD function within an IF statement, which does not work as expected because the COUNTD function cannot conditionally count within the scope of the IF statement.
Correct Calculation: COUNTD(IF YEAR([Date]) = 2010 THEN [OrderID] END). This calculation checks each order date; if the year is 2010, it returns the OrderID. The COUNTD function then counts all unique OrderIDs that meet this condition.
Why It Works: This method ensures that each order is first checked for the year condition before being counted, effectively filtering and counting in one step. It efficiently processes the data by focusing the distinct count operation only on relevant records.
Reference This approach is consistent with Tableau's guidance on using conditional logic inside aggregation functions for accurate and efficient data calculations, as detailed in the Tableau User Guide under 'Aggregations and Calculations'.
A stakeholder has multiple files saved (CSV/Tables) in a single location. A few files from the location are required for analysis. Data transformation (calculations)
is required for the files before designing the visuals. The files have the following attributes:
. All files have the same schema.
. Multiple files have something in common among their file names.
. Each file has a unique key column.
Which data transformation strategy should the consultant use to deliver the best optimized result?
Answer : B
Given that all files share the same schema and have a common element in their file names, the wildcard union is an optimal approach to combine these files before performing any transformations. This strategy offers the following advantages:
Efficient Data Combination: Wildcard union allows multiple files with a common naming scheme to be combined into a single dataset in Tableau, streamlining the data preparation process.
Uniform Schema Handling: Since all files share the same schema, wildcard union ensures that the combined dataset maintains consistency in data structure, making further data manipulation more straightforward.
Pre-Transformation Combination: Combining the files before applying transformations is generally more efficient as it reduces redundancy in transformation logic across multiple files. This means transformations are written and processed once on the unified dataset, rather than repeatedly for each individual file.
Wildcard Union in Tableau: This feature simplifies the process of combining multiple similar files into a single Tableau data source, ensuring a seamless and efficient approach to data integration and preparation.
A client wants guidance for Creators to build efficient extracts from large data sources.
What are three Tableau best practices that the Creators should use? Choose three.
Answer : A, B, E
To build efficient extracts from large data sources, it is crucial to minimize the load and optimize the performance of the extracts:
A . Keep only the data required for analysis by using extract filters: This best practice involves using filters to reduce the volume of data extracted, thus focusing only on the data necessary for analysis.
B . Use aggregate data for visible dimensions, whenever possible: Aggregating data at the time of extraction reduces the granularity of the data, which can significantly improve performance and reduce the size of the extract.
E . Hide all unused fields: Removing fields that are not needed for analysis from the extract reduces the complexity and size of the data model, which in turn enhances performance and speeds up load times.
These practices are endorsed in Tableau's official documentation and training sessions as effective ways to enhance the performance of Tableau extracts and optimize dashboard responsiveness.
A client is using the Tableau Content Migration Tool to move content from an old Tableau Server to a new Tableau Server.
Which content will need to be moved using a different tool or process?
Answer : B
When migrating content between Tableau servers, certain types of content may require special consideration or different tools for migration:
Tableau Prep Flows: These are specific to Tableau Prep and are not included in the standard content migration capabilities of the Tableau Content Migration Tool. Tableau Prep flows often require separate processes for migration due to their distinct setup and integration with data sources and workflows.
Published Data Sources and Workbooks: These can typically be migrated directly using the Tableau Content Migration Tool, which supports moving published data sources (both live connections and extracts) and workbooks without requiring additional tools.
Tableau Help and Support: Offers comprehensive tutorials and guidelines on using different tools for migrating various types of content, including the specific requirements for migrating Tableau Prep flows which are not covered by the standard content migration tool.
Topic 2, hands-on labLab Section