You have an Azure SQL database that has masked columns.
You need to identify when a user attempts to infer data from the masked columns.
What should you use?
Answer : D
Dynamic Data Masking is designed to simplify application development by limiting data exposure in a set of pre-defined queries used by the application. While Dynamic Data Masking can also be useful to prevent accidental exposure of sensitive data when accessing a production database directly, it is important to note that unprivileged users with ad-hoc query permissions can apply techniques to gain access to the actual data. If there is a need to grant such ad-hoc access, Auditing should be used to monitor all database activity and mitigate this scenario.
References:
https://docs.microsoft.com/en-us/sql/relational-databases/security/dynamic-data-masking
You are the data engineer tor your company. An application uses a NoSQL database to store dat
a. The database uses the key-value and wide-column NoSQL database type.
Developers need to access data in the database using an API.
You need to determine which API to use for the database model and type.
Which two APIs should you use? Each correct answer presents a complete solution.
NOTE: Each correct selection s worth one point.
Use the following login credentials as needed:
Azure Username: xxxxx
Azure Password: xxxxx
The following information is for technical support purposes only:
Lab Instance: 10277521
You need to increase the size of db2 to store up to 250 GB of data.
To complete this task, sign in to the Azure portal.
Answer : A
You develop data engineering solutions for a company.
You need to ingest and visualize real-time Twitter data by using Microsoft Azure.
Which three technologies should you use? Each correct answer presents part of the solution.
NOTE: Each correct selection is worth one point.
Answer : B, D, F
You can use Azure Logic apps to send tweets to an event hub and then use a Stream Analytics job to read from event hub and send them to PowerBI.
References:
Note: This question is part of series of questions that present the same scenario. Each question in the series contain a unique solution. Determine whether the solution meets the stated goals.
You develop data engineering solutions for a company.
A project requires the deployment of resources to Microsoft Azure for batch data processing on Azure
HDInsight. Batch processing will run daily and must:
Scale to minimize costs
Be monitored for cluster performance
You need to recommend a tool that will monitor clusters and provide information to suggest how to scale.
Solution: Download Azure HDInsight cluster logs by using Azure PowerShell.
Does the solution meet the goal?
Answer : B
Instead monitor clusters by using Azure Log Analytics and HDInsight cluster management solutions.
References:
https://docs.microsoft.com/en-us/azure/hdinsight/hdinsight-hadoop-oms-log-analytics-tutorial
Your company uses several Azure HDInsight clusters.
The data engineering team reports several errors with some application using these clusters.
You need to recommend a solution to review the health of the clusters.
What should you include in you recommendation?
Answer : B
Azure Monitor logs integration. Azure Monitor logs enables data generated by multiple resources such as HDInsight clusters, to be collected and aggregated in one place to achieve a unified monitoring experience.
As a prerequisite, you will need a Log Analytics Workspace to store the collected data. If you have not already created one, you can follow the instructions for creating a Log Analytics Workspace.
You can then easily configure an HDInsight cluster to send many workload-specific metrics to Log Analytics.
References:
You plan to implement an Azure Cosmos DB database that will write 100,000 JSON every 24 hours. The
database will be replicated to three regions. Only one region will be writable.
You need to select a consistency level for the database to meet the following requirements:
Guarantee monotonic reads and writes within a session.
Provide the fastest throughput.
Provide the lowest latency.
Which consistency level should you select?
Answer : C, D, E
Session: Within a single client session reads are guaranteed to honor the consistent-prefix (assuming a single ''writer'' session), monotonic reads, monotonic writes, read-your-writes, and write-follows-reads guarantees.
Clients outside of the session performing writes will see eventual consistency.
References:
https://docs.microsoft.com/en-us/azure/cosmos-db/consistency-levels