You are designing an Azure Databricks interactive cluster. The cluster will be used infrequently and will be configured for auto-termination.You need to ensure that the cluster configuration is retained indefinitely after the cluster is terminated. The solution must minimize costs.What should you do?
You are planning a solution to aggregate streaming data that originates in Apache Kafka and is output to Azure Data Lake Storage Gen2. The developers who will implement the stream processing solution use Java, Which service should you recommend using to process the streaming data?
You need to trigger an Azure Data Factory pipeline when a file arrives in an Azure Data Lake Storage Gen2 container.Which resource provider should you enable?
You have an Azure Synapse Analytics dedicated SQL pool that contains a large fact table. The table contains 50 columns and 5 billion rows and is a heap.Most queries against the table aggregate values from approximately 100 million rows and return only two columns.You discover that the queries against the fact table are very slow.Which type of index should you add to provide the fastest query times?
You have an Azure subscription that contains an Azure Synapse Analytics workspace and a user named Used.You need to ensure that User1 can review the Azure Synapse Analytics database templates from the gallery. The solution must follow the principle of least privilege.Which role should you assign to User1?
You need to implement the surrogate key for the retail store table. The solution must meet the sales transactiondataset requirements.What should you create?
What should you do to improve high availability of the real-time data processing solution?
You are designing a security model for an Azure Synapse Analytics dedicated SQL pool that will support multiple companies. You need to ensure that users from each company can view only the data of their respective company. Which two objects should you include in the solution? Each correct answer presents part of the solutionNOTE: Each correct selection it worth one point.
You are creating an Azure Data Factory data flow that will ingest data from a CSV file, cast columns to specified types of data, and insert the data into a table in an Azure Synapse Analytics dedicated SQL pool. The CSV file contains columns named username, comment and date.The data flow already contains the following:* A source transformation* A Derived Column transformation to set the appropriate types of data* A sink transformation to land the data in the poolYou need to ensure that the data flow meets the following requirements;* All valid rows must be written to the destination table.* Truncation errors in the comment column must be avoided proactively.* Any rows containing comment values that will cause truncation errors upon insert must be written to a file in blob storage.Which two actions should you perform? Each correct answer presents part of the solution. NOTE: Each correct selection is worth one point
You are designing a highly available Azure Data Lake Storage solution that will induce geo-zone-redundant storage (GZRS).You need to monitor for replication delays that can affect the recovery point objective (RPO).What should you include m the monitoring solution?