Latest DP-300 Exam Premium Dumps provide by TrainingQuiz.com to help you Passing DP-300 Exam! TrainingQuiz.com offers the updated DP-300 exam dumps, the TrainingQuiz.com DP-300 exam questions has been updated to correct Answer. Get the latest TrainingQuiz.com DP-300 pdf dumps with Exam Engine here:
(464 Q&As Dumps, 40%OFF Special Discount: DumpsDB)
You have an on-premises Microsoft SQL Server 2019 instance that hosts a database named DB1. You plan to perform an online migration of DB1 to an Azure SQL managed instance by using the Azure Database Migration Service. You need to create a backup of DB1 that is accessible to the Azure Database Migration Service. What should you run for the backup and where should you store the backup? To answer, select the appropriate options in the answer area. NOTE: Each correct selection is worth one point.
You have an Azure virtual machine named VM1 on a virtual network named VNet1. Outbound traffic from VM1 to the internet is blocked. You have an Azure SQL database named SqlDb1 on a logical server named SqlSrv1. You need to implement connectivity between VM1 and SqlDb1 to meet the following requirements: Ensure that VM1 cannot connect to any Azure SQL Server other than SqlSrv1. Restrict network connectivity to SqlSrv1. What should you create on VNet1?
Correct Answer: C
Section: [none] Explanation: Azure Private Link enables you to access Azure PaaS Services (for example, Azure Storage and SQL Database) and Azure hosted customer-owned/partner services over a private endpoint in your virtual network. Traffic between your virtual network and the service travels the Microsoft backbone network. Exposing your service to the public internet is no longer necessary. Reference: https://docs.microsoft.com/en-us/azure/private-link/private-link-overview Testlet 1 This is a case study. Case studies are not timed separately. You can use as much exam time as you would like to complete each case. However, there may be additional case studies and sections on this exam. You must manage your time to ensure that you are able to complete all questions included on this exam in the time provided. To answer the questions included in a case study, you will need to reference information that is provided in the case study. Case studies might contain exhibits and other resources that provide more information about the scenario that is described in the case study. Each question is independent of the other questions in this case study. At the end of this case study, a review screen will appear. This screen allows you to review your answers and to make changes before you move to the next section of the exam. After you begin a new section, you cannot return to this section. To start the case study To display the first question in this case study, click the Next button. Use the buttons in the left pane to explore the content of the case study before you answer the questions. Clicking these buttons displays information such as business requirements, existing environment, and problem statements. If the case study has an All Information tab, note that the information displayed is identical to the information displayed on the subsequent tabs. When you are ready to answer a question, click the Question button to return to the question. Overview Litware, Inc. is a renewable energy company that has a main office in Boston. The main office hosts a sales department and the primary datacenter for the company. Physical Locations Existing Environment Litware has a manufacturing office and a research office is separate locations near Boston. Each office has its own datacenter and internet connection. The manufacturing and research datacenters connect to the primary datacenter by using a VPN. Network Environment The primary datacenter has an ExpressRoute connection that uses both Microsoft peering and private peering. The private peering connects to an Azure virtual network named HubVNet. Identity Environment Litware has a hybrid Azure Active Directory (Azure AD) deployment that uses a domain named litwareinc.com. All Azure subscriptions are associated to the litwareinc.com Azure AD tenant. Database Environment The sales department has the following database workload: * An on-premises named SERVER1 hosts an instance of Microsoft SQL Server 2012 and two 1-TB databases. * A logical server named SalesSrv01A contains a geo-replicated Azure SQL database named SalesSQLDb1. SalesSQLDb1 is in an elastic pool named SalesSQLDb1Pool. SalesSQLDb1 uses database firewall rules and contained database users. * An application named SalesSQLDb1App1 uses SalesSQLDb1. The manufacturing office contains two on-premises SQL Server 2016 servers named SERVER2 and SERVER3. The servers are nodes in the same Always On availability group. The availability group contains a database named ManufacturingSQLDb1 Database administrators have two Azure virtual machines in HubVnet named VM1 and VM2 that run Windows Server 2019 and are used to manage all the Azure databases. Licensing Agreement Litware is a Microsoft Volume Licensing customer that has License Mobility through Software Assurance. Current Problems SalesSQLDb1 experiences performance issues that are likely due to out-of-date statistics and frequent blocking queries. Requirements Planned Changes Litware plans to implement the following changes: * Implement 30 new databases in Azure, which will be used by time-sensitive manufacturing apps that have varying usage patterns. Each database will be approximately 20 GB. * Create a new Azure SQL database named ResearchDB1 on a logical server named ResearchSrv01. ResearchDB1 will contain Personally Identifiable Information (PII) data. * Develop an app named ResearchApp1 that will be used by the research department to populate and access ResearchDB1. * Migrate ManufacturingSQLDb1 to the Azure virtual machine platform. * Migrate the SERVER1 databases to the Azure SQL Database platform. Technical Requirements Litware identifies the following technical requirements: * Maintenance tasks must be automated. * The 30 new databases must scale automatically. * The use of an on-premises infrastructure must be minimized. * Azure Hybrid Use Benefits must be leveraged for Azure SQL Database deployments. * All SQL Server and Azure SQL Database metrics related to CPU and storage usage and limits must be analyzed by using Azure built-in functionality. Security and Compliance Requirements Litware identifies the following security and compliance requirements: * Store encryption keys in Azure Key Vault. * Retain backups of the PII data for two months. * Encrypt the PII data at rest, in transit, and in use. * Use the principle of least privilege whenever possible. * Authenticate database users by using Active Directory credentials. * Protect Azure SQL Database instances by using database-level firewall rules. * Ensure that all databases hosted in Azure are accessible from VM1 and VM2 without relying on public endpoints. Business Requirements Litware identifies the following business requirements: * Meet an SLA of 99.99% availability for all Azure deployments. * Minimize downtime during the migration of the SERVER1 databases. * Use the Azure Hybrid Use Benefits when migrating workloads to Azure. * Once all requirements are met, minimize costs whenever possible.
Question 88
You need to apply 20 built-in Azure Policy definitions to all new and existing Azure SQL Database deployments in an Azure subscription. The solution must minimize administrative effort. Which three actions should you perform in sequence? To answer, move the appropriate actions from the list of actions to the answer area and arrange them in the correct order.
Correct Answer:
1 - Create an Azure Policy Initiative 2 - Create an Azure Policy Initiative assignment 3 - Run Azure Policy remediation tasks Reference: https://docs.microsoft.com/en-us/azure/governance/policy/tutorials/create-and-manage
Question 89
You need to recommend the appropriate purchasing model and deployment option for the 30 new databases. The solution must meet the technical requirements and the business requirements. What should you recommend? To answer, select the appropriate options in the answer area. NOTE: Each correct selection is worth one point.
Note: This question is part of a series of questions that present the same scenario. Each question in the series contains a unique solution that might meet the stated goals. Some question sets might have more than one correct solution, while others might not have a correct solution. After you answer a question in this section, you will NOT be able to return to it. As a result, these questions will not appear in the review screen. You have an Azure Data Lake Storage account that contains a staging zone. You need to design a daily process to ingest incremental data from the staging zone, transform the data by executing an R script, and then insert the transformed data into a data warehouse in Azure Synapse Analytics. Solution: You use an Azure Data Factory schedule trigger to execute a pipeline that executes an Azure Databricks notebook, and then inserts the data into the data warehouse. Does this meet the goal?
Correct Answer: B
Section: [none] Explanation: If you need to transform data in a way that is not supported by Data Factory, you can create a custom activity, not an Azure Databricks notebook, with your own data processing logic and use the activity in the pipeline. You can create a custom activity to run R scripts on your HDInsight cluster with R installed. Reference: https://docs.microsoft.com/en-US/azure/data-factory/transform-data