Exam Code: 70-535 (Practice Exam Latest Test Questions VCE PDF)
Exam Name: Architecting Microsoft Azure Solutions
Certification Provider: Microsoft
Free Today! Guaranteed Training- Pass 70-535 Exam.
♥♥ 2021 NEW RECOMMEND ♥♥
Free VCE & PDF File for Microsoft 70-535 Real Exam (Full Version!)
★ Pass on Your First TRY ★ 100% Money Back Guarantee ★ Realistic Practice Exam Questions
Free Instant Download NEW 70-535 Exam Dumps (PDF & VCE):
Available on:
http://www.surepassexam.com/70-535-exam-dumps.html
P.S. Precise 70-535 bible are available on Google Drive, GET MORE: https://drive.google.com/open?id=1lOIcwe4sj0Mb2GbvXvAdbPWJdlMgS-j5
New Microsoft 70-535 Exam Dumps Collection (Question 4 - Question 13)
New Questions 4
You are designing an Azure application that will use a worker role. The worker role will create temporary files.
You need to minimize storage transaction charges. Where should you create the files?
A. In Azure local storage
B. In Azure Storage page blobs
C. On an Azure Drive
D. In Azure Storage block blobs
Answer: A
Explanation:
Local storage is temporary in Azure. So, if the virtual machine supporting your role dies and cannot recover, your local storage is lost! Therefore, Azure developers will tell you, only volatile data should ever be stored in local storage of Azure.
References: http://www.intertech.com/Blog/windows-azure-local-file-storage-how-to-guide-
and-warnings/
http://blog.codingoutloud.com/2011/06/12/azure-faq-can-i-write-to-the-file-system-on- windows-azure/
New Questions 5
A company has a very large dataset that includes sensitive information. The dataset is over 30 TB in size.
You have a standard business-class ISP internet connection that is rated at 100 megabits/second.
You have 10 4-TB hard drives that are approved to work with the Azure Import/Export Service.
You need to migrate the dataset to Azure. The solution must meet the following requirements:
* The dataset must be transmitted securely to Azure.
* Network bandwidth must not increase.
* Hardware costs must be minimized.
What should you do?
A. Prepare the drives with the Azure Import/Export tool and then create the import job. Ship the drives to Microsoft via a supported carrier service.
B. Create an export job and then encrypt the data on the drives by using the Advanced Encryption Standard (AES). Create a destination Blob to store the export data.
C. Create an import job and then encrypt the data on the drives by using the Advanced Encryption Standard (AES). Create a destination Blob to store the import data.
D. Prepare the drives by using Sysprep.exe and then create the import job. Ship the drives to Microsoft via a supported carrier service.
Answer: A
Explanation:
You can use the Microsoft Azure Import/Export service to transfer large amounts of file data to Azure Blob storage in situations where uploading over the network is prohibitively expensive or not feasible.
References: http://azure.microsoft.com/en-gb/documentation/articles/storage-import- export-service/
New Questions 6
You administer an Azure Storage account with a blob container. You enable Storage account logging for read, write and delete requests. You need to reduce the costs associated with storing the logs. What should you do?
A. Execute Delete Blob requests over https.
B. Create an export job for your container.
C. Set up a retention policy.
D. Execute Delete Blob requests over http.
Answer: C
Explanation:
There are two ways to delete Storage Analytics data: by manually making deletion requests or by setting a data retention policy. Manual requests to delete Storage Analytics data are billable, but delete requests resulting from a retention policy are not billable.
References: https://docs.microsoft.com/en-us/rest/api/storageservices/Setting-a-Storage-Analytics-Data-Retention-Policy?redirectedfrom=MSDN
New Questions 7
You are designing an Azure application that provides online backup storage for hundreds of media files. Each file is larger than 1GB.
The data storage solution has the following requirements:
* Be capable of storing an average of 2 terabytes (TB) of data for each user.
* Support sharing of data between all Microsoft Azure instances
* Provide random read/write access
You need to recommend a durable data storage solution. What should you recommend?
A. Store data in the VHD file
B. Azure Page Blob service
C. Azure Block Blob service
D. Local storage on the VM
Answer: B
Explanation:
References: https://msdn.microsoft.com/en-us/library/azure/ee691964.aspx
New Questions 8
An application currently resides on an on-premises virtual machine that has 2 CPU cores, 4 GB of RAM, 20 GB of hard disk space, and a 10 megabit/second network connection.
You plan to migrate the application to Azure. You have the following requirements:
* You must not make changes to the application.
* You must minimize the costs for hosting the application.
You need to recommend the appropriate virtual machine instance type. Which virtual machine tier should you recommend?
A. Network Optimized (A Series)
B. General Purpose Compute, Basic Tier (A Series)
C. General Purpose Compute, Standard Tier (A Series)
D. Optimized Compute (D Series)
Answer: B
Explanation:
General purpose compute: Basic tier
An economical option for development workloads, test servers, and other applications that don't require load balancing, auto-scaling, or memory-intensive virtual machines.
CPU core range: 1-8 RAM range: 0.75 u2013 14 GB
Disk size: 20-240 GB References:
New Questions 9
A company has 10 on-premises SQL databases. The company plans to move the databases to SQL Server 2012 that runs in Azure Infrastructure-as-a-Service (IaaS). After migration, the databases will support a limited number of Azure websites in the same Azure Virtual Network.
You have the following requirements:
* You must restore copies of existing on-premises SQL databases to the SQL servers that
run in Azure IaaS.
* You must be able to manage the SQL databases remotely.
* You must not open a direct connection from all of the machines on the on-premises network to Azure.
* Connections to the databases must originate from only five Windows computers.
You need to configure remote connectivity to the databases. Which technology solution should you implement?
A. Azure Virtual Network site-to-site VPN
B. Azure Virtual Network multi-point VPN
C. Azure Virtual Network point-to-site VPN
D. Azure ExpressRoute
Answer: C
Explanation:
A point-to-site VPN would meet the requirements.
References: https://azure.microsoft.com/en-us/documentation/articles/vpn-gateway-point- to-site-create/
New Questions 10
You manage a web application published to Azure Cloud Services. Your service level agreement (SLA) requires that you are notified in the event of poor performance from customer locations in the US, Asia, and Europe.
You need to configure the Azure Management Portal to notify you when the SLA performance targets are not met. What should you do?
A. Create an alert rule to monitor web endpoints
B. Create a Notification Hub alert with response time metrics.
C. Add an endpoint monitor and alert rule to the Notification Hub.
D. Configure the performance counter on the cloud service.
Answer: A
Explanation:
References: https://docs.microsoft.com/en-us/azure/monitoring-and-diagnostics/insights- alerts-portal
New Questions 11
Your company network includes two branch offices. Users at the company access internal virtual machines (VMs). You want to ensure secure communications between the branch offices and the internal VMs and network.
You need to create a site-to-site VPN connection. What are two possible ways to achieve this goal? Each correct answer presents a complete solution
A. a private IPv4 IP address and a compatible VPN device
B. a private IPv4 IP address and a RRAS running on Windows Server 2012
C. a public-facing IPv4 IP address and a compatible VPN device
D. a public-facing IPv4 IP address and a RRAS running on Windows Server 2012
Answer: C,D
Explanation:
You must have an externally facing IPv4 IP address and a VPN device or RRAS to configure a site-to-site VPN connection.
References: https://docs.microsoft.com/en-us/azure/vpn-gateway/vpn-gateway-vpn-faq
New Questions 12
You have several virtual machines (VMs) that run in Azure. You also have a single System Center 2012 R2 Configuration Manager (SCCM) primary site on-premises.
You have the following requirements:
* All VMs must run on the same virtual network.
* Network traffic must be minimized between the on-premises datacenter and Azure.
* The solution minimize complexity.
You need to use SCCM to collect inventory and deploy software to Azure VMs. What should you do first?
A. Configure client push for the Azure virtual network.
B. Enable and configure Operations Insights in Azure.
C. Install a cloud distribution point on an Azure VM.
D. Install a secondary site underneath the primary site onto an Azure VM.
Answer: C
Explanation:
Cloud-based distribution Point, a Configuration Manager Site System Role in the Cloud
Much of the Configuration Manager topology is made up of distribution points, they are very helpful in many situations where bandwidth and geographical separation are the facts of life, but also hard to manage if you have hundreds or even thousands of them.
This feature started with the vision that it makes perfect sense to have big distribution points in the Windows Azure cloud where one should not worry about things like (but not limited to) size, performance, reliability, security, access from all around the world, hardware/software update issues etc.
Note: Content management in System Center 2012 Configuration Manager provides the tools for you to manage content files for applications, packages, software updates, and operating system deployment. Configuration Manager uses distribution points to store files that are required for software to run on client computers. These distribution points function as distribution centers for the content files and let users download and run the software. Clients must have access to at least one distribution point from which they can download the files.
References: http://blogs.technet.com/b/configmgrteam/archive/2013/01/31/new-distribution-points-in-configuration-manager-sp1.aspx
New Questions 13
Case Study
This is a case study. Case studies are not timed separately. You can use as much exam time as you would like to complete each case. However, there may be additional case studies and sections on this exam. You must manage your time to ensure that you are able to complete all questions included on this exam in the time provided.
To answer the questions included in a case study, you will need to reference information that is provided in the case study. Case studies might contain exhibits and other resources that provide more information about the scenario that is described in the case study. Each question is independent of the other question on this case study.
At the end of this case study, a review screen will appear. This screen allows you to review your answers and to make changes before you move to the next sections of the exam. After you begin a new section, you cannot return to this section.
To start the case study
To display the first question on this case study, click the Next button. Use the buttons in the left pane to explore the content of the case study before you answer the questions. Clicking these buttons displays information such as business requirements, existing environment, and problem statements. If the case study has an All Information tab, note that the information displayed is identical to the information displayed on the subsequent tabs. When you are ready to answer a question, click the Question button to return to the question.
Background Overview
Woodgrove Bank has 20 regional offices and operates 1,500 branch office locations. Each regional office hosts the servers, infrastructure, and applications that support that region.
Woodgrove Bank plans to move all of their on-premises resources to Azure, including virtual machine (VM)-based, line-of-business workloads, and SQL databases. You are the owner of the Azure subscription that Woodgrove Bank is using. Your team is using Git repositories hosted on GitHub for source control.
Security
Currently, Woodgrove Banku2021s Computer Security Incident Response Team (CSIRT) has a problem investigating security issues due to the lack of security intelligence integrated with their current incident response tools. This lack of integration introduces a problem during the detection (too many false positives), assessment, and diagnose stages. You decide to use Azure Security Center to help address this problem.
Woodgrove Bank has several apps with regulates data such as Personally Identifiable Information (PII) that require a higher level of security. All apps are currently secured by using an on-premises Active Directory Domain Services (ADDS). The company depends on following mission-critical apps: WGBLoanMaster, WGBLeaseLeader, and WGBCreditCruncher apps. You plan to move each of these apps to Azure as part of an app migration project.
Apps
The WGBLoanMaster app has been audited for transaction loss. Many transactions have been lost is processing and monetary write-offs have cost the bank. The app runs on two VMs that include several public endpoints.
The WGBLeaseLeader app has been audited for several data breaches. The app includes a SQL Server database and a web-based portal. The portal uses an ASP.NET Web API
function to generate a monthly aggregate report from the database.
The WGBCreditCruncher app runs on a VM and is load balanced at the network level. The app includes several stateless components and must accommodate scaling of increased credit processing. The app runs on a nightly basis to process credit transactions that are batched during the day. The app includes a web-based portal where customers can check their credit information. A mobile version of the app allows users to upload check images.
Business Requirements WGBLoanMaster app
The app audit revealed a need for zero transaction loss. The business is losing money due to the app losing and not processing loan information. In addition, transactions fail to process after running for a long time. The business has requested the aggregation processing to be scheduled for 01:00 to prevent system slowdown.
WGBLeaseLeader app
The app should be secured to stop data breaches. If the data is breached, it must not be readable. The app is continuing to see increased volume and the business does not want the issues presented in the WGBLoanMaster app. Transaction loss is unacceptable, and although the lease monetary amounts are smaller than loans, they are still an important profit center for Woodgrove Bank. The business would also like the monthly report to be automatically generated on the first of the month. Currently, a user must log in to the portal and click a button to generate the report.
WGBCreditCruncher app
The web-based portal area of the app must allow users to sign in with their Facebook credentials. The bank would like to allow this feature to enable more users to check their credit within the app.
Woodgrove Bank needs to develop a new financial risk modeling feature that they can include in the WGBCreditCruncher app. The financial risk modeling feature has not been developed due to costs associated with processing, transforming, and analyzing the large volumes of data that are collected. You need to find a way to implement parallel processing to ensure that the features run efficiently, reliably, and quickly. The feature must scale based on computing demand to process the large volumes of data and output several financial risk models.
Technical Requirements
WGBLoanMaster app
The app uses several compute-intensive tasks that create long-running requests to the system. The app is critical to the business and must be scalable to increased loan processing demands. The VMs that run the app include a Windows Task Scheduler task that aggregates loan information from the app to send to a third party. This task runs a console app on the VM.
The app requires a messaging system to handle transaction processing. The messaging system must meet the following requirements:
Allow messages to reside in the queue for up to a month. Be able to publish and consume batches of messages.
Allow full integration with the Windows Communication Foundation (WCF) communication stack.
Provide a role-based access model to the queues, including different permissions for senders and receivers.
You develop an Azure Resource Manager (ARM) template to deploy the VMs used to support the app. The template must be deployed to a new resource group and you must validate your deployment settings before creating actual resources.
WGBLeaseLeader app
The app must use Azure SQL Databases as a replacement to the current Microsoft SQL Server environment. The monthly report must be automatically generated.
The app requires a messaging system to handle transaction processing. The messaging system must meet the following requirements:
Require server-side logs of all of the transactions run against your queues. Track progress of a message within the queue.
Process the messages within 7 days.
Provide a differing timeout value per message.
WGBCreditCruncher app
The app must:
Secure inbound and outbound traffic.
Analyze inbound network traffic for vulnerabilities.
Use an instance-level public IP and allow web traffic on port 443 only.
Upgrade the portal to a Single Page Application (SPA) that uses JavaScript, Azure Active Directory (Azure AD), and the OAuth 2.0 implicit authorization grant to secure the Web API back end.
Cache authentication and host the Web API back end using the Open Web Interface for
.NET (OWIN) middleware.
Immediately compress check images received from the mobile web app. Schedule processing of the batched credit transactions on a nightly basis.
Provide parallel processing and scalable computing resources to output financial risk models.
Use simultaneous computer nodes to enable high performance computing and updating of the financial risk models.
Key security area
You need to recommend a business continuity and disaster recovery solution for all of the existing line of business applications.
What are two ways to achieve the goal? Each correct answer presents a complete solution.
A. Create new virtual machines (VMs) in Azure and migrate the line of business applications to the VMs. Migrate any backend databases to SQL Database.
B. Migrate the virtual machines to the Hyper-V cluster and enable Hyper-V replica.
C. Configure ExpressRoute to enable migration to Azure.
D. Install the Azure Backup agent on the virtual machines.
Answer: A,B
Explanation:
References:
https://docs.microsoft.com/en-us/azure/site-recovery/site-recovery-hyper-v-site-to-azure
Recommend!! Get the Precise 70-535 dumps in VCE and PDF From Examcollection, Welcome to download: http://www.examcollectionuk.com/70-535-vce-download.html (New Q&As Version)
