Braindump2go Microsoft Exam 70-534 Updated For Free Download (51-60)

All Latest Updated Questions and Answers in Braindump2go 70-534 Exam Dumps will not take you a lot of time to comprehend and you can easily cover up the entire Microsoft 70-534 syllabus for your examination.Download Braindump2go Free 70-534 Sample Questions Now, Pass 70-534 Exam in advance!

Vendor: Microsoft
Exam Code: 70-534
Exam Name: Architecting Microsoft Azure Solutions

Keywords: 70-534 Exam Dumps,70-534 Practice Tests,70-534 Practice Exams,70-534 Exam Questions,70-534 PDF,70-534 VCE Free,70-534 Book,70-534 E-Book,70-534 Study Guide,70-534 Braindump,70-534 Prep Guide

QUESTION 51
You are designing an Azure application that provides online backup storage for hundreds of media files. Each file is larger than 1GB.
The data storage solution has the following requirements:
– It must be capable of storing an average of 1TB of data for each user.
– It must support sharing of data between all Windows Azure instances.
– It must provide random read/write access.
You need to recommend a durable data storage solution.
What should you recommend?

A.    Azure Drive
B.    Azure Page Blob service
C.    Azure Block Blob service
D.    Local storage on an Azure instance

Answer: D

QUESTION 52
You are designing an Azure web application.
All users must authenticate by using Active Directory Domain Services (AD DS) credentials.
You need to recommend an approach to enable single sign-on to the application for domain-authenticated users.
Which two actions should you recommend? Each correct answer presents part of the solution.

A.    Use Forms authentication to generate claims.
B.    Use the SQL membership provider in the web application.
C.    Use Windows Identity Foundation in the web application.
D.    Use Active Directory Federation Services (AD FS) to generate claims.

Answer: AD

Case Study 2 – Trey Research (Question 53 – Question 57)
Background
Overview
Trey Research conducts agricultural research and sells the results to the agriculture and food industries. The company uses a combination of on-premises and third-party server clusters to meet its storage needs. Trey Research has seasonal demands on its services, with up to 50 percent drops in data capacity and bandwidth demand during low-demand periods. They plan to host their websites in an agile, cloud environment where the company can deploy and remove its websites based on its business requirements rather than the requirements of the hosting company.
A recent fire near the datacenter that Trey Research uses raises the management team’s awareness of the vulnerability of hosting all of the company’s websites and data at any single location. The management team is concerned about protecting its data from loss as a result of a disaster.
Websites
Trey Research has a portfolio of 300 websites and associated background processes that are currently hosted in a third-party datacenter. All of the websites are written in ASP.NET, and the background processes use Windows Services. The hosting environment costs Trey Research approximately S25 million in hosting and maintenance fees.
Infrastructure
Trey Research also has on-premises servers that run VMs to support line-of-business applications. The company wants to migrate the line-of-business applications to the cloud, one application at a time. The company is migrating most of its production VMs from an aging VMWare ESXi farm to a Hyper-V cluster that runs on Windows Server 2012.
Applications
DistributionTracking
Trey Research has a web application named Distribution!racking. This application constantly collects realtime data that tracks worldwide distribution points to customer retail sites. This data is available to customers at all times. The company wants to ensure that the distribution tracking data is stored at a location that is geographically close to the customers who will be using the information. The system must continue running in the event of VM failures without corrupting data. The system is processor intensive and should be run in a multithreading environment.
HRApp
The company has a human resources (HR) application named HRApp that stores data in an on-premises SQL Server database. The database must have at feast two copies, but data to support backups and business continuity must stay in Trey Research locations only.
The data must remain on-premises and cannot be stored in the cloud.
HRApp was written by a third party, and the code cannot be modified. The human resources data is used by all business offices, and each office requires access to the entire database. Users report that HRApp takes all night to generate the required payroll reports, and they would like to reduce this time.
MetricsTracking
Trey Research has an application named MetricsTracking that is used to track analytics for the DistributionTracking web application. The data MetricsTracking collects is not customer-facing. Data is stored on an on-premises SQL Server database, but this data should be moved to the cloud. Employees at other locations access this data by using a remote desktop connection to connect to the application, but latency issues degrade the functionality.
Trey Research wants a solution that allows remote employees to access metrics data without using a remote desktop connection. MetricsTracking was written in-house, and the development team is available to make modifications to the application if necessary. However, the company wants to continue to use SQL Server for MetricsTracking.
Business Requirements
Business Continuity
You have the following requirements:
– Move all customer-facing data to the cloud.
– Web servers should be backed up to geographically separate locations, If one website becomes unavailable, customers should automatically be routed to websites that are still operational.
– Data must be available regardless of the operational status of any particular website.
– The HRApp system must remain on-premises and must be backed up.
– The Met ricsTrac king data must be replicated so that it is locally available to all Trey Research offices.
Auditing and Security
You have the following requirements:
– Both internal and external consumers should be able to access research results.
– Internal users should be able to access data by using their existing company credentials without requiring multiple logins.
– Consumers should be able to access the service by using their Microsoft credentials.
– Applications written to access the data must be authenticated. Access and activity must be monitored and audited.
– Ensure the security and integrity of the data collected from the worldwide distribution points for the distribution tracking application.
Storage and Processing
You have the following requirements:
– Provide real-time analysis of distribution tracking data by geographic location.
– Collect and store large datasets in real-time data for customer use.
– Locate the distribution tracking data as close to the central office as possible to improve bandwidth.
– Co-locate the distribution tracking data as close to the customer as possible based on the customer’s location.
– Distribution tracking data must be stored in the JSON format and indexed by metadata that is stored in a SQL Server database.
– Data in the cloud must be stored in geographically separate locations, but kept with the same political boundaries.
Technical Requirements
Migration
You have the following requirements:
– Deploy all websites to Azure.
– Replace on-premises and third-party physical server clusters with cloud-based solutions.
– Optimize the speed for retrieving exiting JSON objects that contain the distribution tracking data.
Recommend strategies for partitioning data for load balancing.
Auditing and Security
You have the following requirements:
– Use Active Directory for internal and external authentication.
– Use OAuth for application authentication.
Business Continuity
You have the following requirements:
– Data must be backed up to separate geographic locations.
– Web servers must run concurrent versions of all websites in distinct geographic locations.
– Use Azure to back up the on-premises MetricsTracking data.
– Use Azure virtual machines as a recovery platform for MetricsTracking and HRApp.
– Ensure that there is at least one additional on-premises recovery environment for the HRApp.

QUESTION 53
Drag and Drop Question
You need to recommend a test strategy for the disaster recovery system.
What should you do? To answer, drag the appropriate test strategy to the correct business application. Each test strategy may be used once, more than once, or not at all. You may need to drag the split bar between panes or scroll to view content.


Answer:
 

QUESTION 54
Hotspot Question
You need to plan the business continuity strategy.
For each requirement, what should you recommend? To answer, select the appropriate option from each list in the answer area.
 
Answer:
 

QUESTION 55
Drag and Drop Question
You need to ensure that customer data is secured both in transit and at rest.
Which technologies should you recommend? To answer, drag the appropriate technology to the correct security requirement. Each technology may be used once, more than once, or not at all. You may need to drag the split bar between panes or scroll to view content.
 
Answer:
 

QUESTION 56
You need to configure the distribution tracking application.
What should you do?

A.    Map each role to a single upgrade domain to optimize resource utilization.
B.    Design all services as stateless services.
C.    Configure operations to queue when a role reaches its capacity.
D.    Configure multiple worker roles to run on each virtual machine.

Answer: C

QUESTION 57
Hotspot Question
You need to design a data storage strategy for each application.
In the table below, identify the strategy that you should use for each application. Make only one selection in each column.
 
Answer:
 

Case Study 3 – Contoso, Ltd (Question 58 – Question 62)
Background
Overview
Contoso, Ltd., manufactures and sells golf clubs and golf balls. Contoso also sells golf accessories under the Contoso Golf and Odyssey brands worldwide.
Most of the company’s IT infrastructure is located in the company’s Carlsbad, California, headquarters. Contoso also has a sizable third-party colocation datacenter that costs the company USD $30,000 to $40,000 a month. Contoso has other servers scattered around the United States.
Contoso, Ltd., has the following goals:
Move many consumer-facing websites, enterprise databases, and enterprise web services to Azure.
Improve the performance for customers and resellers who are access company websites from around the world.
Provide support for provisioning resources to meet bursts of demand. Consolidate and improve the utilization of website- and database-hosting resources.
Avoid downtime, particularly that caused by web and database server updating. Leverage familiarity with Microsoft server management tools.
Infrastructure
Contoso’s datacenters are filled with dozens of smaller web servers and databases that run on under-utilized hardware. This creates issues for data backup. Contoso currently backs up data to tape by using System Center Data Protection Manager. System Center Operations Manager is not deployed in the enterprise.
All of the servers are expensive to acquire and maintain, and scaling the infrastructure takes significant time. Contoso conducts weekly server maintenance, which causes downtime for some of its global offices. Special events, such as high-profile golf tournaments, create a large increase in site traffic. Contoso has difficulty scaling the web- hosting environment fast enough to meet these surges in site traffic.
Contoso has resellers and consumers in Japan and China. These resellers must use applications that run in a datacenter that is located in the state of Texas, in the United States. Because of the physical distance, the resellers experience slow response times and downtime.
Business Requirements
Management and Performance
Management
Web servers and databases must automatically apply updates to the operating system and products.
Automatically monitor the health of worldwide sites, databases, and virtual machines.
Automatically back up the website and databases. Manage hosted resources by using on-premises tools.
Performance
The management team would like to centralize data backups and eliminate the use of tapes.
The website must automatically scale without code changes or redeployment. Support changes in service tier without reconfiguration or redeployment. Site-hosting must automatically scale to accommodate data bandwidth and number of connections.
Scale databases without requiring migration to a larger server. Migrate business critical applications to Azure. Migrate databases to the cloud and centralize databases where possible.
Business Continuity and Support
Business Continuity
Minimize downtime in the event of regional disasters. Recover data if unintentional modifications or deletions are discovered. Run the website on multiple web server instances to minimize downtime and support a high service level agreement (SLA).
Connectivity
Allow enterprise web services to access data and other services located on- premises.
Provide and monitor lowest latency possible to website visitors. Automatically balance traffic among all web servers. Provide secure transactions for users of both legacy and modern browsers. Provide automated auditing and reporting of web servers and databases.
Support single sign-on from multiple domains.
Development Environment
You identify the following requirements for the development environment:
– Support the current development team’s knowledge of Microsoft web development and SQL Service tools.
– Support building experimental applications by using data from the Azure deployment and on-premises data sources.
– Mitigate the need to purchase additional tools for monitoring and debugging.
– System designers and architects must be able to create custom Web APIs without requiring any coding.
– Support automatic website deployment from source control.
– Support automated build verification and testing to mitigate bugs introduced during builds.
– Manage website versions across all deployments.
– Ensure that website versions are consistent across ail deployments.
Technical Requirement
Management and Performance
Management
Use build automation to deploy directly from Visual Studio. Use build-time versioning of assets and builds/releases. Automate common IT tasks such as VM creation by using Windows PowerShell workflows.
Use advanced monitoring features and reports of workloads in Azure by using existing Microsoft tools.
Performance
Websites must automatically load balance across multiple servers to adapt to varying traffic.
In production, websites must run on multiple instances. First-time published websites must be published by using Visual Studio and scaled to a single instance to test publishing.
Data storage must support automatic load balancing across multiple servers. Websites must adapt to wide increases in traffic during special events. Azure virtual machines (VMs) must be created in the same datacenter when applicable.
Business Continuity and Support
Business Continuity
Automatically co-locate data and applications in different geographic locations. Provide real-time reporting of changes to critical data and binaries. Provide real-time alerts of security exceptions. Unwanted deletions or modifications of data must be reversible for up to one month, especially in business critical applications and databases. Any cloud-hosted servers must be highly available.
Enterprise Support
The solution must use stored procedures to access on-premises SQL Server data from Azure.
A debugger must automatically attach to websites on a weekly basis. The scripts that handle the configuration and setup of debugging cannot work if there is a delay in attaching the debugger.

QUESTION 58
Drag and Drop Question
You need to deploy the virtual machines to Azure.
Which four Azure PowerShell scripts should you run in sequence? To answer, move the appropriate scripts from the list of scripts to the answer area and arrange them in the correct order.
 
Answer:
 

QUESTION 59
You need to recommend a solution for publishing one of the company websites to Azure and configuring it for remote debugging.
Which two actions should you perform? Each correct answer presents part of the solution.

A.    From Visual Studio, attach the debugger to the solution.
B.    Set the application logging level to Verbose and enable logging.
C.    Set the Web Server logging level to Information and enable logging.
D.    Set the Web Server logging level to Verbose and enable logging.
E.    From Visual Studio, configure the site to enable Debugger Attaching and then publish the site.

Answer: AD

QUESTION 60
Drag and Drop Question
You need to recommend network connectivity solutions for the experimental applications.
What should you recommend? To answer, drag the appropriate solution to the correct network connection requirements. Each solution may be used once, more than once, or not at all. You may need to drag the split bar between panes or scroll to view content.
 
Answer:
 


Thanks For Trying Braindump2go Latest Microsoft 70-534 Dumps Questions! Braindump2go Exam Dumps ADVANTAGES:
☆ 100% Pass Guaranteed Or Full Money Back!
☆ Instant Download Access After Payment!
☆ One Year Free Updation!
☆ Well Formated: PDF,VCE,Exam Software!
☆ Multi-Platform capabilities – Windows, Laptop, Mac, Android, iPhone, iPod, iPad.
☆ Professional, Quick,Patient IT Expert Team 24/7/375 Onlinen Help You!
☆ We served more than 35,000 customers all around the world in last 5 years with 98.99% PASS RATE!
☆ Guaranteed Secure Shopping! Your Transcations are protected by Braindump2go all the time!
☆ Pass any exams at the FIRST try!

http://www.braindump2go.com/70-534.html

         

Comments are closed.