P.S. Printable 70-535 faq are available on Google Drive, GET MORE: https://drive.google.com/open?id=1QGQh8lSQv2kpYQewvx2Fa025vtCRw5Vh
Question No: 1
You need to ensure that the repository runs the script when new software is released. Which technology should you use?
A. Azure Function
B. Azure App Service Logic App
C. Azure App Service API App
D. Azure Webjob
Answer: B
Question No: 2
You have several virtual machines (VMs) that run in Azure. You also have a single System Center 2012 R2 Configuration Manager (SCCM) primary site on-premises.
You have the following requirements:
* All VMs must run on the same virtual network.
* Network traffic must be minimized between the on-premises datacenter and Azure.
* The solution minimize complexity.
You need to use SCCM to collect inventory and deploy software to Azure VMs. What should you do first?
A. Configure client push for the Azure virtual network.
B. Enable and configure Operations Insights in Azure.
C. Install a cloud distribution point on an Azure VM.
D. Install a secondary site underneath the primary site onto an Azure VM.
Answer: C
Explanation:
Cloud-based distribution Point, a Configuration Manager Site System Role in the Cloud
Much of the Configuration Manager topology is made up of distribution points, they are very helpful in many situations where bandwidth and geographical separation are the facts of life, but also hard to manage if you have hundreds or even thousands of them.
This feature started with the vision that it makes perfect sense to have big distribution points in the Windows Azure cloud where one should not worry about things like (but not limited to) size, performance, reliability, security, access from all around the world, hardware/software update issues etc.
Note: Content management in System Center 2012 Configuration Manager provides the tools for you to manage content files for applications, packages, software updates, and operating system deployment. Configuration Manager uses distribution points to store files that are required for software to run on client computers. These distribution points function as distribution centers for the content files and let users download and run the software. Clients must have access to at least one distribution point from which they can download the files.
References: http://blogs.technet.com/b/configmgrteam/archive/2013/01/31/new-distribution-points-in-configuration-manager-sp1.aspx
Question No: 3
Case Study
This is a case study. Case studies are not timed separately. You can use as much exam time as you would like to complete each case. However, there may be additional case studies and sections on this exam. You must manage your time to ensure that you are able to complete all questions included on this exam in the time provided.
To answer the questions included in a case study, you will need to reference information that is provided in the case study. Case studies might contain exhibits and other resources that provide more information about the scenario that is described in the case study. Each question is independent of the other question on this case study.
At the end of this case study, a review screen will appear. This screen allows you to review your answers and to make changes before you move to the next sections of the exam. After you begin a new section, you cannot return to this section.
To start the case study
To display the first question on this case study, click the Next button. Use the buttons in the left pane to explore the content of the case study before you answer the questions. Clicking these buttons displays information such as business requirements, existing environment, and problem statements. If the case study has an All Information tab, note that the information displayed is identical to the information displayed on the subsequent tabs. When you are ready to answer a question, click the Question button to return to the question.
Background
Overview
Woodgrove Bank has 20 regional offices and operates 1,500 branch office locations. Each regional office hosts the servers, infrastructure, and applications that support that region.
Woodgrove Bank plans to move all of their on-premises resources to Azure, including virtual machine (VM)-based, line-of-business workloads, and SQL databases. You are the owner of the Azure subscription that Woodgrove Bank is using. Your team is using Git repositories hosted on GitHub for source control.
Security
Currently, Woodgrove Banku2021s Computer Security Incident Response Team (CSIRT) has a problem investigating security issues due to the lack of security intelligence integrated with their current incident response tools. This lack of integration introduces a problem during the detection (too many false positives), assessment, and diagnose stages. You decide to use Azure Security Center to help address this problem.
Woodgrove Bank has several apps with regulates data such as Personally Identifiable Information (PII) that require a higher level of security. All apps are currently secured by using an on-premises Active Directory Domain Services (ADDS). The company depends on following mission-critical apps: WGBLoanMaster, WGBLeaseLeader, and WGBCreditCruncher apps. You plan to move each of these apps to Azure as part of an app migration project.
Apps
The WGBLoanMaster app has been audited for transaction loss. Many transactions have been lost is processing and monetary write-offs have cost the bank. The app runs on two VMs that include several public endpoints.
The WGBLeaseLeader app has been audited for several data breaches. The app includes a SQL Server database and a web-based portal. The portal uses an ASP.NET Web API function to generate a monthly aggregate report from the database.
The WGBCreditCruncher app runs on a VM and is load balanced at the network level. The app includes several stateless components and must accommodate scaling of increased credit processing. The app runs on a nightly basis to process credit transactions that are batched during the day. The app includes a web-based portal where customers can check their credit information. A mobile version of the app allows users to upload check images.
Business Requirements
WGBLoanMaster app
The app audit revealed a need for zero transaction loss. The business is losing money due to the app losing and not processing loan information. In addition, transactions fail to process after running for a long time. The business has requested the aggregation processing to be scheduled for 01:00 to prevent system slowdown.
WGBLeaseLeader app
The app should be secured to stop data breaches. If the data is breached, it must not be readable. The app is continuing to see increased volume and the business does not want the issues presented in the WGBLoanMaster app. Transaction loss is unacceptable, and although the lease monetary amounts are smaller than loans, they are still an important profit center for Woodgrove Bank. The business would also like the monthly report to be automatically generated on the first of the month. Currently, a user must log in to the portal and click a button to generate the report.
WGBCreditCruncher app
The web-based portal area of the app must allow users to sign in with their Facebook credentials. The bank would like to allow this feature to enable more users to check their credit within the app.
Woodgrove Bank needs to develop a new financial risk modeling feature that they can include in the WGBCreditCruncher app. The financial risk modeling feature has not been developed due to costs associated with processing, transforming, and analyzing the large volumes of data that are collected. You need to find a way to implement parallel processing to ensure that the features run efficiently, reliably, and quickly. The feature must scale based on computing demand to process the large volumes of data and output several financial risk models.
Technical Requirements WGBLoanMaster app
The app uses several compute-intensive tasks that create long-running requests to the system. The app is critical to the business and must be scalable to increased loan processing demands. The VMs that run the app include a Windows Task Scheduler task that aggregates loan information from the app to send to a third party. This task runs a console app on the VM.
The app requires a messaging system to handle transaction processing. The messaging system must meet the following requirements:
Allow messages to reside in the queue for up to a month. Be able to publish and consume batches of messages.
Allow full integration with the Windows Communication Foundation (WCF) communication stack.
Provide a role-based access model to the queues, including different permissions for senders and receivers.
You develop an Azure Resource Manager (ARM) template to deploy the VMs used to support the app. The template must be deployed to a new resource group and you must validate your deployment settings before creating actual resources.
WGBLeaseLeader app
The app must use Azure SQL Databases as a replacement to the current Microsoft SQL Server environment. The monthly report must be automatically generated.
The app requires a messaging system to handle transaction processing. The messaging system must meet the following requirements:
Require server-side logs of all of the transactions run against your queues. Track progress of a message within the queue.
Process the messages within 7 days.
Provide a differing timeout value per message.
WGBCreditCruncher app The app must:
Secure inbound and outbound traffic.
Analyze inbound network traffic for vulnerabilities.
Use an instance-level public IP and allow web traffic on port 443 only.
Upgrade the portal to a Single Page Application (SPA) that uses JavaScript, Azure Active Directory (Azure AD), and the OAuth 2.0 implicit authorization grant to secure the Web API back end.
Cache authentication and host the Web API back end using the Open Web Interface for
.NET (OWIN) middleware.
Immediately compress check images received from the mobile web app. Schedule processing of the batched credit transactions on a nightly basis.
Provide parallel processing and scalable computing resources to output financial risk models.
Use simultaneous computer nodes to enable high performance computing and updating of the financial risk models.
Key security area
You need to run the script for a new release. Which technology should you use?
A. Azure WebJob
B. Azure App Service API App
C. Azure Function
D. Azure App Service Logic App
Answer: B
Question No: 4
You are designing a Windows Azure application that will store data in two SQL Azure databases. The application will insert data in both databases as part of a single logical operation.
You need to recommend an approach for maintaining data consistency across the databases. What should you recommend?
A. Execute database calls on parallel threads
B. Wrap the database calls in a single transaction scope.
C. Use Microsoft Distributed Transaction Coordinator (MSDTC).
D. Handle errors resulting from the database calls by using compensatory logic.
Answer: B
Explanation:
The TransactionScope class establishes an ambient transaction in .NET. (An u201cambient transactionu201d is one that lives in the current thread.) All connections opened within the TransactionScope participate in the transaction. If different databases participate, the transaction is automatically elevated to a distributed transaction. The outcome of the transaction is controlled by setting the scope to complete to indicate a commit.
Note: Elastic database transactions for SQL DB enable applications to make atomic changes to data stored in several different SQL Databases.
Elastic database transactions targets the following scenarios:
* Multi-database applications in Azure: With this scenario, data is vertically partitioned across several databases in SQL DB such that different kinds of data reside on different databases. Some operations require changes to data which is kept in two or more databases. The application uses elastic database transactions to coordinate the changes across databases and ensure atomicity.
Etc.
References: https://docs.microsoft.com/en-us/azure/sql-database/sql-database-elastic-transactions-overview
Question No: 5
You are designing an Azure application that processes graphical image files. The graphical Images are processed in batches by remote applications that run on multiple servers.
You have the following requirements:
*The application must remain operational during batch-processing operations.
*Users must be able to roll back each image to a previous version.
You need to ensure that each remote application has exclusive access to an image while the application processes the image. Which type of storage should you use to store the images?
A. Table service
B. Queue service
C. Blob service
D. A single Azure VHD that is attached to the web role
Answer: C
Explanation:
* Blob Leases allow you to claim ownership to a Blob. Once you have the lease you can then update the Blob or delete the Blob without worrying about another process changing it underneath you. When a Blob is leased, other processes can still read it, but any attempt to update it will fail. You can update Blobs without taking a lease first, but you do run the chance of another process also attempting to modify it at the same time.
* You can opt to use either optimistic or pessimistic concurrency models to manage access to blobs and containers in the blob service.
References: http://justazure.com/azure-blob-storage-part-8-blob-leases/ http://www.azurefromthetrenches.com/?p=1371
Question No: 6
DRAG DROPYou need to automate tasks with Azure by using Azure PowerShell workflows.
How should you complete the Azure PowerShell script? To answer, drag the appropriate cmdlet to the correct location. Each cmdlet may be used once, more than once, or not at all. You may need to drag the split bar between panes or scroll to view content.
Answer:
Explanation:
workflow Use-WorkflowCheckpointSample
{
# An exception occurs if 'HasBeenSuspended' does not already exist.
# Exceptions that are not caught with a try/catch will cause the runbook to suspend. Set-AutomationVariable -Name 'HasBeenSuspended' -Value $False
# This line occurs before the checkpoint. When the runbook is resumed after
# suspension, 'Before Checkpoint' will not be output a second time. Write-Output "Before Checkpoint"
# A checkpoint is created. Checkpoint-Workflow
# This line occurs after the checkpoint. The runbook will start here on resume. Write-Output "After Checkpoint"
$HasBeenSuspended = Get-AutomationVariable -Name 'HasBeenSuspended'
# If branch only executes if the runbook has not previously suspended. if (!$HasBeenSuspended) {
Set-AutomationVariable -Name 'HasBeenSuspended' -Value $True
# This will cause a runtime exception. Any runtime exception in a runbook
# will cause the runbook to suspend. 1 + "abc"
}
Write-Output "Runbook Complete"
}
References: https://gallery.technet.microsoft.com/scriptcenter/How-to-use-workflow- cd57324f
Question No: 7
A company hosts a website and exposes web services on the company intranet.
The intranet is secured by using a firewall. Company policies prohibit changes to firewall rules.
Devices outside the firewall must be able to access the web services. You need to recommend an approach to enable inbound communication.
What should you recommend?
A. The Azure Access Control Service
B. Windows Azure Pack
C. The Azure Service Bus
D. A web service in an Azure role that relays data to the internal web services
Answer: C
Explanation:
The Service Bus Relay is designed for the use-case of taking existing Windows Communication Foundation (WCF) web services and making those services securely accessible to solutions that reside outside the corporate perimeter without requiring intrusive changes to the corporate network infrastructure. Such Service Bus relay services are still hosted inside their existing environment, but they delegate listening for incoming sessions and requests to the cloud-hosted Service Bus.
References: http://azure.microsoft.com/en-gb/documentation/articles/cloud-services-dotnet-hybrid-app-using-service-bus-relay/
Question No: 8
A company creates an API and makes it accessible on an Azure website. External partners use the API occasionally. The website uses the Standard web hosting plan.
Partners report that the first API call in a sequence of API calls occasionally takes longer than expected to run. Subsequent API calls consistently perform as expected.
You need to ensure that all API calls perform consistently. What should you do?
A. Configure the website to use the Basic web hosting plan
B. Enable Always On support.
C. Configure the website to automatically scale.
D. Add a trigger to the web.config file for the website that causes the website to recycle periodically.
Answer: B
Explanation:
By default, web apps are unloaded if they are idle for some period of time. This lets the system conserve resources. In Basic or Standard mode, you can enable Always On to keep the app loaded all the time. If your app runs continuous WebJobs or runs WebJobs triggered using a CRON expression, you should enable Always On, or the web jobs may not run reliably.
References: https://docs.microsoft.com/en-us/azure/app-service-web/web-sites-configure
Question No: 9
A company has a very large dataset that includes sensitive information. The dataset is over 30 TB in size.
You have a standard business-class ISP internet connection that is rated at 100 megabits/second.
You have 10 4-TB hard drives that are approved to work with the Azure Import/Export Service.
You need to migrate the dataset to Azure. The solution must meet the following requirements:
* The dataset must be transmitted securely to Azure.
* Network bandwidth must not increase.
* Hardware costs must be minimized.
What should you do?
A. Prepare the drives with the Azure Import/Export tool and then create the import job. Ship the drives to Microsoft via a supported carrier service.
B. Create an export job and then encrypt the data on the drives by using the Advanced Encryption Standard (AES). Create a destination Blob to store the export data.
C. Create an import job and then encrypt the data on the drives by using the Advanced Encryption Standard (AES). Create a destination Blob to store the import data.
D. Prepare the drives by using Sysprep.exe and then create the import job. Ship the drives to Microsoft via a supported carrier service.
Answer: A
Explanation:
You can use the Microsoft Azure Import/Export service to transfer large amounts of file data to Azure Blob storage in situations where uploading over the network is prohibitively expensive or not feasible.
References: http://azure.microsoft.com/en-gb/documentation/articles/storage-import- export-service/
Question No: 10
You administer of a set of virtual machine (VM) guests hosted in Hyper-V on Windows
Server 2012 R2. The virtual machines run the following operating systems:
* Windows Server 2008
* Windows Server 2008 R2
* Linux (openSUSE 13.1)
All guests currently are provisioned with one or more network interfaces with static bindings and VHDX disks. You need to move the VMs to Azure Virtual Machines hosted in an Azure subscription. Which three actions should you perform?
Each correct answer presents part of the solution.
A. Install the WALinuxAgent on Linux servers.
B. Ensure that all servers can acquire an IP by means of Dynamic Host Configuration Protocol (DHCP).
C. Upgrade all Windows VMs to Windows Server 2008 R2 or higher.
D. Sysprep all Windows servers
E. Convert the existing virtual disks to the virtual hard disk (VHD) format.
Answer: A,B,E
Explanation:
The Azure Linux Agent is installed on the Linux VM and is responsible to communicate with the Azure Frabric Controller.
It is a prerequisite that the Virtual Machines can receive ip addresses from DHCP. Azure does not use VHDX, only VHD. We are required to convert VHDX to VHD.
P.S. Easily pass 70-535 Exam with Examcollectionplus Printable Dumps & pdf vce, Try Free: https://www.examcollectionplus.net/vce-70-535/ ( New Questions)