70-475 | Microsoft 70-475 Free Practice Questions 2019

We offers 70 475 exam. "Designing and Implementing Big Data Analytics Solutions", also known as 70-475 exam, is a Microsoft Certification. This set of posts, Passing the 70-475 exam with exam 70 475, will help you answer those questions. The 70 475 exam covers all the knowledge points of the real exam. 100% real 70 475 exam and revised by experts!

Online 70-475 free questions and answers of New Version:

NEW QUESTION 1
You have a Microsoft Azure Data Factory that loads data to an analytics solution. You receive an alert that an error occurred during the last processing of a data stream. You debug the problem and solve an error.
You need to process the data stream that caused the error. What should you do?

  • A. From Azure Cloud Shell, run the az dla job command.
  • B. From Azure Cloud Shell, run the az batch job enable command.
  • C. From PowerShell, run the Resume-AzureRmDataFactoryPipeline cmdlet.
  • D. From PowerShell, run the Set-AzureRmDataFactorySliceStatus cmdlet.

Answer: D

Explanation: ADF operates on data in batches known as slices. Slices are obtained by querying data over a date-time window—for example, a slice may contain data for a specific hour, day, or week.
References:
https://blogs.msdn.microsoft.com/bigdatasupport/2016/08/31/rerunning-many-slices-and-activities-in-azure-data

NEW QUESTION 2
You plan to implement a Microsoft Azure Data Factory pipeline. The pipeline will have custom business logic that requires a custom processing step.
You need to implement the custom processing step by using C#.
Which interface and method should you use? To answer, select the appropriate options in the answer area. NOTE: Each correct selection is worth one point.
70-475 dumps exhibit

    Answer:

    Explanation: References:
    https://github.com/MicrosoftDocs/azure-docs/blob/master/articles/data-factory/v1/data-factory-use-custom-activ

    NEW QUESTION 3
    You need to implement a solution that meets the data refresh requirement for DB1.
    Which three actions should you perform in sequence? To answer, move the appropriate actions from the list of actions to the answer area and arrange them in the correct order.
    70-475 dumps exhibit

      Answer:

      Explanation: Azure Data Factory can be used to orchestrate the execution of stored procedures. This allows more complex pipelines to be created and extends Azure Data Factory's ability to leverage the computational power of SQL Data Warehouse.
      From scenario:
      Relecloud has a Microsoft SQL Server database named DB1 that stores information about the advertisers. DB1 is hosted on a Microsoft Azure virtual machine.
      Relecloud identifies the following requirements for DB1:
      70-475 dumps exhibit Data generated by the streaming analytics platform must be stored in DB1.
      70-475 dumps exhibit The advertisers in DB1 must be stored in a table named Table1 and must be refreshed nightly.

      Topic 3, Litware, Inc
      Overview
      General Overview
      Litware, Inc. is a company that manufactures personal devices to track physical activity and other health-related data.
      Litware has a health tracking application that sends health-related data horn a user's personal device to Microsoft Azure.
      Physical Locations
      Litware has three development and commercial offices. The offices are located in the Untied States, Luxembourg, and India.
      Litware products are sold worldwide. Litware has commercial representatives in more than 80 countries.
      Existing Environment Environment
      In addition to using desktop computers in all of the offices. Litware recently started using Microsoft Azure resources and services for both development and operations.
      Litware has an Azure Machine Learning Solution.
      Litware Health Tracking Application
      Litware recently extended its platform to provide third-party companies with the ability to upload data from devices to Azure. The data can be aggregated across multiple devices to provide users with a comprehensive view of their global health activity.
      While the upload from each device is small, potentially more than 100 million devices will upload data daily by using an Azure event hub.
      Each health activity has a small amount of data, such as activity type, start date/time, and end date/time. Each activity is limited to a total of 3 KB and includes a customer Identification key.
      In addition to the Litware health tracking application, the users' activities can be reported to Azure by using an open API.
      Machine Learning Experiments
      The developers at Litware perform Machine Learning experiments to recommend an appropriate health activity based on the past three activities of a user.
      The Litware developers train a model to recommend the best activity for a user based on the hour of the day.
      Requirements Planned Changes
      Litware plans to extend the existing dashboard features so that health activities can be compared between the users based on age, gender, and geographic region.
      Business Goals
      Minimize the costs associated with transferring data from the event hub to Azure Storage.
      Technical Requirements
      Litware identities the following technical requirements:
      Data from the devices must be stored from three years in a format that enables the fast processing of data fields and Filtering.
      The third-party companies must be able to use the Litware Machine learning models to generate recommendations to their users by using a third-party application.
      Any changes to the health tracking application must ensure that the Litware developers can run the experiments without interrupting or degrading the performance of the production environment.
      Privacy Requirements
      Activity tracking data must be available to all of the Litware developers for experimentation. The developers must be prevented from accessing the private information of the users.
      Other Technical Requirements
      When the Litware health tracking application asks users how they feel, their responses must be reported to Azure.

      Topic 2, Mix Questions

      NEW QUESTION 4
      You have a Microsoft Azure Stream Analytics job that contains several pipelines.
      The Stream Analytics job is configured to trigger an alert when the sale of products in specific categories exceeds a specified threshold.
      You plan to change the product-to-category mappings next month to meet future business requirements.
      You need to create the new product-to-category mappings to prepare for the planned change. The solution must ensure that the Stream Analytics job only uses the new product-to-category mappings when the
      mappings are ready to be activated.
      Which naming structure should you use for the file that contains the product-to-category mappings?

      • A. Use any date after the day the file becomes active.
      • B. Use any date before the day the categories become active.
      • C. Use the date and hour that the categories are to become active.
      • D. Use the current date and time.

      Answer: C

      NEW QUESTION 5
      You are designing a solution that will use Apache HBase on Microsoft Azure HDInsight.
      You need to design the row keys for the database to ensure that client traffic is directed over all of the nodes in the cluster.
      What are two possible techniques that you can use? Each correct answer presents a complete solution. NOTE: Each correct selection is worth one point.

      • A. padding
      • B. trimming
      • C. hashing
      • D. salting

      Answer: CD

      Explanation: There are two strategies that you can use to avoid hotspotting:
      * Hashing keys
      To spread write and insert activity across the cluster, you can randomize sequentially generated keys by hashing the keys, inverting the byte order. Note that these strategies come with trade-offs. Hashing keys, for example, makes table scans for key subranges inefficient, since the subrange is spread across the cluster.
      * Salting keys
      Instead of hashing the key, you can salt the key by prepending a few bytes of the hash of the key to the actual key.
      Note. Salted Apache HBase tables with pre-split is a proven effective HBase solution to provide uniform workload distribution across RegionServers and prevent hot spots during bulk writes. In this design, a row key is made with a logical key plus salt at the beginning. One way of generating salt is by calculating n (number of regions) modulo on the hash code of the logical row key (date, etc).
      Reference:
      https://blog.cloudera.com/blog/2015/06/how-to-scan-salted-apache-hbase-tables-with-region-specific-key-range
      http://maprdocs.mapr.com/51/MapR-DB/designing_row_keys_for_mapr_db_binary_tables.html

      NEW QUESTION 6
      You need to implement a security solution for Microsoft Azure SQL database. The solution must meet the following requirements:
      70-475 dumps exhibit Ensure that users can see the data from their respective department only.
      70-475 dumps exhibit Prevent administrators from viewing the data.
      Which feature should you use for each requirement? To answer, drag the appropriate features to the correct requirements. Each feature may be used once, more than once, or not at all. You may need to drag the split bar between panes or scroll to view content.
      NOTE: Each correct selection is worth one point.
      70-475 dumps exhibit

        Answer:

        Explanation: 70-475 dumps exhibit

        NEW QUESTION 7
        You have a Microsoft Azure data factory named ADF1 that contains a pipeline named Pipeline1. You plan to automate updates to Pipeline1.
        You need to build the URL that must be called to update the pipeline from the REST API.
        How should you complete the URL? To answer, drag the appropriate URL elements to the correct locations. Each URL element may be used once, more than once, or not at all. You may need to drag the split bar between panes or scroll to view content.
        NOTE: Each correct selection is worth one point.
        70-475 dumps exhibit

          Answer:

          Explanation: 70-475 dumps exhibit

          NEW QUESTION 8
          You are planning a solution that will have multiple data files stored in Microsoft Azure Blob storage every hour. Data processing will occur once a day at midnight only.
          You create an Azure data factory that has blob storage as the input source and an Azure HD Insight activity that uses the input to create an output Hive table.
          You need to identify a data slicing strategy for the data factory.
          What should you identify? To answer, select the appropriate options in the answer area. NOTE: Each correct selection is worth one point.
          70-475 dumps exhibit

            Answer:

            Explanation: 70-475 dumps exhibit

            NEW QUESTION 9
            A company named Fabrikam, Inc. has a Microsoft Azure web app. Billions of users visit the app daily.
            The web app logs all user activity by using text files in Azure Blob storage. Each day, approximately 200 GB of text files are created.
            Fabrikam uses the log files from an Apache Hadoop cluster on Azure DHlnsight.
            You need to recommend a solution to optimize the storage of the log files for later Hive use.
            What is the best property to recommend adding to the Hive table definition to achieve the goal? More than one answer choice may achieve the goal. Select the BEST answer.

            • A. STORED AS RCFILE
            • B. STORED AS GZIP
            • C. STORED AS ORC
            • D. STORED AS TEXTFILE

            Answer: C

            Explanation: The Optimized Row Columnar (ORC) file format provides a highly efficient way to store Hive data. It was designed to overcome limitations of the other Hive file formats. Using ORC files improves performance when Hive is reading, writing, and processing data.
            Compared with RCFile format, for example, ORC file format has many advantages such as:
            70-475 dumps exhibit a single file as the output of each task, which reduces the NameNode's load
            70-475 dumps exhibit Hive type support including datetime, decimal, and the complex types (struct, list, map, and union)
            70-475 dumps exhibit light-weight indexes stored within the file
            70-475 dumps exhibit skip row groups that don't pass predicate filtering
            70-475 dumps exhibit seek to a given row
            70-475 dumps exhibit block-mode compression based on data type
            70-475 dumps exhibit run-length encoding for integer columns
            70-475 dumps exhibit dictionary encoding for string columns
            70-475 dumps exhibit concurrent reads of the same file using separate RecordReaders
            70-475 dumps exhibit ability to split files without scanning for markers
            70-475 dumps exhibit bound the amount of memory needed for reading or writing
            70-475 dumps exhibit metadata stored using Protocol Buffers, which allows addition and removal of fields

            NEW QUESTION 10
            You have an Apache Hadoop system that contains 5 TB of data.
            You need to create queries to analyze the data in the system. The solution must ensure that the queries execute as quickly as possible.
            Which language should you use to create the queries?

            • A. Apache Pig
            • B. Java
            • C. Apache Hive
            • D. MapReduce

            Answer: D

            NEW QUESTION 11
            You plan to implement a Microsoft Azure Data Factory pipeline. The pipeline will have custom business logic that requires a custom processing step.
            You need to implement the custom processing step by using C#.
            Which interface and method should you use? To answer, select the appropriate options in the answer area. NOTE: Each correct selection is worth one point.
            70-475 dumps exhibit

              Answer:

              Explanation: References:
              https://github.com/MicrosoftDocs/azure-docs/blob/master/articles/data-factory/v1/data-factory-use-custom-activ

              NEW QUESTION 12
              A company named Fabrikam, Inc. plans to monitor financial markets and social networks, and then to correlate global stock movements to social network activity.
              You need to recommend a Microsoft Azure HDInsight cluster solution that meets the following requirements: 70-475 dumps exhibitProvides continuous availability
              70-475 dumps exhibit Can process asynchronous feeds
              What is the best type of cluster to recommend to achieve the goal? More than one answer choice may achieve the goal. Select the BEST answer.

              • A. Apache Hbase
              • B. Apache Hadoop
              • C. Apache Spark
              • D. Apache Storm

              Answer: C

              NEW QUESTION 13
              You have a Microsoft Azure Machine Learning application named App1 that is used by several departments in your organization.
              App 1 connects to an Azure database named DB1. DB1 contains several tables that store sensitive information. You plan to implement a security solution for the tables.
              You need to prevent the users of App1 from viewing the data of users in other departments in the tables. The solution must ensure that the users can see only data of the users in their respective department.
              Which feature should you implement?

              • A. Cell-level encryption
              • B. Row-Level Security (RLS)
              • C. Transparent Data Encryption (TDE)
              • D. Dynamic Data Masking

              Answer: D

              NEW QUESTION 14
              You extend the dashboard of the health tracking application to summarize fields across several users. You need to recommend a file format for the activity data in Azure that meets the technical requirements.
              What is the best recommendation to achieve the goal? More than one answer choice may achieve the goal. Select the BEST answer.

              • A. ORC
              • B. TSV
              • C. CSV
              • D. JSON
              • E. XML

              Answer: E

              NEW QUESTION 15
              You are designing a data-driven data flow in Microsoft Azure Data Factory to copy data from Azure Blob storage to Azure SQL Database.
              You need to create the copy activity.
              How should you complete the JSON code? To answer, drag the appropriate code elements to the correct targets. Each element may be used once, more than once, or not at all. You may need to drag the split bar between panes or scroll to view content
              NOTE: Each correct selection is worth one point.
              70-475 dumps exhibit

                Answer:

                Explanation: 70-475 dumps exhibit

                NEW QUESTION 16
                Users report that when they access data that is more than one year old from a dashboard, the response time is slow.
                You need to resolve the issue that causes the slow response when visualizing older data. What should you do?

                • A. Process the event hub data first, and then process the older data on demand.
                • B. Process the older data on demand first, and then process the event hub data.
                • C. Aggregate the older data by time, and then save the aggregated data to reference data streams.
                • D. Store all of the data from the event hub in a single partition.

                Answer: C

                P.S. Certleader now are offering 100% pass ensure 70-475 dumps! All 70-475 exam questions have been updated with correct answers: https://www.certleader.com/70-475-dumps.html (102 New Questions)