70-475 | Microsoft 70-475 Dumps Questions 2019

We provide microsoft 70 475 in two formats. Download PDF & Practice Tests. Pass Microsoft 70-475 Exam quickly & easily. The 70-475 PDF type is available for reading and printing. You can print more and practice many times. With the help of our exam 70 475 product and material, you can easily pass the 70-475 exam.

Free demo questions for Microsoft 70-475 Exam Dumps Below:

NEW QUESTION 1
A Company named Fabrikam, Inc. has a web app. Millions of users visit the app daily.
Fabrikam performs a daily analysis of the previous day’s logs by scheduling the following Hive query.
70-475 dumps exhibit
You need to recommend a solution to gather the log collections from the web app. What should you recommend?

  • A. Generate a single directory that contains multiple files for each da
  • B. Name the file by using the syntax of{date}_{randomsuffix}.txt.
  • C. Generate a directory that is named by using the syntax of "LogDate={date}” and generate a set of files for that day.
  • D. Generate a directory each day that has a single file.
  • E. Generate a single directory that has a single file for each day.

Answer: B

NEW QUESTION 2
You have an application that displays data from a Microsoft Azure SQL database. The database contains credit card numbers.
You need to ensure that the application only displays the last four digits of each credit card number when a credit card number is returned from a query. The solution must NOT require any changes to the data in the database.
What should you use?

  • A. Dynamic Data Masking
  • B. cell-level security
  • C. Transparent Data Encryption (TDE)
  • D. row-level security

Answer: A

NEW QUESTION 3
You plan to deploy a Microsoft Azure Data Factory pipeline to run an end-to-end data processing workflow. You need to recommend winch Azure Data Factory features must be used to meet the Following requirements: Track the run status of the historical activity.
Enable alerts and notifications on events and metrics.
Monitor the creation, updating, and deletion of Azure resources.
Which features should you recommend? To answer, drag the appropriate features to the correct requirements. Each feature may be used once, more than once, or not at all. You may need to drag the split bar between panes or scroll to view content.
NOTE: Each correct selection is worth one point.
70-475 dumps exhibit

    Answer:

    Explanation: Box 1: Azure Hdinsight logs Logs contain historical activities. Box 2: Azure Data Factory alerts Box 3: Azure Data Factory events

    NEW QUESTION 4
    You have an Apache Hive cluster in Microsoft Azure HDInsight. The cluster contains 10 million data files. You plan to archive the data.
    The data will be analyzed monthly.
    You need to recommend a solution to move and store the data. The solution must minimize how long it takes to move the data and must minimize costs.
    Which two services should you include in the recommendation? Each correct answer presents part of the solution.
    NOTE: Each correct selection is worth one point.

    • A. Azure Queue storage
    • B. Microsoft SQL Server Integration Services (SSIS)
    • C. Azure Table Storage
    • D. Azure Data Lake
    • E. Azure Data Factory

    Answer: DE

    Explanation: D: To analyze data in HDInsight cluster, you can store the data either in Azure Storage, Azure Data Lake Storage Gen 1/Azure Data Lake Storage Gen 2, or both. Both storage options enable you to safely delete HDInsight clusters that are used for computation without losing user data.
    E: The Spark activity in a Data Factory pipeline executes a Spark program on your own or on-demand HDInsight cluster. It handles data transformation and the supported transformation activities.
    References:
    https://docs.microsoft.com/en-us/azure/hdinsight/hdinsight-hadoop-use-data-lake-store https://docs.microsoft.com/en-us/azure/data-factory/transform-data-using-spark

    NEW QUESTION 5
    You are designing an application that will perform real-time processing by using Microsoft Azure Stream Analytics.
    You need to identify the valid outputs of a Stream Analytics job.
    What are three possible outputs that you can use? Each correct answer presents a complete solution.
    NOTE: Each correct selection is worth one point.

    • A. Microsoft Power BI
    • B. Azure SQL Database
    • C. a Hive table in Azure HDInsight
    • D. Azure Blob storage
    • E. Azure Redis Cache

    Answer: ABD

    Explanation: https://docs.microsoft.com/en-us/azure/stream-analytics/stream-analytics-define-outputs

    NEW QUESTION 6
    You are building an Azure Analysis Services cube.
    The source data for the cube is located on premises in a Microsoft SQL Server database. You need to ensure that the Azure Analysis Services service can access the source data. What should you deploy to your Azure subscription?

    • A. a site-to-site VPN
    • B. Azure Data Factory
    • C. a network gateway in Azure
    • D. a data gateway in Azure

    Answer: D

    Explanation: Connecting to on-premises data sources from and Azure AS server require an On-premises gateway.
    70-475 dumps exhibit
    References:
    https://azure.microsoft.com/en-in/blog/on-premises-data-gateway-support-for-azure-analysis-services/

    NEW QUESTION 7
    You have a web application that generates several terabytes (TB) of financial documents each day. The application processes the documents in batches.
    You need to store the documents in Microsoft Azure. The solution must ensure that a user can restore the previous version of a document.
    Which type of storage should you use for the documents?

    • A. Azure Cosmos DB
    • B. Azure File Storage
    • C. Azure Data Lake
    • D. Azure Blob storage

    Answer: A

    NEW QUESTION 8
    Note: This question is part of a series of questions that present the same scenario. Each question in the series contains a unique solution that might meet the states goals. Some question sets might have more than one correct solution, while the others might not have a correct solution.
    After you answer a question in this section, you will NOT be able to return to it. As a result, these questions will not appear in the review screen.
    You plan to implement a new data warehouse.
    You have the following information regarding the data warehouse:
    70-475 dumps exhibit The first data files for the data warehouse will be available in a few days.
    70-475 dumps exhibit Most queries that will be executed against the data warehouse are ad-hoc.
    70-475 dumps exhibit The schemas of data files that will be loaded to the data warehouse change often.
    70-475 dumps exhibit One month after the planned implementation, the data warehouse will contain 15 TB of data. You need to recommend a database solution to support the planned implementation.
    Solution: You recommend an Apache Hadoop system. Does this meet the goal?

    • A. Yes
    • B. No

    Answer: A

    NEW QUESTION 9
    Your company has a Microsoft Azure environment that contains an Azure HDInsight Hadoop cluster and an Azure SQL data warehouse. The Hadoop cluster contains text files that are formatted by using UTF-8 character encoding.
    You need to implement a solution to ingest the data to the SQL data warehouse from the Hadoop cluster. The solution must provide optimal read performance for the data after ingestion.
    Which three actions should you perform in sequence? To answer, move the appropriate actions from the list of actions to the answer area and arrange them in the correct order.
    70-475 dumps exhibit

      Answer:

      Explanation: SQL Data Warehouse supports loading data from HDInsight via PolyBase. The process is the same as loading data from Azure Blob Storage - using PolyBase to connect to HDInsight to load data.
      Use PolyBase and T-SQL Summary of loading process: Recommendations
      Create statistics on newly loaded data. Azure SQL Data Warehouse does not yet support auto create or auto update statistics. In order to get the best performance from your queries, it's important to create statistics on all columns of all tables after the first load or any substantial changes occur in the data.

      NEW QUESTION 10
      You are using a Microsoft Azure Data Factory pipeline to copy data to an Azure SQL database. You need to prevent the insertion of duplicate data for a given dataset slice.
      Which two actions should you perform? Each correct answer presents part of the solution. NOTE: Each correct selection is worth one point.

      • A. Set the External property to true.
      • B. Add a column named SliceIdentifierColumnName to the output dataset.
      • C. Set the SqlWriterCleanupScript property to true.
      • D. Remove the duplicates in post-processing.
      • E. Manually delete the duplicate data before running the pipeline activity.

      Answer: BC

      NEW QUESTION 11
      Your company builds hardware devices that contain sensors. You need to recommend a solution to process the sensor data and. What should you include in the recommendation?

      • A. Microsoft Azure Event Hubs
      • B. API apps in Microsoft Azure App Service
      • C. Microsoft Azure Notification Hubs
      • D. Microsoft Azure IoT Hub

      Answer: A

      NEW QUESTION 12
      You have a Microsoft Azure Data Factory pipeline.
      You discover that the pipeline fails to execute because data is missing. You need to rerun the failure in the pipeline.
      Which cmdlet should you use?

      • A. Set-AzureRmAutomationJob
      • B. Set-AzureRmDataFactorySliceStatus
      • C. Resume-AzureRmDataFactoryPipeline
      • D. Resume-AzureRmAutomationJob

      Answer: B

      Explanation: Use some PowerShell to inspect the ADF activity for the missing file error. Then simply set the dataset slice to either skipped or ready using the cmdlet to override the status.
      For example:
      Set-AzureRmDataFactorySliceStatus `
      -ResourceGroupName $ResourceGroup `
      -DataFactoryName $ADFName.DataFactoryName `
      -DatasetName $Dataset.OutputDatasets `
      -StartDateTime $Dataset.WindowStart `
      -EndDateTime $Dataset.WindowEnd `
      -Status "Ready" `
      -UpdateType "Individual" References:
      https://stackoverflow.com/questions/42723269/azure-data-factory-pipelines-are-failing-when-no-files-available-

      NEW QUESTION 13
      You plan to deploy Microsoft Azure HDInsight clusters for business analytics and data pipelines. The clusters must meet the following requirements:
      70-475 dumps exhibit Business users must use a language that is similar to SQL.
      70-475 dumps exhibit The authoring of data pipelines must occur in a dataflow language. You need to identify which language must be used for each requirement.
      Which languages should you identify? To answer, drag the appropriate languages to the correct requirements. Each language may be used once, more than once, or not at all. You may need to drag the split bar between panes or scroll to view content.
      70-475 dumps exhibit

        Answer:

        Explanation: 70-475 dumps exhibit

        NEW QUESTION 14
        You have data generated by sensors. The data is sent to Microsoft Azure Event Hubs.
        You need to have an aggregated view of the data in near real-time by using five minute tumbling windows to identity short-term trends. You must also have hourly and a daily aggregated views of the data.
        Which technology should you use for each task? To answer, drag the appropriate technologies to the correct tasks. Each technology may be used once, more than once, or not at all. You may need to drag the split bar between panes or scroll to view content.
        NOTE: Each correct selection is worth one point.
        70-475 dumps exhibit

          Answer:

          Explanation: Box 1: Azure HDInsight MapReduce
          Azure Event Hubs allows you to process massive amounts of data from websites, apps, and devices. The Event Hubs spout makes it easy to use Apache Storm on HDInsight to analyze this data in real time.
          Box 2: Azure Event Hub
          Box 3: Azure Stream Analytics
          Stream Analytics is a new service that enables near real time complex event processing over streaming data. Combining Stream Analytics with Azure Event Hubs enables near real time processing of millions of events per second. This enables you to do things such as augment stream data with reference data and output to storage (or even output to another Azure Event Hub for additional processing).

          NEW QUESTION 15
          You have an Apache Spark cluster on Microsoft Azure HDInsight for all analytics workloads.
          You plan to build a Spark streaming application that processes events ingested by using Azure Event Hubs. You need to implement checkpointing in the Spark streaming application for high availability of the event
          data.
          In which order should you perform the actions? To answer, move all actions from the list of actions to the answer area and arrange them in the correct order.
          70-475 dumps exhibit

            Answer:

            Explanation: 70-475 dumps exhibit

            NEW QUESTION 16
            You plan to create a Microsoft Azure Data Factory pipeline that will connect to an Azure HDInsight cluster that uses Apache Spark.
            You need to recommend which file format must be used by the pipeline. The solution must meet the following requirements:
            70-475 dumps exhibit Store data in the columnar format
            70-475 dumps exhibit Support compression
            Which file format should you recommend?

            • A. XML
            • B. AVRO
            • C. text
            • D. Parquet

            Answer: D

            Explanation: Apache Parquet is a columnar storage format available to any project in the Hadoop ecosystem, regardless of the choice of data processing framework, data model or programming language.
            Apache Parquet supports compression.

            Thanks for reading the newest 70-475 exam dumps! We recommend you to try the PREMIUM Surepassexam 70-475 dumps in VCE and PDF here: https://www.surepassexam.com/70-475-exam-dumps.html (102 Q&As Dumps)