Valid DP-200 Dumps shared by ExamDiscuss.com for Helping Passing DP-200 Exam! ExamDiscuss.com now offer the newest DP-200 exam dumps, the ExamDiscuss.com DP-200 exam questions have been updated and answers have been corrected get the newest ExamDiscuss.com DP-200 dumps with Test Engine here:
You need to develop a pipeline for processing dat a. The pipeline must meet the following requirements. * Scale up and down resources for cost reduction. * Use an in-memory data processing engine to speed up ETL and machine learning operations. * Use streaming capabilities. * Provide the ability to code in SQL, Python, Scala, and R. * Integrate workspace collaboration with Git. What should you use?
Correct Answer: A
Aparch Spark is an open-source, parallel-processing framework that supports in-memory processing to boost the performance of big-data analysis applications. HDInsight is a managed Hadoop service. Use it deploy and manage Hadoop clusters in Azure. For batch processing, you can use Spark, Hive, Hive LLAP, MapReduce. Languages: R, Python, Java, Scala, SQL You can create an HDInsight Spark cluster using an Azure Resource Manager template. The template can be found in GitHub. References: https://docs.microsoft.com/en-us/azure/architecture/data-guide/technology-choices/batch-processing