Data Science - Big Data Assessment
Exam Questions with 100% Verified
Answers
when building a standalone application, you need to create the SparkContext. To
do this in Scala, you would include which of the following within the main
method?
val conf=new SparkConf().setAppName("AuctionsApp")
val sc= new SparkContext(conf)
val sc = SparkContext().setAppName("AuctionsApp"_
val conf= SparkConf(sc)
val sc=new SparkContext()
val conf= new SparkConf().setAppName("AuctionsApp") - ✔ ✔ val
conf=new SparkConf().setAppName("AuctionsApp")
val sc= new SparkContext(conf)
,which of these are responsible to output key-value pairs from the Reducer phase
to output files?
reducer Writer
RecordReader
RecordWriter
none of these - ✔ ✔ RecordWriter
The HDFS block size configuration is 128MB in your hadoop cluster. The HDFS
directory contains 50 small files each of 200 MB in size. How many map tasks
will be created when the inputformat for your job is TextInputFormat?
100
128
50
200 - ✔ ✔ 100
which of the following application types can Spark run in addition to batch-
processing jobs?
all
graph processing
, machine learning
stream processin - ✔ ✔ graph processing
in what year was apache spark made an open-source technology? - ✔ ✔ 2010
what kind of data can be handled by spark?
semi-structured
unstructured
structured
all - ✔ ✔ all
spark is 100x faster than MapReduce due to development in scala - ✔ ✔ false
the following are characteristics shared by hadoop and spark, except ____
- both use open source api's to link between different tools
- both are data processing platforms
-both have their own file system
-both are cluster computing environments - ✔ ✔ both have their own file system
spark can store its data in ___
- all
Exam Questions with 100% Verified
Answers
when building a standalone application, you need to create the SparkContext. To
do this in Scala, you would include which of the following within the main
method?
val conf=new SparkConf().setAppName("AuctionsApp")
val sc= new SparkContext(conf)
val sc = SparkContext().setAppName("AuctionsApp"_
val conf= SparkConf(sc)
val sc=new SparkContext()
val conf= new SparkConf().setAppName("AuctionsApp") - ✔ ✔ val
conf=new SparkConf().setAppName("AuctionsApp")
val sc= new SparkContext(conf)
,which of these are responsible to output key-value pairs from the Reducer phase
to output files?
reducer Writer
RecordReader
RecordWriter
none of these - ✔ ✔ RecordWriter
The HDFS block size configuration is 128MB in your hadoop cluster. The HDFS
directory contains 50 small files each of 200 MB in size. How many map tasks
will be created when the inputformat for your job is TextInputFormat?
100
128
50
200 - ✔ ✔ 100
which of the following application types can Spark run in addition to batch-
processing jobs?
all
graph processing
, machine learning
stream processin - ✔ ✔ graph processing
in what year was apache spark made an open-source technology? - ✔ ✔ 2010
what kind of data can be handled by spark?
semi-structured
unstructured
structured
all - ✔ ✔ all
spark is 100x faster than MapReduce due to development in scala - ✔ ✔ false
the following are characteristics shared by hadoop and spark, except ____
- both use open source api's to link between different tools
- both are data processing platforms
-both have their own file system
-both are cluster computing environments - ✔ ✔ both have their own file system
spark can store its data in ___
- all