How to set Spark recovery to Zookeeper

0 votes
Hi. I want to set a recovery system for my Spark application. I have a Zookeeper running for this purpose. Now, I want to know how to set Zookeeper as the recovery setting for the Spark application. Please help
Mar 25, 2019 in Apache Spark by Hanuman
766 views

1 answer to this question.

0 votes

There is a property of Spark which lets you set the recovery mode. To set Zookeeper as the recovery mode, try the below commands:

val sc = new SparkContext(new SparkConf())

./bin/spark-submit <all your existing options> --spark.deploy.recoveryMode=ZOOKEEPER
answered Mar 25, 2019 by Hari

Related Questions In Apache Spark

0 votes
1 answer

How to connect to Zookeeper after setting Spark recovery mode?

You have set Zookeeper as the recovery ...READ MORE

answered Mar 25, 2019 in Apache Spark by Hari
1,403 views
0 votes
1 answer
0 votes
1 answer

How to set cpu cores for spark task?

By default, each task is allocated with ...READ MORE

answered Mar 12, 2019 in Apache Spark by Veer
4,420 views
0 votes
1 answer

How to set stage abort in Spark?

You can do it as follows. Use ...READ MORE

answered Mar 12, 2019 in Apache Spark by veer
1,110 views
+1 vote
2 answers
+1 vote
1 answer

Hadoop Mapreduce word count Program

Firstly you need to understand the concept ...READ MORE

answered Mar 16, 2018 in Data Analytics by nitinrawat895
• 11,380 points
11,035 views
0 votes
1 answer

hadoop.mapred vs hadoop.mapreduce?

org.apache.hadoop.mapred is the Old API  org.apache.hadoop.mapreduce is the ...READ MORE

answered Mar 16, 2018 in Data Analytics by nitinrawat895
• 11,380 points
2,544 views
+2 votes
11 answers

hadoop fs -put command?

Hi, You can create one directory in HDFS ...READ MORE

answered Mar 16, 2018 in Big Data Hadoop by nitinrawat895
• 11,380 points
108,864 views
0 votes
1 answer

How to set executors for static allocation in Spark Yarn?

Open Spark shell and run the following ...READ MORE

answered Mar 28, 2019 in Apache Spark by Raj
1,497 views
0 votes
1 answer

How to set extra JVM options for Spark application?

You cans set extra JVM options that ...READ MORE

answered Mar 28, 2019 in Apache Spark by Raj
3,967 views
webinar REGISTER FOR FREE WEBINAR X
REGISTER NOW
webinar_success Thank you for registering Join Edureka Meetup community for 100+ Free Webinars each month JOIN MEETUP GROUP