Hello R-help list,
I am a new list member. My first question: I was trying out sparklyr (in R ver 3.3.2) on my Red Hat Linux workstation, following the instructions at spark.rstudio.com as to how to download and use a local copy of Spark. The Spark download appears to work. However, when I try to issue the spark_connect, to get started, I get the error msgs that you see below.
I cannot find any guidance as to how to fix this. Quite frustrating. Can somebody give me a bit of help? Does something need to be added to my PATH env var in my .mycshrc file, for example? Is there a closed port problem? Has anybody run into this type of error msg? Do I need to do something additional to start up the local copy of Spark that is not mentioned in the RStudio online documentation?
- Ron
%%%%%%%%%%%%%%%%%%%%
Here is the spark_install (apparently successful) and then the error msg on the spark_connect():
Error in force(code) :
Failed while connecting to sparklyr to port (8880) for sessionid (3689): Gateway in port (8880) did not respond.
Path: /home/rtaylor/.cache/spark/spark-1.6.2-bin-hadoop2.6/bin/spark-submit
Parameters: --class, sparklyr.Backend, --jars, '/usr/lib64/R/library/sparklyr/java/spark-csv_2.11-1.3.0.jar','/usr/lib64/R/library/sparklyr/java/commons-csv-1.1.jar','/usr/lib64/R/library/sparklyr/java/univocity-parsers-1.5.1.jar', '/usr/lib64/R/library/sparklyr/java/sparklyr-1.6-2.10.jar', 8880, 3689
---- Output Log ----
/home/rtaylor/.cache/spark/spark-1.6.2-bin-hadoop2.6/bin/spark-class: line 86: /usr/local/bin/bin/java: No such file or directory
---- Error Log ----
%%%%%%%%%%%%%%%%%%
And here is the entire screen output of my R session, from the R invocation on:
sidney115% R
R version 3.3.2 (2016-10-31) -- "Sincere Pumpkin Patch"
Copyright (C) 2016 The R Foundation for Statistical Computing
Platform: x86_64-redhat-linux-gnu (64-bit)
Error in force(code) :
Failed while connecting to sparklyr to port (8880) for sessionid (3689): Gateway in port (8880) did not respond.
Path: /home/rtaylor/.cache/spark/spark-1.6.2-bin-hadoop2.6/bin/spark-submit
Parameters: --class, sparklyr.Backend, --jars, '/usr/lib64/R/library/sparklyr/java/spark-csv_2.11-1.3.0.jar','/usr/lib64/R/library/sparklyr/java/commons-csv-1.1.jar','/usr/lib64/R/library/sparklyr/java/univocity-parsers-1.5.1.jar', '/usr/lib64/R/library/sparklyr/java/sparklyr-1.6-2.10.jar', 8880, 3689
---- Output Log ----
/home/rtaylor/.cache/spark/spark-1.6.2-bin-hadoop2.6/bin/spark-class: line 86: /usr/local/bin/bin/java: No such file or directory
---- Error Log ----
%%%%%%%%%%%%%%%%%%
Ronald C. Taylor, Ph.D.
Computational Biology & Bioinformatics Group
Pacific Northwest National Laboratory (U.S. Dept of Energy/Battelle)
Richland, WA 99352
phone: (509) 372-6568, email: ronald.taylor at pnnl.gov
web page: http://www.pnnl.gov/science/staff/staff_info.asp?staff_num=7048
On Feb 1, 2017, at 3:23 PM, Taylor, Ronald C <Ronald.Taylor at pnnl.gov> wrote:
Hello R-help list,
I am a new list member. My first question: I was trying out sparklyr (in R ver 3.3.2) on my Red Hat Linux workstation, following the instructions at spark.rstudio.com as to how to download and use a local copy of Spark. The Spark download appears to work. However, when I try to issue the spark_connect, to get started, I get the error msgs that you see below.
I cannot find any guidance as to how to fix this. Quite frustrating. Can somebody give me a bit of help? Does something need to be added to my PATH env var in my .mycshrc file, for example? Is there a closed port problem? Has anybody run into this type of error msg? Do I need to do something additional to start up the local copy of Spark that is not mentioned in the RStudio online documentation?
- Ron
%%%%%%%%%%%%%%%%%%%%
Here is the spark_install (apparently successful) and then the error msg on the spark_connect():
Error in force(code) :
Failed while connecting to sparklyr to port (8880) for sessionid (3689): Gateway in port (8880) did not respond.
Path: /home/rtaylor/.cache/spark/spark-1.6.2-bin-hadoop2.6/bin/spark-submit
Parameters: --class, sparklyr.Backend, --jars, '/usr/lib64/R/library/sparklyr/java/spark-csv_2.11-1.3.0.jar','/usr/lib64/R/library/sparklyr/java/commons-csv-1.1.jar','/usr/lib64/R/library/sparklyr/java/univocity-parsers-1.5.1.jar', '/usr/lib64/R/library/sparklyr/java/sparklyr-1.6-2.10.jar', 8880, 3689
---- Output Log ----
/home/rtaylor/.cache/spark/spark-1.6.2-bin-hadoop2.6/bin/spark-class: line 86: /usr/local/bin/bin/java: No such file or directory
So this makes me wonder if you do not have a proper installation of java for one of those other packages.
David,
You were right. I was using Java 1.6 instead of Java 1.7 or later. Mea culpa. I am now up and running, and looking to do many things with R and Spark. Thank you.
Ron
Ronald C. Taylor, Ph.D.
Computational Biology & Bioinformatics Group
Pacific Northwest National Laboratory (U.S. Dept of Energy/Battelle)
Richland, WA 99352
phone: (509) 372-6568, email: ronald.taylor at pnnl.gov
web page: ?http://www.pnnl.gov/science/staff/staff_info.asp?staff_num=7048
-----Original Message-----
From: David Winsemius [mailto:dwinsemius at comcast.net]
Sent: Wednesday, February 01, 2017 4:40 PM
To: Taylor, Ronald C
Cc: r-help at r-project.org; ronald.taylor24 (ronald.taylor24 at gmail.com)
Subject: Re: [R] need help in trying out sparklyr - spark_connect will not work on local copy of Spark
On Feb 1, 2017, at 3:23 PM, Taylor, Ronald C <Ronald.Taylor at pnnl.gov> wrote:
Hello R-help list,
I am a new list member. My first question: I was trying out sparklyr (in R ver 3.3.2) on my Red Hat Linux workstation, following the instructions at spark.rstudio.com as to how to download and use a local copy of Spark. The Spark download appears to work. However, when I try to issue the spark_connect, to get started, I get the error msgs that you see below.
I cannot find any guidance as to how to fix this. Quite frustrating. Can somebody give me a bit of help? Does something need to be added to my PATH env var in my .mycshrc file, for example? Is there a closed port problem? Has anybody run into this type of error msg? Do I need to do something additional to start up the local copy of Spark that is not mentioned in the RStudio online documentation?
- Ron
%%%%%%%%%%%%%%%%%%%%
Here is the spark_install (apparently successful) and then the error msg on the spark_connect():
Error in force(code) :
Failed while connecting to sparklyr to port (8880) for sessionid (3689): Gateway in port (8880) did not respond.
Path: /home/rtaylor/.cache/spark/spark-1.6.2-bin-hadoop2.6/bin/spark-submit
Parameters: --class, sparklyr.Backend, --jars, '/usr/lib64/R/library/sparklyr/java/spark-csv_2.11-1.3.0.jar','/usr/lib64/R/library/sparklyr/java/commons-csv-1.1.jar','/usr/lib64/R/library/sparklyr/java/univocity-parsers-1.5.1.jar', '/usr/lib64/R/library/sparklyr/java/sparklyr-1.6-2.10.jar', 8880, 3689
---- Output Log ----
/home/rtaylor/.cache/spark/spark-1.6.2-bin-hadoop2.6/bin/spark-class: line 86: /usr/local/bin/bin/java: No such file or directory