Hadoop error when using spark-submit - hadoop

I am trying to spark-submit using Amazon ec2 with the following:
spark-submit --packages org.apache.hadoop:hadoop-aws:2.7.1 --master spark://amazonaws.com SimpleApp.py
and I end up with the following error. It seems to be that it is looking for hadoop. My ec2 cluster was created using spark-ec2 command.
Ivy Default Cache set to: /home/adas/.ivy2/cache
The jars for the packages stored in: /home/adas/.ivy2/jars
:: loading settings :: url = jar:file:/home/adas/spark/spark-2.1.0-bin-hadoop2.7/jars/ivy-2.4.0.jar!/org/apache/ivy/core/settings/ivysettings.xml
org.apache.hadoop#hadoop-aws added as a dependency
:: resolving dependencies :: org.apache.spark#spark-submit-parent;1.0
confs: [default]
:: resolution report :: resolve 66439ms :: artifacts dl 0ms
:: modules in use:
---------------------------------------------------------------------
| | modules || artifacts |
| conf | number| search|dwnlded|evicted|| number|dwnlded|
---------------------------------------------------------------------
| default | 1 | 0 | 0 | 0 || 0 | 0 |
---------------------------------------------------------------------
:: problems summary ::
:::: WARNINGS
module not found: org.apache.hadoop#hadoop-aws;2.7.1
==== local-m2-cache: tried
file:/home/adas/.m2/repository/org/apache/hadoop/hadoop-aws/2.7.1/hadoop-aws-2.7.1.pom
-- artifact org.apache.hadoop#hadoop-aws;2.7.1!hadoop-aws.jar:
file:/home/adas/.m2/repository/org/apache/hadoop/hadoop-aws/2.7.1/hadoop-aws-2.7.1.jar
==== local-ivy-cache: tried
/home/adas/.ivy2/local/org.apache.hadoop/hadoop-aws/2.7.1/ivys/ivy.xml
-- artifact org.apache.hadoop#hadoop-aws;2.7.1!hadoop-aws.jar:
/home/adas/.ivy2/local/org.apache.hadoop/hadoop-aws/2.7.1/jars/hadoop-aws.jar
==== central: tried
https://repo1.maven.org/maven2/org/apache/hadoop/hadoop-aws/2.7.1/hadoop-aws-2.7.1.pom
-- artifact org.apache.hadoop#hadoop-aws;2.7.1!hadoop-aws.jar:
https://repo1.maven.org/maven2/org/apache/hadoop/hadoop-aws/2.7.1/hadoop-aws-2.7.1.jar
==== spark-packages: tried
http://dl.bintray.com/spark-packages/maven/org/apache/hadoop/hadoop-aws/2.7.1/hadoop-aws-2.7.1.pom
-- artifact org.apache.hadoop#hadoop-aws;2.7.1!hadoop-aws.jar:
http://dl.bintray.com/spark-packages/maven/org/apache/hadoop/hadoop-aws/2.7.1/hadoop-aws-2.7.1.jar
::::::::::::::::::::::::::::::::::::::::::::::
:: UNRESOLVED DEPENDENCIES ::
::::::::::::::::::::::::::::::::::::::::::::::
:: org.apache.hadoop#hadoop-aws;2.7.1: not found
::::::::::::::::::::::::::::::::::::::::::::::
:::: ERRORS
Server access error at url https://repo1.maven.org/maven2/org/apache/hadoop/hadoop-aws/2.7.1/hadoop-aws-2.7.1.pom (java.net.NoRouteToHostException: No route to host (Host unreachable))
Server access error at url https://repo1.maven.org/maven2/org/apache/hadoop/hadoop-aws/2.7.1/hadoop-aws-2.7.1.jar (java.net.NoRouteToHostException: No route to host (Host unreachable))
Server access error at url http://dl.bintray.com/spark-packages/maven/org/apache/hadoop/hadoop-aws/2.7.1/hadoop-aws-2.7.1.pom (java.net.NoRouteToHostException: No route to host (Host unreachable))
Server access error at url http://dl.bintray.com/spark-packages/maven/org/apache/hadoop/hadoop-aws/2.7.1/hadoop-aws-2.7.1.jar (java.net.NoRouteToHostException: No route to host (Host unreachable))
:: USE VERBOSE OR DEBUG MESSAGE LEVEL FOR MORE DETAILS
Exception in thread "main" java.lang.RuntimeException: [unresolved dependency: org.apache.hadoop#hadoop-aws;2.7.1: not found]
at org.apache.spark.deploy.SparkSubmitUtils$.resolveMavenCoordinates(SparkSubmit.scala:1078)
at org.apache.spark.deploy.SparkSubmit$.prepareSubmitEnvironment(SparkSubmit.scala:296)
at org.apache.spark.deploy.SparkSubmit$.submit(SparkSubmit.scala:160)
at org.apache.spark.deploy.SparkSubmit$.main(SparkSubmit.scala:126)
at org.apache.spark.deploy.SparkSubmit.main(SparkSubmit.scala)

You are submitting the job with --packages org.apache.hadoop:hadoop-aws:2.7.1 option and job is attempting to resolve the dependencies by downloading the packages from public maven repo. However, this error indicates it's unable to reach the maven repo.
Server access error at url https://repo1.maven.org/maven2/org/apache/hadoop/hadoop-aws/2.7.1/hadoop-aws-2.7.1.pom (java.net.NoRouteToHostException: No route to host (Host unreachable))
You might want to check if the spark master has access to internet.

Sorry everyone, it was just some local proxy issues.

Related

503 from server: Service Unavailable: Back-end server is at capacity

I'm trying to build a jib based container with gradle.
I'm getting a 503 error when it tries to GET an xml file from jfrog.org.
+ eval set -- '"-Xmx64m"' '"-Xms64m"' '"-Xdock:name=Gradle"' '"-Xdock:icon=/Users/username/VSCode/globality-karate-testing/media/gradle.icns"' '"-Dorg.gradle.appname=gradlew"' -classpath '"/Users/username/VSCode/globality-karate-testing/gradle/wrapper/gradle-wrapper.jar"' org.gradle.wrapper.GradleWrapperMain ''\''jibDockerBuild'\'' \
'
++ set -- -Xmx64m -Xms64m -Xdock:name=Gradle -Xdock:icon=/Users/username/VSCode/globality-karate-testing/media/gradle.icns -Dorg.gradle.appname=gradlew -classpath /Users/username/VSCode/globality-karate-testing/gradle/wrapper/gradle-wrapper.jar org.gradle.wrapper.GradleWrapperMain jibDockerBuild
+ exec /Users/username/.sdkman/candidates/java/current/bin/java -Xmx64m -Xms64m -Xdock:name=Gradle -Xdock:icon=/Users/username/VSCode/globality-karate-testing/media/gradle.icns -Dorg.gradle.appname=gradlew -classpath /Users/username/VSCode/globality-karate-testing/gradle/wrapper/gradle-wrapper.jar org.gradle.wrapper.GradleWrapperMain jibDockerBuild
FAILURE: Build failed with an exception.
* What went wrong:
A problem occurred configuring root project 'karate-base'.
> Could not resolve all artifacts for configuration ':classpath'.
> Could not resolve org.jfrog.buildinfo:build-info-extractor-gradle:4.+.
Required by:
project :
> Failed to list versions for org.jfrog.buildinfo:build-info-extractor-gradle.
> Unable to load Maven meta-data from https://repo.jfrog.org/artifactory/gradle/org/jfrog/buildinfo/build-info-extractor-gradle/maven-metadata.xml.
> Could not get resource 'https://repo.jfrog.org/artifactory/gradle/org/jfrog/buildinfo/build-info-extractor-gradle/maven-metadata.xml'.
> Could not GET 'https://repo.jfrog.org/artifactory/gradle/org/jfrog/buildinfo/build-info-extractor-gradle/maven-metadata.xml'.
Received status code 503 from server: Service Unavailable: Back-end server is at capacity
This appears to be a problem with the jfrog.org server.
It only seems to happen if this is the first time someone has built the project.
Another engineer was able to replicate my error, who had not built this prior.
An engineer who had built the project earlier does not have problems. Seems to be an initialization file.
Heard back from Jfrog support.
The repo https://repo.jfrog.org/artifactory/ has been deprecated.
Replace it with: https://repo1.maven.org/maven2

Bamboo-artifactory plugin dosen't work at elactic agent

I have a problem with bamboo-artifactory-plugin at bamboo elastic-agent. Build fail with error:
Could not execute task no Plugin with key 'org.jfrog.bamboo.bamboo-artifactory-plugin:maven3Task' is installed.
Bamboo agents on hardware run same build successfully.
I use Bamboo version 7.2, bamboo-artifactory-plugin version 3.2.1 and JFrog Artifactory version 6.18.1
As it looks like, the plugin is not installed on your agent. This may occur due to many reasons. Please try to reinstall the plugin and check the agent's logs.
For example, one possible reason is the lack of memory in the agent. If you see errors such as:
INFO | jvm 1 | 2020/11/10 11:12:36 | 2020-11-10 11:12:36,120 ERROR [remoteEventRebroadcasterMessageListenerContainer-1] [ScanningPluginLoader] Unable to deploy plugin 'org.jfrog.bamboo.bamboo-artifactory-plugin' from 'Unit: /Users/yahavi/bamboo-agent-home/plugins/user-installed/plugin_8527020839401694412_bamboo-artifactory-plugin-3.1.0.jar (1604999532000)'.
INFO | jvm 1 | 2020/11/10 11:12:36 | 2020-11-10 11:12:36,121 ERROR [remoteEventRebroadcasterMessageListenerContainer-1] [ScanningPluginLoader] Because of the following exception:
INFO | jvm 1 | 2020/11/10 11:12:36 | java.lang.OutOfMemoryError: GC overhead limit exceeded
That means your agent has insufficient heap size. I'd recommend on the resolution in this guide: https://confluence.atlassian.com/bamkb/outofmemory-errors-in-bamboo-server-or-in-remote-agent-307038120.html
Edit:
For Elastic agents, you can use this guide:
https://confluence.atlassian.com/bamkb/outofmemory-errors-in-bamboo-elastic-agent-826899144.html

RedHat Fuse 7.5 - Can't deploy feature

Using RedHat Fuse 7.5 for Apache Karaf platform on Windows 10. I'm defining the feature shown below:
...
<feature name="customers-data" version="${project.version}">
<feature>transaction</feature>
<feature>jndi</feature>
<feature>pax-jdbc-config</feature>
<feature>pax-jdbc-oracle</feature>
<feature>pax-jdbc-pool-dbcp2</feature>
<feature>jdbc</feature>
<feature dependency="true">aries-blueprint</feature>
<feature>jpa</feature>
<feature>hibernate</feature>
<bundle>...</bundle>
</feature>
...
Adding the the new repo works as expected but trying to install the feature:
karaf#root()> feature:install customers-data
raises the following exception:
2019-12-12 17:43:02,385 | WARN | ownloader-5-thread-7 | o.o.p.u.m.i.AetherBasedResolver | 5 - org.ops4j.pax.url.mvn - 2.6.1 | Error resolving artifact
org.apache.aries.jpa.javax.persistence:javax.persistence_2.1:jar:2.7.2: [Could not transfer artifact org.apache.aries.jpa.javax.persistence:javax.persistence_2.1:jar:2.7.2 from/to maven.central.repo
(https://repo1.maven.org/maven2/): Connect to repo1.maven.org:443 [repo1.maven.org/151.101.36.209] failed: connect timed out, Could not transfer artifact
org.apache.aries.jpa.javax.persistence:javax.persistence_2.1:jar:2.7.2 from/to redhat.ga.repo
(https://maven.repository.redhat.com/ga/): Connect to maven.repository.redhat.com:443
[maven.repository.redhat.com/23.60.31.183] failed: connect timed out
Looking up the features I got this:
karaf#root()> feature:list | grep aries-jpa
jpa | 2.7.2 | | Uninstalled | aries-jpa-2.7.2 | OSGi Persistence Container
karaf#root()>
So the feature aries-jpa-2.7.2 is registered. Trying to install it:
karaf#root()> feature:install jpa
raises the same exception as mentioned above.
Could any one please shad some light here ?
Many thanks in advance.
Kind regards,
Nicolas
The issue was caused by the fact that Karaf doesn't use the local maven install and its settings but its own one. So, when behind an HTTP proxy, as it was my case, one needs either to configure Karaf to use the HTTP proxy or to tell it to use the local maven settings. This happens in $KARAF_HOME/etc/org.ops4j.pax.url.mvn.cfg file.

Staging Error while Pushing a Spring Application to Cloud Foundry

I am getting the following Error while pushing the Sample Hello World spring application on CloudFoundry.
Using manifest file C:\Users\I321571\Desktop\helo\Hello\manifest.yml
Updating app Hello in org trial / space I321571 as I321571...
OK
Uploading Hello...
Uploading app files from: C:\Users\I321571\Desktop\helo\Hello
Uploading 20.1K, 46 files
Done uploading
OK
Stopping app Hello in org trial / space I321571 as I321571...
OK
Starting app Hello in org trial / space I321571 as I321571...
-----> Downloaded app package (12K)
Cloning into '/tmp/buildpacks/java-buildpack'...
-----> Java Buildpack Version: b050954 | https://github.com/cloudfoundry/java-buildpack.git#b050954
[Buildpack] ERROR Compile failed with exception #<RuntimeError: No container can run this application. Please ensure that you've pushed a valid JVM artifact or artifacts using the
-p command line argument or path manifest entry. Information about valid JVM artifacts can be found at https://github.com/cloudfoundry/java-buildpack#additional-documentation. >
No container can run this application. Please ensure that you've pushed a valid JVM artifact or artifacts using the -p command line argument or path manifest entry. Information about valid JVM artifac
ts can be found at https://github.com/cloudfoundry/java-buildpack#additional-documentation.
Staging failed: Buildpack compilation step failed
FAILED
Error restarting application: BuildpackCompileFailed
TIP: use 'cf logs Hello --recent' for more information
this is my manifest.yml
applications:
- name: Hello
memory: 512M
instances: 1
Please help me in resolving the issue.
I encountered this error too!
Make sure the command given is valid
cf push {your-app-name} -p {path to your executable jar}

Apache Spark: Unable to build: [error] Server access Error..jetty

I have downloaded spark-1.4.1.tgz, unzipped it.
Now, when i try to build as follows, it gets stuck:
$ ./sbt/sbt assembly
Invoking 'build/sbt assembly' now
[info] Loading project definition...
[warn] Multiple resolvers having different access mechanism configured with same name 'sbt-plugin-releases'.
...
[info] Resolving org.eclipse.jetty#jetty-parent;18
[error] Server access Error: Connection reset url=http://download.eclipse.org/jgit/maven/org/eclipse/jetty/jetty-parent/18/jetty-parent-18.jar
[error] Server access Error: Server redirected too many times (20) url=http://scalasbt.artifactoryonline.com/scalasbt/sbt-plugin-releases/org.eclipse.jetty.orbit/jetty-orbit/1/jars/jetty-orbit.jar
The build gets stuck at this point. Am i missing any configuration/libraries?
You have a network connectivity issue and probably need to retry.

Resources