Ant BuildException error building Hadoop 2.4.1 - windows

I am trying to install Hadoop 2.4.1 from source on Windows 7, I got the following error on the last stage. I searched for a solution, but to no avail.
O/S Windows 7: 6.1 (32-bit)
Java: java version "1.8.0_11"
protoc 2.5.0
Apache Maven 3.2.2
I use this tutorials: https://wiki.apache.org/hadoop/Hadoop2OnWindows and https://www.srccodes.com/p/article/38/build-install-configure-run-apache-hadoop-2.2.0-microsoft-windows-os
Maven command:mvn package -Pdist,native-win -DskipTests -Dtar
Please help.
Error code:
[INFO] ------------------------------------------------------------------------
[INFO] Building Apache Hadoop Distribution 2.4.1
[INFO] ------------------------------------------------------------------------
[INFO]
[INFO] --- maven-antrun-plugin:1.7:run (create-testdirs) # hadoop-dist ---
[INFO] Executing tasks
main:
[INFO] Executed tasks
[INFO]
[INFO] --- maven-resources-plugin:2.2:resources (default-resources) # hadoop-dist ---
[INFO] Using default encoding to copy filtered resources.
[INFO]
[INFO] --- maven-compiler-plugin:2.5.1:compile (default-compile) # hadoop-dist ---
[INFO] No sources to compile
[INFO]
[INFO] --- maven-resources-plugin:2.2:testResources (default-testResources) # hadoop-dist ---
[INFO] Using default encoding to copy filtered resources.
[INFO]
[INFO] --- maven-compiler-plugin:2.5.1:testCompile (default-testCompile) # hadoop-dist ---
[INFO] No sources to compile
[INFO]
[INFO] --- maven-surefire-plugin:2.16:test (default-test) # hadoop-dist ---
[INFO] Tests are skipped.
[INFO]
[INFO] --- maven-antrun-plugin:1.7:run (dist) # hadoop-dist ---
[INFO] Executing tasks
main:
[exec]
[exec] Current directory /cygdrive/c/hdc/hadoop-dist/target
[exec]
[exec] $ rm -rf hadoop-2.4.1
[exec] $ mkdir hadoop-2.4.1
[exec] $ cd hadoop-2.4.1
[exec] $ cp -r /cygdrive/c/hdc/hadoop-common-project/hadoop-common/target/hadoop-common-2.4.1/bin /cygdrive/c/hdc/ha
doop-common-project/hadoop-common/target/hadoop-common-2.4.1/etc /cygdrive/c/hdc/hadoop-common-project/hadoop-common/targ
et/hadoop-common-2.4.1/libexec /cygdrive/c/hdc/hadoop-common-project/hadoop-common/target/hadoop-common-2.4.1/sbin /cygdr
ive/c/hdc/hadoop-common-project/hadoop-common/target/hadoop-common-2.4.1/share .
[exec] cp.exe: /cygdrive/c/hdc/hadoop-common-project/hadoop-common/target/hadoop-common-2.4.1/bin: No such file or d
irectory
[exec] cp.exe: /cygdrive/c/hdc/hadoop-common-project/hadoop-common/target/hadoop-common-2.4.1/etc: No such file or d
irectory
[exec] cp.exe: /cygdrive/c/hdc/hadoop-common-project/hadoop-common/target/hadoop-common-2.4.1/libexec: No such file
or directory
[exec] cp.exe: /cygdrive/c/hdc/hadoop-common-project/hadoop-common/target/hadoop-common-2.4.1/sbin: No such file or
directory
[exec] cp.exe: /cygdrive/c/hdc/hadoop-common-project/hadoop-common/target/hadoop-common-2.4.1/share: No such file or
directory
[exec]
[exec] Failed!
[exec]
[INFO] ------------------------------------------------------------------------
[INFO] Reactor Summary:
[INFO]
[INFO] Apache Hadoop Main ................................. SUCCESS [ 4.006 s]
[INFO] Apache Hadoop Project POM .......................... SUCCESS [ 1.592 s]
[INFO] Apache Hadoop Annotations .......................... SUCCESS [ 4.343 s]
[INFO] Apache Hadoop Assemblies ........................... SUCCESS [ 0.233 s]
[INFO] Apache Hadoop Project Dist POM ..................... SUCCESS [ 3.262 s]
[INFO] Apache Hadoop Maven Plugins ........................ SUCCESS [ 4.871 s]
[INFO] Apache Hadoop MiniKDC .............................. SUCCESS [ 4.450 s]
[INFO] Apache Hadoop Auth ................................. SUCCESS [ 3.150 s]
[INFO] Apache Hadoop Auth Examples ........................ SUCCESS [ 3.384 s]
[INFO] Apache Hadoop Common ............................... SUCCESS [02:26 min]
[INFO] Apache Hadoop NFS .................................. SUCCESS [ 9.297 s]
[INFO] Apache Hadoop Common Project ....................... SUCCESS [ 0.058 s]
[INFO] Apache Hadoop HDFS ................................. SUCCESS [03:31 min]
[INFO] Apache Hadoop HttpFS ............................... SUCCESS [ 34.454 s]
[INFO] Apache Hadoop HDFS BookKeeper Journal .............. SUCCESS [ 9.568 s]
[INFO] Apache Hadoop HDFS-NFS ............................. SUCCESS [ 4.962 s]
[INFO] Apache Hadoop HDFS Project ......................... SUCCESS [ 0.071 s]
[INFO] hadoop-yarn ........................................ SUCCESS [ 0.055 s]
[INFO] hadoop-yarn-api .................................... SUCCESS [ 42.210 s]
[INFO] hadoop-yarn-common ................................. SUCCESS [ 36.877 s]
[INFO] hadoop-yarn-server ................................. SUCCESS [ 0.048 s]
[INFO] hadoop-yarn-server-common .......................... SUCCESS [ 7.941 s]
[INFO] hadoop-yarn-server-nodemanager ..................... SUCCESS [ 14.402 s]
[INFO] hadoop-yarn-server-web-proxy ....................... SUCCESS [ 2.625 s]
[INFO] hadoop-yarn-server-applicationhistoryservice ....... SUCCESS [ 4.543 s]
[INFO] hadoop-yarn-server-resourcemanager ................. SUCCESS [ 16.911 s]
[INFO] hadoop-yarn-server-tests ........................... SUCCESS [ 1.438 s]
[INFO] hadoop-yarn-client ................................. SUCCESS [ 3.495 s]
[INFO] hadoop-yarn-applications ........................... SUCCESS [ 0.042 s]
[INFO] hadoop-yarn-applications-distributedshell .......... SUCCESS [ 2.033 s]
[INFO] hadoop-yarn-applications-unmanaged-am-launcher ..... SUCCESS [ 1.682 s]
[INFO] hadoop-yarn-site ................................... SUCCESS [ 0.051 s]
[INFO] hadoop-yarn-project ................................ SUCCESS [ 3.453 s]
[INFO] hadoop-mapreduce-client ............................ SUCCESS [ 0.117 s]
[INFO] hadoop-mapreduce-client-core ....................... SUCCESS [ 33.057 s]
[INFO] hadoop-mapreduce-client-common ..................... SUCCESS [ 23.963 s]
[INFO] hadoop-mapreduce-client-shuffle .................... SUCCESS [ 2.181 s]
[INFO] hadoop-mapreduce-client-app ........................ SUCCESS [ 13.202 s]
[INFO] hadoop-mapreduce-client-hs ......................... SUCCESS [ 8.889 s]
[INFO] hadoop-mapreduce-client-jobclient .................. SUCCESS [ 7.842 s]
[INFO] hadoop-mapreduce-client-hs-plugins ................. SUCCESS [ 1.379 s]
[INFO] Apache Hadoop MapReduce Examples ................... SUCCESS [ 5.019 s]
[INFO] hadoop-mapreduce ................................... SUCCESS [ 3.700 s]
[INFO] Apache Hadoop MapReduce Streaming .................. SUCCESS [ 3.600 s]
[INFO] Apache Hadoop Distributed Copy ..................... SUCCESS [ 12.786 s]
[INFO] Apache Hadoop Archives ............................. SUCCESS [ 2.083 s]
[INFO] Apache Hadoop Rumen ................................ SUCCESS [ 6.695 s]
[INFO] Apache Hadoop Gridmix .............................. SUCCESS [ 4.275 s]
[INFO] Apache Hadoop Data Join ............................ SUCCESS [ 1.995 s]
[INFO] Apache Hadoop Extras ............................... SUCCESS [ 2.223 s]
[INFO] Apache Hadoop Pipes ................................ SUCCESS [ 0.042 s]
[INFO] Apache Hadoop OpenStack support .................... SUCCESS [ 4.186 s]
[INFO] Apache Hadoop Client ............................... SUCCESS [ 7.937 s]
[INFO] Apache Hadoop Mini-Cluster ......................... SUCCESS [ 0.158 s]
[INFO] Apache Hadoop Scheduler Load Simulator ............. SUCCESS [ 3.874 s]
[INFO] Apache Hadoop Tools Dist ........................... SUCCESS [ 5.658 s]
[INFO] Apache Hadoop Tools ................................ SUCCESS [ 0.038 s]
[INFO] Apache Hadoop Distribution ......................... FAILURE [ 2.740 s]
[INFO] ------------------------------------------------------------------------
[INFO] BUILD FAILURE
[INFO] ------------------------------------------------------------------------
[INFO] Total time: 12:23 min
[INFO] Finished at: 2014-07-18T17:43:34+04:00
[INFO] Final Memory: 65M/238M
[INFO] ------------------------------------------------------------------------
[ERROR] Failed to execute goal org.apache.maven.plugins:maven-antrun-plugin:1.7:run (dist) on project hadoop-dist: An Ant
BuildException has occured: exec returned: 1
[ERROR] around Ant part ...<exec failonerror="true" dir="C:\hdc\hadoop-dist\target" executable="sh">... # 34:76 in C:\hdc
\hadoop-dist\target\antrun\build-main.xml
[ERROR] -> [Help 1]
[ERROR]
[ERROR] To see the full stack trace of the errors, re-run Maven with the -e switch.
[ERROR] Re-run Maven using the -X switch to enable full debug logging.
[ERROR]
[ERROR] For more information about the errors and possible solutions, please read the following articles:
[ERROR] [Help 1] http://cwiki.apache.org/confluence/display/MAVEN/MojoExecutionException
[ERROR]
[ERROR] After correcting the problems, you can resume the build with the command
[ERROR] mvn <goals> -rf :hadoop-dist

I met the same problem.
There are some information above the "Reactor Summar",
such as
"cp: cannot open `/cygdrive/c/hadoop/hadoop-common-project/hadoop-common/target/hadoop-common-3.0.0-SNAPSHOT/bin/hadoop.dll' for reading: Permission denied".
Just run the command prompt as admin

I just had the same problem - run the Maven command from within Cygwin and follow the instructions from here up to step 5.

I had the same issue and i fixed it by doing:
sudo apt-get install zlib1g-dev
sudo apt-get install libssl-dev
Source: https://stackoverflow.com/a/23385780/2245718
Cheers

Related

mvn sonar:sonar overrides the top level pom build time and runs sequentially at the end

After integrating sonar qube into my maven multi-module build and running
mvn -U -B -Dmaven.test.failure.ignore -Psonar verify sonar:sonar
I see the following build times
[2023-01-06T17:00:17.831Z] [INFO] ------------------------------------------------------------------------
[2023-01-06T17:00:17.831Z] [INFO] Reactor Summary for my-project 1.0.1:
[2023-01-06T17:00:17.831Z] [INFO]
[2023-01-06T17:00:17.831Z] [INFO] my-project ............................................ SUCCESS [01:52 min]
[2023-01-06T17:00:17.831Z] [INFO] my-project-common ..................................... SUCCESS [ 22.046 s]
[2023-01-06T17:00:17.831Z] [INFO] my-project-migration .................................. SUCCESS [ 3.998 s]
[2023-01-06T17:00:17.831Z] [INFO] my-project-test-common ................................ SUCCESS [ 33.710 s]
[2023-01-06T17:00:17.831Z] [INFO] my-project-dto ........................................ SUCCESS [ 3.010 s]
[2023-01-06T17:00:17.831Z] [INFO] my-project-producer ................................... SUCCESS [ 31.773 s]
[2023-01-06T17:00:17.831Z] [INFO] my-project-db ......................................... SUCCESS [01:07 min]
[2023-01-06T17:00:17.831Z] [INFO] my-project-utils ...................................... SUCCESS [ 23.637 s]
[2023-01-06T17:00:17.831Z] [INFO] my-project-server ..................................... SUCCESS [05:50 min]
[2023-01-06T17:00:17.831Z] [INFO] my-project-client ..................................... SUCCESS [ 32.758 s]
[2023-01-06T17:00:17.831Z] [INFO] my-project-http-client ................................ SUCCESS [ 5.322 s]
[2023-01-06T17:00:17.831Z] [INFO] my-project-integration ................................ SUCCESS [02:43 min]
[2023-01-06T17:00:17.831Z] [INFO] jacoco-coverage-report ................................ SUCCESS [ 8.389 s]
[2023-01-06T17:00:17.831Z] [INFO] my-project-other ...................................... SUCCESS [01:00 min]
[2023-01-06T17:00:17.831Z] [INFO] ------------------------------------------------------------------------
[2023-01-06T17:00:17.831Z] [INFO] BUILD SUCCESS
[2023-01-06T17:00:17.831Z] [INFO] ------------------------------------------------------------------------
[2023-01-06T17:00:17.831Z] [INFO] Total time: 15:22 min
[2023-01-06T17:00:17.831Z] [INFO] Finished at: 2023-01-06T17:00:17Z
[2023-01-06T17:00:17.831Z] [INFO] ------------------------------------------------------------------------
where the my-project .... build time of ~2 minutes took me by surprise.
Then on the console output I noticed the following
Building my-project 1.0.1 [15/14]
i.e. It builds the 15th step out of 14.
Examining the output it seems this is an extra step added by sonar. But it re-uses the the top level pom.xml name masking it's build time and hiding its own.
Is this expected? Is there a workaround to avoid the override?
I'm also curious if there is a way to integrate sonar and run its analysis in parallel rather than sequentially at the end?
i.e. It could analyse the finished modules while the next module is being built in parallel.
Note: The integration is via jacoco
<plugin>
<groupId>org.jacoco</groupId>
<artifactId>jacoco-maven-plugin</artifactId>
<version>0.8.5</version>
</plugin>

Failed to execute goal antrun plugin while building hadoop

While building the hadoop using maven, I encountered one error that one command in ant script exited abnormally.
more details are in following error.
[INFO] Apache Hadoop NFS .................................. SUCCESS [ 2.372 s]
[INFO] Apache Hadoop KMS .................................. FAILURE [ 5.222 s]
[INFO] Apache Hadoop Common Project ....................... SKIPPED
[INFO] Apache Hadoop HDFS ................................. SKIPPED
[INFO] Apache Hadoop HttpFS ............................... SKIPPED
[INFO] Apache Hadoop HDFS BookKeeper Journal .............. SKIPPED
[INFO] Apache Hadoop HDFS-NFS ............................. SKIPPED
[INFO] Apache Hadoop HDFS Project ......................... SKIPPED
[INFO] hadoop-yarn ........................................ SKIPPED
[INFO] hadoop-yarn-api .................................... SKIPPED
[INFO] hadoop-yarn-common ................................. SKIPPED
[INFO] hadoop-yarn-server ................................. SKIPPED
[INFO] hadoop-yarn-server-common .......................... SKIPPED
[INFO] hadoop-yarn-server-nodemanager ..................... SKIPPED
[INFO] hadoop-yarn-server-web-proxy ....................... SKIPPED
[INFO] hadoop-yarn-server-applicationhistoryservice ....... SKIPPED
[INFO] hadoop-yarn-server-resourcemanager ................. SKIPPED
[INFO] hadoop-yarn-server-tests ........................... SKIPPED
[INFO] hadoop-yarn-client ................................. SKIPPED
[INFO] hadoop-yarn-applications ........................... SKIPPED
[INFO] hadoop-yarn-applications-distributedshell .......... SKIPPED
[INFO] hadoop-yarn-applications-unmanaged-am-launcher ..... SKIPPED
[INFO] ------------------------------------------------------------------------
[INFO] BUILD FAILURE
[INFO] ------------------------------------------------------------------------
[INFO] Total time: 51.908 s
[INFO] Finished at: 2017-03-18T00:01:56+08:00
[INFO] Final Memory: 77M/771M
[INFO] ------------------------------------------------------------------------
[ERROR] Failed to execute goal org.apache.maven.plugins:maven-antrun-plugin:1.7:run (dist) on project hadoop-kms: An Ant BuildException has occured: exec returned: 1
[ERROR] around Ant part ...<exec failonerror="true" dir="/Users/jinteng/work/hadoop-compile/hadoop-2.6.5-src/hadoop-common-project/hadoop-kms/target" executable="sh">... # 10:142 in /Users/jinteng/work/hadoop-compile/hadoop-2.6.5-src/hadoop-common-project/hadoop-kms/target/antrun/build-main.xml
[ERROR] -> [Help 1]
[ERROR]
[ERROR] To see the full stack trace of the errors, re-run Maven with the -e switch.
[ERROR] Re-run Maven using the -X switch to enable full debug logging.
[ERROR]
I asked this question for sharing.
Background:
I killed the building process for the slow internet speed while downloading tomcat's tar packet.
Because of the interruption of downloading, the packet is not complete and tomcat-untar.sh would occured an exception while trying to untar the tomcat.
Later, I downloaded the packet separately and pushed it into downloads dir and found it works

apache atlas install error in web application

[Run command] export MAVEN_OPTS="-Xmx1536m -XX:MaxPermSize=512m" &&
mvn clean install
[Error message] Running org.apache.atlas.web.adapters.TestEntityREST
Tests run: 11, Failures: 1, Errors: 0, Skipped: 8, Time elapsed:
13.995 sec <<< FAILURE! - in org.apache.atlas.web.adapters.TestEntityREST enter code here
cleanup(org.apache.atlas.web.adapters.TestEntityREST) Time elapsed:
0.217 sec <<< FAILURE!
java.lang.NullPointerException: null at org.apache.atlas.RequestContext.clear(RequestContext.java:97) at
org.apache.atlas.web.adapters.TestEntityREST.cleanup(TestEntityREST.java:80)
...
[INFO]
------------------------------------------------------------------------ [INFO] Reactor Summary: [INFO] [INFO] Apache Atlas Server Build
Tools .................... SUCCESS [ 1.489 s] [INFO] apache-atlas
....................................... SUCCESS [ 8.805 s] [INFO]
Apache Atlas Integration ........................... SUCCESS [
42.642 s] [INFO] Apache Atlas Common ................................ SUCCESS [ 14.276 s] [INFO] Apache
Atlas Typesystem ............................ SUCCESS [01:15 min]
[INFO] Apache Atlas Client ................................ SUCCESS
[ 17.168 s] [INFO] Apache Atlas Server API
............................ SUCCESS [ 8.440 s] [INFO] Apache Atlas
Notification .......................... SUCCESS [ 38.640 s] [INFO]
Apache Atlas Graph Database Projects ............... SUCCESS [
0.553 s] [INFO] Apache Atlas Graph Database API .................... SUCCESS [ 5.432 s] [INFO] Graph Database Common Code
......................... SUCCESS [ 5.636 s] [INFO] Shaded version
of Apache hbase client .............. SUCCESS [ 10.728 s] [INFO]
Apache Atlas Titan 0.5.4 Graph DB Impl ............. SUCCESS [01:58
min] [INFO] Apache Atlas Graph Database Implementation Dependencies
SUCCESS [ 0.846 s] [INFO] Shaded version of Apache hbase server
.............. SUCCESS [ 23.850 s] [INFO] Apache Atlas Repository
............................ SUCCESS [13:25 min] [INFO] Apache Atlas
Authorization ......................... SUCCESS [ 13.707 s] [INFO]
Apache Atlas Business Catalog ...................... SUCCESS [
31.334 s] [INFO] Apache Atlas UI .................................... SUCCESS [01:11 min] [INFO]
Apache Atlas Web Application ....................... FAILURE [04:29
min] [INFO] Apache Atlas Documentation .........................
SKIPPED [INFO] Apache Atlas FileSystem Model ......................
SKIPPED [INFO] Apache Atlas Plugin Classloader ....................
SKIPPED [INFO] Apache Atlas Hive Bridge Shim ......................
SKIPPED [INFO] Apache Atlas Hive Bridge ...........................
SKIPPED [INFO] Apache Atlas Falcon Bridge Shim ....................
SKIPPED [INFO] Apache Atlas Falcon Bridge .........................
SKIPPED [INFO] Apache Atlas Sqoop Bridge Shim .....................
SKIPPED [INFO] Apache Atlas Sqoop Bridge ..........................
SKIPPED [INFO] Apache Atlas Storm Bridge Shim .....................
SKIPPED [INFO] Apache Atlas Storm Bridge ..........................
SKIPPED [INFO] Apache Atlas Distribution ..........................
SKIPPED [INFO]
------------------------------------------------------------------------ [INFO] BUILD FAILURE [INFO]
------------------------------------------------------------------------ [INFO] Total time: 26:06 min [INFO] Finished at:
2017-02-09T01:20:53+09:00 [INFO] Final Memory: 165M/1437M [INFO]
------------------------------------------------------------------------ [ERROR] Failed to execute goal
org.apache.maven.plugins:maven-surefire-plugin:2.18.1:test
(default-test) on project atlas-webapp: There are test failures.
[ERROR] [ERROR] Please refer to
/Users/dongkillee/dev/bin/atlas/webapp/target/surefire-reports for
the individual test results. [ERROR] -> [Help 1] [ERROR] [ERROR] To
see the full stack trace of the errors, re-run Maven with the -e
switch. [ERROR] Re-run Maven using the -X switch to enable full
debug logging. [ERROR] [ERROR] For more information about the
errors and possible solutions, please read the following articles:
[ERROR] [Help 1]
```
http://cwiki.apache.org/confluence/display/MAVEN/MojoFailureException
[ERROR] [ERROR] After correcting the problems, you can resume the
build with the command [ERROR] mvn -rf :atlas-webapp
This is testcase error. Use following command. It will create your build successfully.
export MAVEN_OPTS="-Xmx1536m -XX:MaxPermSize=512m" && mvn clean install -DskipTests
I have explained the complete solution, here. briefly, cause it uses some test cases in maven plugins in pom.xml file, we need to add -DskipTests argument when building the Apache Atlas from scratch.
Do the following items to run the project without any error:
git clone https://github.com/apache/atlas
mvn clean install -DskipTests
mvn clean package -Pdist -DskipTests

Hawkular server setup using maven give me error during build

Hawkular server setup using maven give me error during build. I am getting below error after run below command :
mvn clean install -Pdev -DskipTests -e
I am using git hub souce code to build this using link : https://github.com/hawkular/hawkular-apm/tree/master
[WARNING] npm WARN deprecated babel#6.5.2: Babel's CLI commands have been moved from the babel package to the babel-cli package
[WARNING] npm WARN deprecated vinyl-map#1.0.1: Missing dependencies resolved in v1.0.2 (Run 'npm update --save vinyl-map' to update package.json)
[WARNING] npm WARN deprecated minimatch#2.0.10: Please update to minimatch 3.0.2 or higher to avoid a RegExp DoS issue
[WARNING] npm WARN deprecated minimatch#0.2.14: Please update to minimatch 3.0.2 or higher to avoid a RegExp DoS issue
[WARNING] npm WARN deprecated graceful-fs#1.2.3: graceful-fs v3.0.0 and before will fail on node releases >= v7.0. Please update to graceful-fs#^4.0.0 as soon as possible. Use 'npm ls graceful-fs' to find it in the tree.
[WARNING] npm WARN deprecated minimatch#0.3.0: Please update to minimatch 3.0.2 or higher to avoid a RegExp DoS issue
[WARNING] npm WARN deprecated node-uuid#1.4.7: use uuid module instead
[WARNING] npm WARN deprecated graceful-fs#2.0.3: graceful-fs v3.0.0 and before will fail on node releases >= v7.0. Please update to graceful-fs#^4.0.0 as soon as possible. Use 'npm ls graceful-fs' to find it in the tree.
[INFO] ------------------------------------------------------------------------
[INFO] Reactor Summary:
[INFO]
[INFO] Hawkular APM ....................................... SUCCESS [ 11.622 s]
[INFO] Hawkular APM::Client::Instrumentation JVM .......... SUCCESS [ 2.024 s]
[INFO] Hawkular APM::API .................................. SUCCESS [ 12.477 s]
[INFO] Hawkular APM::Client::API .......................... SUCCESS [ 1.301 s]
[INFO] Hawkular APM::Server::API .......................... SUCCESS [ 4.397 s]
[INFO] Hawkular APM::Tests::Parent ........................ SUCCESS [ 0.743 s]
[INFO] Hawkular APM::Tests::Common ........................ SUCCESS [ 0.942 s]
[INFO] Hawkular APM::Client::Collector .................... SUCCESS [ 1.769 s]
[INFO] Hawkular APM::Client::Trace Publisher REST Client .. SUCCESS [ 0.600 s]
[INFO] Hawkular APM::Client::Instrumenter ................. SUCCESS [ 3.242 s]
[INFO] Hawkular APM::Client::Config Service REST Client ... SUCCESS [ 0.672 s]
[INFO] Hawkular APM::Client::Agent ........................ SUCCESS [ 11.770 s]
[INFO] Hawkular APM::Client::OpenTracing .................. SUCCESS [ 2.608 s]
[INFO] Hawkular APM::Client::OpenTracing Agent ............ SUCCESS [ 2.005 s]
[INFO] Hawkular APM::Client::Analytics Service REST Client SUCCESS [ 0.599 s]
[INFO] Hawkular APM::Client::Kafka ........................ SUCCESS [ 0.877 s]
[INFO] Hawkular APM::Client::Trace Service REST Client .... SUCCESS [ 0.556 s]
[INFO] Hawkular APM::Examples::Vertx Opentracing .......... SUCCESS [ 0.550 s]
[INFO] Hawkular APM::Examples::Vertx Opentracing::Common .. SUCCESS [ 1.142 s]
[INFO] Hawkular APM::Examples::Vertx Opentracing::Account Manager SUCCESS [ 3.121 s]
[INFO] Hawkular APM::Examples::Vertx Opentracing::Inventory Manager SUCCESS [ 2.613 s]
[INFO] Hawkular APM::Examples::Vertx Opentracing::Order Log SUCCESS [ 2.547 s]
[INFO] Hawkular APM::Examples::Vertx Opentracing::Order Manager SUCCESS [ 2.939 s]
[INFO] Hawkular APM::Server::Elasticsearch ................ SUCCESS [ 4.698 s]
[INFO] Hawkular APM::Server::Infinispan ................... SUCCESS [ 1.564 s]
[INFO] Hawkular APM::Server::Processors ................... SUCCESS [ 1.926 s]
[INFO] Hawkular APM::Server::Processors::Zipkin ........... SUCCESS [ 1.488 s]
[INFO] Hawkular APM::Server::JMS .......................... SUCCESS [ 2.246 s]
[INFO] Hawkular APM::Server::Kafka ........................ SUCCESS [ 1.072 s]
[INFO] Hawkular APM::Server::Processors::Alerts Publisher . SUCCESS [ 1.325 s]
[INFO] Hawkular APM::Server::REST ......................... SUCCESS [ 2.697 s]
[INFO] Hawkular APM::Server::Security Provider::JAAS ...... SUCCESS [ 0.566 s]
[INFO] Hawkular APM::Server::Zipkin REST .................. SUCCESS [ 0.690 s]
[INFO] Hawkular APM::Server::Zipkin WAR ................... SUCCESS [ 2.372 s]
[INFO] Hawkular APM::Server::WAR .......................... SUCCESS [ 2.235 s]
[INFO] Hawkular APM::UI::Console .......................... FAILURE [ 28.713 s]
[INFO] Hawkular APM::Distribution ......................... SKIPPED
[INFO] Hawkular APM::Tests::App::Swarm polyglot OpenTracing SKIPPED
[INFO] Hawkular APM::Tests::OpenTracing Agent ............. SKIPPED
[INFO] Hawkular APM::Tests::Containers::Jetty ............. SKIPPED
[INFO] Hawkular APM::Tests::Containers::Standalone ........ SKIPPED
[INFO] Hawkular APM::Tests::Containers::Wildfly ........... SKIPPED
[INFO] Hawkular APM::Tests::Instrumentation ............... SKIPPED
[INFO] Hawkular APM::Tests::Instrumentation Test Framework SKIPPED
[INFO] Hawkular APM::Tests::Distribution .................. SKIPPED
[INFO] ------------------------------------------------------------------------
[INFO] BUILD FAILURE
[INFO] ------------------------------------------------------------------------
[INFO] Total time: 02:05 min
[INFO] Finished at: 2017-01-18T03:45:12-05:00
[INFO] Final Memory: 127M/239M
[INFO] ------------------------------------------------------------------------
[ERROR] Failed to execute goal com.github.eirslett:frontend-maven-plugin:1.0:npm (npm install) on project hawkular-apm-ui: Failed to run task: 'npm install' failed. (error code 137) -> [Help 1]
org.apache.maven.lifecycle.LifecycleExecutionException: Failed to execute goal com.github.eirslett:frontend-maven-plugin:1.0:npm (npm install) on project hawkular-apm-ui: Failed to run task
at org.apache.maven.lifecycle.internal.MojoExecutor.execute(MojoExecutor.java:212)
at org.apache.maven.lifecycle.internal.MojoExecutor.execute(MojoExecutor.java:153)
at org.apache.maven.lifecycle.internal.MojoExecutor.execute(MojoExecutor.java:145)
at org.apache.maven.lifecycle.internal.LifecycleModuleBuilder.buildProject(LifecycleModuleBuilder.java:116)
at org.apache.maven.lifecycle.internal.LifecycleModuleBuilder.buildProject(LifecycleModuleBuilder.java:80)
at org.apache.maven.lifecycle.internal.builder.singlethreaded.SingleThreadedBuilder.build(SingleThreadedBuilder.java:51)
at org.apache.maven.lifecycle.internal.LifecycleStarter.execute(LifecycleStarter.java:128)
at org.apache.maven.DefaultMaven.doExecute(DefaultMaven.java:307)
at org.apache.maven.DefaultMaven.doExecute(DefaultMaven.java:193)
at org.apache.maven.DefaultMaven.execute(DefaultMaven.java:106)
at org.apache.maven.cli.MavenCli.execute(MavenCli.java:863)
at org.apache.maven.cli.MavenCli.doMain(MavenCli.java:288)
at org.apache.maven.cli.MavenCli.main(MavenCli.java:199)
at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62)
at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
at java.lang.reflect.Method.invoke(Method.java:497)
at org.codehaus.plexus.classworlds.launcher.Launcher.launchEnhanced(Launcher.java:289)
at org.codehaus.plexus.classworlds.launcher.Launcher.launch(Launcher.java:229)
at org.codehaus.plexus.classworlds.launcher.Launcher.mainWithExitCode(Launcher.java:415)
at org.codehaus.plexus.classworlds.launcher.Launcher.main(Launcher.java:356)
Caused by: org.apache.maven.plugin.MojoFailureException: Failed to run task
at com.github.eirslett.maven.plugins.frontend.mojo.AbstractFrontendMojo.execute(AbstractFrontendMojo.java:95)
at org.apache.maven.plugin.DefaultBuildPluginManager.executeMojo(DefaultBuildPluginManager.java:134)
at org.apache.maven.lifecycle.internal.MojoExecutor.execute(MojoExecutor.java:207)
... 20 more
Caused by: com.github.eirslett.maven.plugins.frontend.lib.TaskRunnerException: 'npm install' failed. (error code 137)
at com.github.eirslett.maven.plugins.frontend.lib.NodeTaskExecutor.execute(NodeTaskExecutor.java:60)
at com.github.eirslett.maven.plugins.frontend.mojo.NpmMojo.execute(NpmMojo.java:62)
at com.github.eirslett.maven.plugins.frontend.mojo.AbstractFrontendMojo.execute(AbstractFrontendMojo.java:89)
... 22 more
[ERROR]
[ERROR] Re-run Maven using the -X switch to enable full debug logging.
[ERROR]
[ERROR] For more information about the errors and possible solutions, please read the following articles:
[ERROR] [Help 1] http://cwiki.apache.org/confluence/display/MAVEN/MojoFailureException
[ERROR]
[ERROR] After correcting the problems, you can resume the build with the command
[ERROR] mvn <goals> -rf :hawkular-apm-ui
Thanks!!
I was fixed this issue at my end. I update ui pom.xml file for npm and bower install with install --allow-root. :)

Compilation error Spark 1.3.1

I tried to compile Spark 1.3.1 using the following flags
mvn -Pyarn -Phadoop-2.4 -Dhadoop.version=2.6.0 \
-Dscala-2.11 \
-Phive -Phive-0.13.1 -Phive-thriftserver \
-DskipTests clean package
The compilation failed with the following the error.
Any suggestions?
Thanks
[INFO] Reactor Summary:
[INFO]
[INFO] Spark Project Parent POM ........................... SUCCESS [01:08 min]
[INFO] Spark Project Core ................................. SUCCESS [02:38 min]
[INFO] Spark Project Bagel ................................ SUCCES [ 17.700 s]
[INFO] Spark Project GraphX ............................... SUCCESS [ 35.732 s]
[INFO] Spark Project ML Library ........................... SUCCESS [01:11 min]
[INFO] Spark Project Tools ................................ SUCCESS [ 6.718 s]
[INFO] Spark Project Networking ........................... SUCCESS [ 6.837 s]
[INFO] Spark Project Shuffle Streaming Service ............ SUCCESS [ 3.534 s]
[INFO] Spark Project Streaming ............................ SUCCESS [ 43.771 s]
[INFO] Spark Project Catalyst ............................. SUCCESS [ 48.411 s]
[INFO] Spark Project SQL .................................. SUCCESS [ 56.046 s]
[INFO] Spark Project Hive ................................. SUCCESS `enter code here`[01:01 min]
[INFO] Spark Project Assembly ............................. FAILURE [ 6.365 s]
[INFO] Spark Project External Twitter ..................... SKIPPED
[INFO] Spark Project External Flume ....................... SKIPPED
[INFO] Spark Project External Flume Sink .................. SKIPPED
[INFO] Spark Project External MQTT ........................ SKIPPED
[INFO] Spark Project External ZeroMQ ...................... SKIPPED
[INFO] Spark Project Examples ............................. SKIPPED
[INFO] Spark Project REPL ................................. SKIPPED
[INFO] Spark Project YARN ................................. SKIPPED
[INFO] Spark Project YARN Shuffle Service ................. SKIPPED
[INFO] Spark Project Hive Thrift Server ................... SKIPPED
[INFO] ------------------------------------------------------------------------
[INFO] BUILD FAILURE
[INFO] ------------------------------------------------------------------------
[INFO] Total time: 09:45 min
[INFO] Finished at: 2015-05-15T11:25:53-06:00
[INFO] Final Memory: 77M/1176M
[INFO] ------------------------------------------------------------------------
[ERROR] Failed to execute goal on project spark-assembly_2.10: Could not resolve dependencies for project org.apache.spark:spark-assembly_2.10:pom:1.3.1: Could not find artifact org.apache.spark:spark-hive-thriftserver_2.11:jar:1.3.1 in central (https://repo1.maven.org/maven2) -> [Help 1]
org.apache.maven.lifecycle.LifecycleExecutionException: Failed to execute goal on project spark-assembly_2.10: Could not resolve dependencies for project org.apache.spark:spark-assembly_2.10:pom:1.3.1: Could not find artifact org.apache.spark:spark-hive-thriftserver_2.11:jar:1.3.1 in central (https://repo1.maven.org/maven2)
Seems no dependency of spark-hive-thriftserver_2.11 exists In maven repository, I also search it manually in maven repo and find nothing. I think thrift server is not ready for scala2.11 for now.

Resources