karaf 3.0.1 not starting up - maven

I'm trying to start karaf 3.0.1 in a solaris box (without internet) but getting the following error:
karaf: Ignoring predefined value for KARAF_HOME
Could not resolve mvn:org.eclipse/org.eclipse.osgi/3.8.2.v20130124-134944
and in karaf.log:
Jun 30, 2014 12:21:09 PM org.apache.karaf.main.Main main
SEVERE: Could not launch framework
java.lang.RuntimeException: Could not resolve mvn:org.eclipse/org.eclipse.osgi/3.8.2.v20130124-134944
at org.apache.karaf.main.util.SimpleMavenResolver.resolve(SimpleMavenResolver.java:59)
at org.apache.karaf.main.Main.createClassLoader(Main.java:315)
at org.apache.karaf.main.Main.launch(Main.java:234)
at org.apache.karaf.main.Main.main(Main.java:171)
the bundles are well in place (system folder) and the org.ops4j.pax.url.mvn.cfg file states:
org.ops4j.pax.url.mvn.repositories=\
file:${karaf.home}/${karaf.default.repository}#id=system.repository, \
file:${karaf.data}/kar#id=kar.repository#multi\
http://repo1.maven.org/maven2#id=central,\
http://repository.springsource.com/maven/bundles/release#id=spring.ebr.release,\
http://repository.springsource.com/maven/bundles/external#id=spring.ebr.external
I've tried running the framework using the three methods (server, service, client) but nothing seems to be working.
my environment is:
KARAF_BASE=/export/home/mehdi/bin/karaf
KARAF_HOME=/export/home/mehdi/bin/karaf
KARAF_ETC=/export/home/mehdi/bin/karaf/etc
KARAF_DATA=/export/home/mehdi/bin/karaf/data
JAVA_HOME=/opt/temp/jre1.7.0_13
I googled a bit and found a workaround which says to add -h 127.0.0.1 to the client script, but still nothing.

Related

Beam / DataFlow unexpected error ProtocolMessageEnum not implemented when using DataFlowRunner

When running my Beam pipeline locally it all works as expected but when trying to run it on the DataflowRunner I suddenly get the error below. Honestly I don't even know where to start evaluating this because the DataflowRunner seems to be a black box.
Jan 14, 2019 11:26:51 AM org.apache.beam.runners.dataflow.DataflowRunner fromOptions
INFO: PipelineOptions.filesToStage was not specified. Defaulting to files from the classpath: will stage 165 files. Enable logging at DEBUG level to see which files will be staged.
Exception in thread "main" java.lang.IncompatibleClassChangeError: Class org.apache.beam.model.pipeline.v1.RunnerApi$StandardPTransforms$Primitives does not implement the requested interface com.google.protobuf.ProtocolMessageEnum
at org.apache.beam.runners.core.construction.BeamUrns.getUrn(BeamUrns.java:27)
at org.apache.beam.runners.core.construction.PTransformTranslation.<clinit>(PTransformTranslation.java:58)
at org.apache.beam.runners.core.construction.UnconsumedReads$1.visitValue(UnconsumedReads.java:49)
at org.apache.beam.sdk.runners.TransformHierarchy$Node.visit(TransformHierarchy.java:666)
at org.apache.beam.sdk.runners.TransformHierarchy$Node.visit(TransformHierarchy.java:649)
at org.apache.beam.sdk.runners.TransformHierarchy$Node.visit(TransformHierarchy.java:649)
at org.apache.beam.sdk.runners.TransformHierarchy$Node.visit(TransformHierarchy.java:649)
at org.apache.beam.sdk.runners.TransformHierarchy$Node.access$600(TransformHierarchy.java:311)
at org.apache.beam.sdk.runners.TransformHierarchy.visit(TransformHierarchy.java:245)
at org.apache.beam.sdk.Pipeline.traverseTopologically(Pipeline.java:458)
at org.apache.beam.runners.core.construction.UnconsumedReads.ensureAllReadsConsumed(UnconsumedReads.java:40)
at org.apache.beam.runners.dataflow.DataflowRunner.replaceTransforms(DataflowRunner.java:868)
at org.apache.beam.runners.dataflow.DataflowRunner.run(DataflowRunner.java:660)
at org.apache.beam.runners.dataflow.DataflowRunner.run(DataflowRunner.java:173)
at org.apache.beam.sdk.Pipeline.run(Pipeline.java:313)
at org.apache.beam.sdk.Pipeline.run(Pipeline.java:299)
at (my code: pipe.run().waitUntilFinish();)
check the versions of beam etc and upgrade your dependencies where possible.
I had the same error and after seeing you get this error, I thought it must be a dependency conflict as it didn't exist before.
I'm using scio to deploy to dataflow and just referenced what they're using. https://github.com/spotify/scio/blob/v0.7.1/build.sbt
I updated guava and protobuf also.
I know you're using java, but try updating beam to 2.9.0 and maybe guava, protobuf...

Error with import BAR-File from OBIEE 12.2.1.0.0 to 12.2.1.2.0

I'm facing an error by importing the BAR-File of Version 12.2.1.0.0 to 12.2.1.2.0 importing it with importserviceinstance().
It is not a Linux machine, so I use the .cmd files like the wlst.cmd to export and import the BAR-File.
Export: exportServiceInstance(r'C:\DOMAIN_HOME','ssi',r'C:\WORK',r'C:\EXPORT','','','',true,'password')
So I used the option to export also all the user content by setting the second last to true.
Import: importServiceInstance(r'C:\DOMAIN_HOME','ssi',r'C:\EXPORT\ssi.bar',true,true,true,'password')
And here I set every option to true, to get everything (RPD, webcat, security Settings).
This is the error message prompted :
...
oracle.bi.bar.si.framework.handler.WebcatServiceInstanceHandler importServiceInstance
SEVERE: Failed in running importServiceInstance for plugin oracle.bi.migration.webcat.WebcatMetadataManagerGenericPlugin
Impl
oracle.bi.bar.si.framework.handler.ServiceInstanceHandlerHelper logErrorMessages
SEVERE: An error has occurred while executing command, Please check Catalog Manager log for more details.
oracle.bi.bar.si.framework.ServiceInstanceFrameWorkImpl importServiceInstance
SEVERE: Error during import service instance for handler oracle.bi.bar.si.framework.handler.WebcatServiceInstanceHandler
PM oracle.bi.bar.si.ServiceInstanceLifeCycleImpl importServiceInstance
SEVERE: Failed in running all the handlers during import service instance.
...
I've already checked the catalog manager logs, but this didn't help much to locate the problem.
Has anyone already faced the same error or has an idea how to solve this issue?
on windows, your path need to be modified from backslashes to forward
example:
instead of C:\DOMAIN_HOME you can use C:/DOMAIN_HOME
this is for all paths in the wlst commands

SonarQube Install - startsonar.bat fails

SonarQube 6.0 Installation
No plugins currently installed
Presently working through what appears to be connect string problem
In a command prompt, running as Administrator, I enter
StartSonar
This fails with rather cryptic output to the Command Prompt window. The following is what I believe is the relevant excerpt from the sonar.log file:
2016.08.11 12:23:29 INFO web[o.sonar.db.Database] Create JDBC data source for jdbc:sqlserver://localhost/DevOps;databaseName=sonar
2016.08.11 12:23:43 ERROR web[o.a.c.c.C.[.[.[/sonar]] Exception sending context initialized event to listener instance of class org.sonar.server.platform.PlatformServletContextListener
java.lang.IllegalStateException: Can not connect to database. Please check connectivity and settings (see the properties prefixed by 'sonar.jdbc.').
at org.sonar.db.DefaultDatabase.checkConnection(DefaultDatabase.java:104) ~[sonar-db-6.0.jar:na]
at org.sonar.db.DefaultDatabase.start(DefaultDatabase.java:71) ~[sonar-db-6.0.jar:na]
Relevant excerpt from .config file:
sonar.jdbc.url=jdbc:sqlserver://localhost/DevOps:1433;databaseName=sonar
sonar.jdbc.username=SonarQube
sonar.jdbc.password=PassWord1234
Finally, victory is mine! To be fair, I looked at the .config file from TeamCity. Thanks JetBrains!
The correct entries from the .config file (Note backslash / escape characters!):
sonar.jdbc.url=jdbc\:sqlserver\://localhost\\DevOps\:1433;databaseName\=Sonar
sonar.jdbc.username=SonarQube
sonar.jdbc.password=PassWord1234

Oracle Webtier 12.2.1.1 on Windows startComponent.cmd "exit status = -1073741515"

I am attempting to install Oracle HTTP server 12.2.1.1 on Windows Server 2012 R2 in standalone mode. I believe I followed the Oracle documentation correctly. I have progressed to this point: https://docs.oracle.com/middleware/1221/core/WTINS/GUID-3ED8EE64-EE8B-4EF3-84A1-19CEF21B8547.htm#WTINS414
My attempts to run "startComponent.cmd ohs1" have resulted in this error:
weblogic.nodemanager.NMException: Received error message from Node Manager Serve
r: [Server start command for OHS server 'ohs1' failed due to: [Failed t
o start the server ohs1
The error message also says to check a log file. Contents are:
<2016-06-29 15:42:10> <INFO> <OHS-4018> <Starting server ohs1>
<2016-06-29 15:42:10> <INFO> <OHS-0> <Running C:\Middleware\Oracle_Home_12\ohs\bin\launch.exe C:\Middleware\Oracle_Home_12\ohs\bin\httpd.exe -DOHS_MPM_WINNT -d C:/Middleware/Oracle_Home_12/user_projects/domains/base_domain/config/fmwconfig/components/OHS/instances/ohs1 -f C:\Middleware\Oracle_Home_12\user_projects\domains\base_domain\config\fmwconfig\components\OHS\instances\ohs1\httpd.conf>
<2016-06-29 15:42:11> <SEVERE> <OHS-0> <C:\Middleware\Oracle_Home_12\ohs\bin\launch.exe C:\Middleware\Oracle_Home_12\ohs\bin\httpd.exe -DOHS_MPM_WINNT -d C:/Middleware/Oracle_Home_12/user_projects/domains/base_domain/config/fmwconfig/components/OHS/instances/ohs1 -f C:\Middleware\Oracle_Home_12\user_projects\domains\base_domain\config\fmwconfig\components\OHS\instances\ohs1\httpd.conf: exit status = -1073741515>
<2016-06-29 15:42:11> <INFO> <OHS-4005> <Check the instance log file for more information: C:\Middleware\Oracle_Home_12\user_projects\domains\base_domain\servers\ohs1\logs\ohs1.log>
<2016-06-29 15:42:11> <SEVERE> <OHS-0> <Failed to start the server ohs1>
The log file referenced here, "C:\Middleware\Oracle_Home_12\user_projects\domains\base_domain\servers\ohs1\logs\ohs1.log", does not exist.
I also made sure that the required Microsoft redistributable was installed.
What am I doing wrong?
I received help from the Oracle Community:
Do you have lastinvocation.log file under
C:/Middleware/Oracle_Home_12/user_projects/domains\base_domain\servers\ohs1\logs?
If yes, rename it to lastinvocation.cmd and run this from command line and check if the ohs1.log file gets generated.
If yes, what is the error message seen?
The error message returned was:
"The program can't start because MSVCR110.dll is missing from your computer. Try reinstalling the program to fix this problem."
The user then pointed me to an Oracle Doc:
This is also described in the below article
Oracle HTTP Server 12.2.1 Fails To Start With Error 'MSVCR110.dll' Missing From Your Computer (Doc ID 2078713.1)
This Oracle article had a link to a DIFFERENT (larger) version of vcredist_x64.exe than I had used when installing the components. This version worked.
Thanks to Aishwarya in the Oracle community!

Spring-xd strange too many open files error

I upgraded from spring-xd 1.2.1 to 1.3.0, and have both under /opt on my system. After starting xd in single node (but configured to use Zookeeper), I tried to create another stream (e.g. "time | log"), and spring-xd throws the following exception:
java.io.FileNotFoundException: /opt/spring-xd-1.2.1.RELEASE/xd/config/modules/modules.yml (Too many open files)
I changed ulimit -n 60000, but it didn't solve the problem. The strange thing is why it still points to spring-xd-1.2.1.RELEASE? I have started both xd-singlenode and xd-shell under /opt/spring-xd-1.3.1.RELEASE
EDIT: add xd-singlenode running process output just to show it's pointing to 1.3.1:
/usr/java/default/bin/java -Dspring.application.name=admin
-Dlogging.config=file:/opt/spring-xd-1.3.0.RELEASE/xd/config//
/xd-singlenode-logback.groovy -Dxd.home=/opt/spring-xd-1.3.0.RELEASE/xd
-Dspring.config.location=file:/opt/spring-xd-1.3.0.RELEASE/xd/config//
-Dxd.config.home=file:/opt
/spring-xd-1.3.0.RELEASE/xd/config//
-Dspring.config.name=servers,application
-Dxd.module.config.location=file:/opt/spring-xd-1.3.0.RELEASE/xd/config//modules/
-Dxd.module.config.name=modules -classpath
/opt/spring-xd-1.3.0.RELEASE/xd/modules/processor/scripts:/opt/spring-xd
-1.3.0.RELEASE/xd/config:/opt/spring-xd-1.3.0.RELEASE/xd/lib/activation-
...
have you updated your environment variables? specifically XD_CONFIG_LOCATION based on the error shown above.

Resources