JDeveloper 12.2.1.1 doesn't work in MacOS Sierra - oracle

Seems like JDeveloper doesn't work with macOS Sierra. I tried to reinstall, it didn't help. It just doesn't launch.
Is there some walkaround to solve the problem?
I tried also 12.2.1.0 version. The result is the same.
Upd. Setting Java home in product.conf doesn't help also.
Oracle JDeveloper 12c Development Build 12.2.1.1.0
Copyright (c) 1997, 2016, Oracle and/or its affiliates. All rights reserved.
java.lang.RuntimeException: Exception in org.eclipse.osgi.framework.internal.core.SystemBundleActivator.start() of bundle org.eclipse.osgi.
at org.eclipse.osgi.framework.internal.core.InternalSystemBundle.resume(InternalSystemBundle.java:233)
at org.eclipse.osgi.framework.internal.core.Framework.launch(Framework.java:664)
at org.eclipse.osgi.framework.internal.core.EquinoxLauncher.internalInit(EquinoxLauncher.java:69)
at org.eclipse.osgi.framework.internal.core.EquinoxLauncher.init(EquinoxLauncher.java:37)
at org.eclipse.osgi.launch.Equinox.init(Equinox.java:178)
at org.netbeans.modules.netbinox.Netbinox.init(Netbinox.java:84)
at org.netbeans.core.netigso.Netigso.prepare(Netigso.java:167)
at org.netbeans.NetigsoHandle.turnOn(NetigsoHandle.java:138)
at org.netbeans.ModuleManager.enable(ModuleManager.java:1346)
at org.netbeans.ModuleManager.enable(ModuleManager.java:1163)
at org.netbeans.core.startup.ModuleList.installNew(ModuleList.java:340)
at org.netbeans.core.startup.ModuleList.trigger(ModuleList.java:276)
at org.netbeans.core.startup.ModuleSystem.restore(ModuleSystem.java:301)
at org.netbeans.core.startup.Main.getModuleSystem(Main.java:181)
at org.netbeans.core.startup.Main.getModuleSystem(Main.java:150)
at org.netbeans.core.startup.Main.start(Main.java:307)
at org.netbeans.core.startup.TopThreadGroup.run(TopThreadGroup.java:123)
at java.lang.Thread.run(Thread.java:745)
Caused by: org.osgi.framework.BundleException: Exception in org.eclipse.osgi.framework.internal.core.SystemBundleActivator.start() of bundle org.eclipse.osgi.
at org.eclipse.osgi.framework.internal.core.BundleContextImpl.startActivator(BundleContextImpl.java:734)
at org.eclipse.osgi.framework.internal.core.BundleContextImpl.start(BundleContextImpl.java:683)
at org.eclipse.osgi.framework.internal.core.InternalSystemBundle.resume(InternalSystemBundle.java:225)
... 17 more
Caused by: java.lang.NumberFormatException: For input string: "2.0"
at java.lang.NumberFormatException.forInputString(NumberFormatException.java:65)
at java.lang.Integer.parseInt(Integer.java:580)
at java.lang.Integer.parseInt(Integer.java:615)
at org.eclipse.osgi.internal.resolver.StateBuilder.createBundleDescription(StateBuilder.java:61)
at org.eclipse.osgi.internal.resolver.StateObjectFactoryImpl.createBundleDescription(StateObjectFactoryImpl.java:33)
at org.eclipse.osgi.internal.baseadaptor.BaseStorage.readStateData(BaseStorage.java:853)
at org.eclipse.osgi.internal.baseadaptor.BaseStorage.getStateManager(BaseStorage.java:799)
at org.eclipse.osgi.baseadaptor.BaseAdaptor.getState(BaseAdaptor.java:387)
at org.eclipse.osgi.internal.baseadaptor.BaseStorage.frameworkStart(BaseStorage.java:923)
at org.eclipse.osgi.baseadaptor.BaseAdaptor.frameworkStart(BaseAdaptor.java:250)
at org.eclipse.osgi.framework.internal.core.SystemBundleActivator.start(SystemBundleActivator.java:60)
at org.eclipse.osgi.framework.internal.core.BundleContextImpl$1.run(BundleContextImpl.java:711)
at java.security.AccessController.doPrivileged(Native Method)
at org.eclipse.osgi.framework.internal.core.BundleContextImpl.startActivator(BundleContextImpl.java:702)
... 19 more

Which Java version do you have installed (java -version)?
Try removing the systemxxx directory (usually under .jdeveloper directory) and restarting JDeveloper.
Works for me fine on El Capitan

If you did an OS update, the java home directory probably changed. Check the product.conf file in your .jdeveloper directory, and make sure the path to the JDK is correct

We need to remove system_cache folder only under path <USER_HOME>/.jdeveloper/system12.2.x.x.x.x.x/system_cache.
Please refer- https://migrmrz.wordpress.com/2016/08/29/jdeveloper-12c-wont-start-in-mac-os-x-el-capitan/

Related

Running liquibase script in google cloud platform

I am trying to run liquibase to create DB schema and tables in GCP.
Below error is coming any idea ?
Class [org.hibernate.boot.jaxb.hbm.spi.package-info] could not be found.
Processing bindings will probably fail.
java.lang.ClassNotFoundException: org.hibernate.boot.jaxb.hbm.spi.package-info
at org.codehaus.plexus.classworlds.strategy.SelfFirstStrategy.loadClass (SelfFirstStrategy.java:50)
at org.codehaus.plexus.classworlds.realm.ClassRealm.unsynchronizedLoadClass (ClassRealm.java:271)
at org.codehaus.plexus.classworlds.realm.ClassRealm.loadClass (ClassRealm.java:247)
at org.codehaus.plexus.classworlds.realm.ClassRealm.loadClass (ClassRealm.java:239)
at java.lang.Class.forName0 (Native Method)
at java.lang.Class.forName (Class.java:315)
I had to downgrade my Liquibase version to 3.6.3 (as suggested by #Jens). Now I am not getting the error.
Stack I use: Ubuntu 18.04 / OpenJDK 11
We had the same issue but it was only warning and it was not a stopper for us. To remove the warning, we upgraded the Liquibase version into version - 4.7.1 and the warning is gone.

KiteSdk 1.1.0 csv-import IOError

with HDP-2.5 on Ubuntu-14.04, running this command and
$ ./kite-dataset csv-import ./test.csv test_schema
trying to import raw csv data into Hive using the KiteSdk ver.1-1-0
and having the following IOError:
1 job failure(s) occurred: org.kitesdk.tools.CopyTask:
Kite(dataset:file:/tmp/444e6fc4-10e2-407d-afaf-723c408a6d... ID=1
(1/1)(1): java.io.FileNotFoundException: File
file:/hdp/apps/2.5.0.0-1245/mapreduce/mapreduce.tar.gz does not exist
at org.apache.hadoop.fs.RawLocalFileSystem.deprecatedGetFileStatus(RawLocalFileSystem.java:624)
at org.apache.hadoop.fs.RawLocalFileSystem.getFileLinkStatusInternal(RawLocalFileSystem.java:850)
at org.apache.hadoop.fs.RawLocalFileSystem.getFileStatus(RawLocalFileSystem.java:614)
at org.apache.hadoop.fs.DelegateToFileSystem.getFileStatus(DelegateToFileSystem.java:125)
at org.apache.hadoop.fs.AbstractFileSystem.resolvePath(AbstractFileSystem.java:468)
at org.apache.hadoop.fs.FilterFs.resolvePath(FilterFs.java:158)
at org.apache.hadoop.fs.FileContext$25.next(FileContext.java:2195)
at org.apache.hadoop.fs.FileContext$25.next(FileContext.java:2191)
at org.apache.hadoop.fs.FSLinkResolver.resolve(FSLinkResolver.java:90)
at org.apache.hadoop.fs.FileContext.resolve(FileContext.java:2191)
at org.apache.hadoop.fs.FileContext.resolvePath(FileContext.java:603)
at org.apache.hadoop.mapreduce.JobSubmitter.addMRFrameworkToDistributedCache(JobSubmitter.java:457)
at org.apache.hadoop.mapreduce.JobSubmitter.submitJobInternal(JobSubmitter.java:142)
at org.apache.hadoop.mapreduce.Job$10.run(Job.java:1290)
at org.apache.hadoop.mapreduce.Job$10.run(Job.java:1287)
at java.security.AccessController.doPrivileged(Native Method)
at javax.security.auth.Subject.doAs(Subject.java:422)
at org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1724)
at org.apache.hadoop.mapreduce.Job.submit(Job.java:1287)
at org.apache.crunch.hadoop.mapreduce.lib.jobcontrol.CrunchControlledJob.submit(CrunchControlledJob.java:329)
at org.apache.crunch.hadoop.mapreduce.lib.jobcontrol.CrunchJobControl.startReadyJobs(CrunchJobControl.java:204)
at org.apache.crunch.hadoop.mapreduce.lib.jobcontrol.CrunchJobControl.pollJobStatusAndStartNewOnes(CrunchJobControl.java:238)
at org.apache.crunch.impl.mr.exec.MRExecutor.monitorLoop(MRExecutor.java:112)
at org.apache.crunch.impl.mr.exec.MRExecutor.access$000(MRExecutor.java:55)
at org.apache.crunch.impl.mr.exec.MRExecutor$1.run(MRExecutor.java:83)
at java.lang.Thread.run(Thread.java:745)
I've checked the file "hdfs:/hdp/apps/2.5.0.0-1245/mapreduce/mapreduce.tar.gz"
exists and can't figure out how to resolve this error for quite a while.
Any help is greatly appreciated.
I was experiencing the same error, and I resolved it by creating /hdp/apps/2.5.0.0-1245/mapreduce and then:
cp /usr/hdp/current/hadoop-client/mapreduce.tar.gz /hdp/apps/2.5.0.0-1245/mapreduce
This then created a new error :org.kitesdk.tools.CopyTask: Kite(dataset:file:/tmp/413a41a2-8813-4056-9433-3c5e073d80... ID=1 (1/1)(1): java.io.FileNotFoundException: File does not exist: hdfs://sandbox.hortonworks.com:8020/tmp/crunch-283520469/p1/REDUCE
Which I'm still trying to troubleshoot.
I think you are getting this error since you are using Kite SDK 1.1.0 version. I also got similar error when I was doing csv-import. When I switched to Kite SDK 1.0.0 version there was no such error.
I would suggest you to switch to Kite SDK 1.0.0 version.
Moreover there has been no new release of Kite SDK after 1.1.0 version and even this release happened in June 2015.

Exception in thread "main" java.lang.UnsupportedClassVersionError: org/apache/hadoop/fs/FsShell : Unsupported major.minor version 51.0

I am trying to execute below command on Hadoop
hadoop fs -ls /
but it is returing with the error
Exception in thread "main" java.lang.UnsupportedClassVersionError: org/apache/hadoop/fs/FsShell : Unsupported major.minor version 51.0
at java.lang.ClassLoader.defineClass1(Native Method)
at java.lang.ClassLoader.defineClassCond(ClassLoader.java:632)
at java.lang.ClassLoader.defineClass(ClassLoader.java:616)
at java.security.SecureClassLoader.defineClass(SecureClassLoader.java:141)
at java.net.URLClassLoader.defineClass(URLClassLoader.java:283)
at java.net.URLClassLoader.access$000(URLClassLoader.java:58)
at java.net.URLClassLoader$1.run(URLClassLoader.java:197)
at java.security.AccessController.doPrivileged(Native Method)
at java.net.URLClassLoader.findClass(URLClassLoader.java:190)
at java.lang.ClassLoader.loadClass(ClassLoader.java:307)
at sun.misc.Launcher$AppClassLoader.loadClass(Launcher.java:301)
at java.lang.ClassLoader.loadClass(ClassLoader.java:248)
Could not find the main class: org.apache.hadoop.fs.FsShell. Program will exit.
I have tried updating the java but it is still giving me the same error.
Note: The same command is working on the other nodes but not on 2 for the cluster nodes.
Try updating the JDK to version 1.7. Perhaps, you have updated JRE not JDK.
If you used the CDH, you may need change the java version to jdk1.7.0_67-cloudera. After I changed the JAVA_HOME from /usr/java/jdk1.6.0_31 to /usr/java/jdk1.7.0_67-cloudera, I solved the problem.
Unsupported major.minor version 51.0
You need Java 7 (or higher) to run this.
See: https://www.java.com/de/download/faq/java_7.xml
If you already installed Java 7 (or higher) then execute it with:
C:\Program Files\Java\<Java Version (jre7/jre8>\bin\java.exe -jar <Path To .Jar>
The command is not working because it is pointing to the newer version of hadoop jars avilable on the nodes instead of the hadoop jars of installed version
It was pointing to jars placed in
/usr/lib/hadoop
Then i tried to execute it from the installation directory as below
/opt/cloudera/parcels/CDH/lib/hadoop/bin/hadoop fs -ls /
It worked for me.

Gradle throws: Unable to establish loopback connection

Gradle throws exception with message "Unable to establish loopback connection". The following is the stack trace thrown at the console.
java.io.IOException: Unable to establish loopback connection
org.gradle.internal.UncheckedException: java.io.IOException: Unable to establish loopback connection
at org.gradle.internal.UncheckedException.throwAsUncheckedException(UncheckedException.java:39)
at org.gradle.messaging.remote.internal.inet.SocketConnection.<init>(SocketConnection.java:58)
at org.gradle.messaging.remote.internal.inet.SocketConnectCompletion.create(SocketConnectCompletion.java:43)
at org.gradle.messaging.remote.internal.hub.MessageHubBackedObjectConnection.connect(MessageHubBackedObjectConnection.java:92)
at org.gradle.api.internal.tasks.testing.worker.ForkingTestClassProcessor.forkProcess(ForkingTestClassProcessor.java:78)
at org.gradle.api.internal.tasks.testing.worker.ForkingTestClassProcessor.processTestClass(ForkingTestClassProcessor.java:56)
at org.gradle.api.internal.tasks.testing.processors.RestartEveryNTestClassProcessor.processTestClass(RestartEveryNTestClassProcessor.java:45)
at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:57)
at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
at java.lang.reflect.Method.invoke(Method.java:606)
at org.gradle.messaging.dispatch.ReflectionDispatch.dispatch(ReflectionDispatch.java:35)
at org.gradle.messaging.dispatch.ReflectionDispatch.dispatch(ReflectionDispatch.java:24)
at org.gradle.messaging.dispatch.FailureHandlingDispatch.dispatch(FailureHandlingDispatch.java:29)
at org.gradle.messaging.dispatch.AsyncDispatch.dispatchMessages(AsyncDispatch.java:132)
at org.gradle.messaging.dispatch.AsyncDispatch.access$000(AsyncDispatch.java:33)
at org.gradle.messaging.dispatch.AsyncDispatch$1.run(AsyncDispatch.java:72)
at org.gradle.internal.concurrent.DefaultExecutorFactory$StoppableExecutorImpl$1.run(DefaultExecutorFactory.java:64)
at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1145)
at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:615)
at java.lang.Thread.run(Thread.java:744)
Gradle version :
Gradle 1.12
Build time: 2014-04-29 09:24:31 UTC
Build number: none
Revision: a831fa866d46cbee94e61a09af15f9dd95987421
Groovy: 1.8.6
Ant: Apache Ant(TM) version 1.9.3 compiled on December 23 2013
Ivy: 2.2.0
JVM: 1.7.0_45 (Oracle Corporation 24.45-b08)
OS: Windows 7 6.1 x86
Jdk is 1.7.45.
Is there any help? I tried stopping the anti virus assuming the connection are getting terminated by AV. However, the issue re-occurred when the AV was turned off.
I assume this is a problem where Gradle cannot connect to the gradle Daemon. Your machine probably has an issue with its networking config, which needs to be worked out.
A possible workaround would be to run without the Daemon, using the --no-daemon option. If that still doesn't work, there is some other reason why it tries to establish a connection, probably due to a custom plugin or build script.
I was using JDK v1.7.0_45 and I now use 1.7.0_79. Ever since the upgrade I am not observing this issue. I switched back to v1.7.0_45 this morning and I am already noticed "Unable to establish loopback connection" twice out of six iterations.

Sonar Runner not starting

I am trying to configure Sonar 3.2 in my Windows 7 64 bit operating system with java 1.7.0_05.
When I start Sonar runner - 1.4, I am getting this exception
Exception in thread "main" org.sonar.batch.bootstrapper.BootstrapException: org.sonar.batch.bootstrapper.BootstrapException: Fail to download the file: http://localhost:9000/batch/hibernate-entitymanager-3.4.0.GA.jar
at org.sonar.batch.bootstrapper.Bootstrapper.downloadBatchFiles(Bootstrapper.java:164)
at org.sonar.batch.bootstrapper.Bootstrapper.createClassLoader(Bootstrapper.java:87)
at org.sonar.runner.Runner.createClassLoader(Runner.java:176)
at org.sonar.runner.Runner.execute(Runner.java:84)
at org.sonar.runner.Main.main(Main.java:67)
Caused by: org.sonar.batch.bootstrapper.BootstrapException: Fail to download the file: http://localhost:9000/batch/hibernate-entitymanager-3.4.0.GA.jar
at org.sonar.batch.bootstrapper.Bootstrapper.remoteContentToFile(Bootstrapper.java:113)
at org.sonar.batch.bootstrapper.Bootstrapper.downloadBatchFiles(Bootstrapper.java:159)
... 4 more
Caused by: java.net.SocketTimeoutException: Read timed out
at java.net.SocketInputStream.socketRead0(Native Method)
at java.net.SocketInputStream.read(SocketInputStream.java:150)
at java.net.SocketInputStream.read(SocketInputStream.java:121)
at java.io.BufferedInputStream.fill(BufferedInputStream.java:235)
at java.io.BufferedInputStream.read1(BufferedInputStream.java:275)
at java.io.BufferedInputStream.read(BufferedInputStream.java:334)
at sun.net.www.http.ChunkedInputStream.fastRead(ChunkedInputStream.java:239)
at sun.net.www.http.ChunkedInputStream.read(ChunkedInputStream.java:680)
at java.io.FilterInputStream.read(FilterInputStream.java:133)
at sun.net.www.protocol.http.HttpURLConnection$HttpInputStream.read(HttpURLConnection.java:3035)
at sun.net.www.protocol.http.HttpURLConnection$HttpInputStream.read(HttpURLConnection.java:3029)
at org.sonar.batch.bootstrapper.BootstrapperIOUtils.copyLarge(BootstrapperIOUtils.java:63)
at org.sonar.batch.bootstrapper.Bootstrapper.remoteContentToFile(Bootstrapper.java:109)
any solution to solve this issue is appreciated.
I am able to download hibernate-entitymanager-3.4.0.GA.jar from http://localhost:9000/batch/hibernate-entitymanager-3.4.0.GA.jar in the browser.
I see in your configuration details that you are running Sonar using a Java 7 VM: could you please try with Java 6?
That might not be the root cause, but I know we have users who reported some issues with Java 7.
This is the error I get when I execute the runner without administrative privileges.
If you're launching the runner from a DOS prompt, try right-clicking on "Windows Command Processor" and choosing "Run as Administrator."

Resources