Unable to resolve dummy/0.0.0: missing requirement - fabric8

I've a problem when try deploy a profile on a cluster of fabric8 CR5.
this's a log part on karaf.log
2014-09-09 12:33:43,142 | INFO | o.fabric8.agent) | DeploymentAgent | 66 - io.fabric8.fabric-agent - 1.1.0.CR5 | DeploymentAgent updated with {repository.karaf-enterprise=mvn:org.apache.karaf.assemblies.features/enterprise/2.3.0.redhat-610379/xml/features, optional.mvn:org.apache.activemq/activemq-osgi/5.10.0=mvn:org.apache.activemq/activemq-osgi/5.10.0, attribute.parents=feature-camel, feature.insight-log=insight-log, org.ops4j.pax.url.mvn.defaultrepositories=file:/opt/fabric8-karaf-1.1.0.CR5/system#snapshots#id=karaf-default,file:/opt/fabric8-karaf-1.1.0.CR5/local-repo#snapshots#id=karaf-local, feature.camel=camel, feature.fabric-jaas=fabric-jaas, attribute.abstract=false, fabric.zookeeper.pid=io.fabric8.agent, bundle.mvn:anagrafiche_bloomberg/mortgages/0.1=mvn:anagrafiche_bloomberg/Mortgages/0.1, feature.camel-blueprint=camel-blueprint, bundle.mvn:anagrafiche_bloomberg/varcomty/0.1=mvn:anagrafiche_bloomberg/Varcomty/0.1, feature.fabric-core=fabric-core, optional.ops4j-base-lang=mvn:org.ops4j.base/ops4j-base-lang/1.4.0, service.pid=io.fabric8.agent, org.ops4j.pax.url.mvn.repositories=http://repo1.maven.org/maven2#id=central, https://repo.fusesource.com/nexus/content/groups/public#id=fusepublic, https://repository.jboss.org/nexus/content/repositories/public#id=jbosspublic, https://repo.fusesource.com/nexus/content/repositories/releases#id=jbossreleases, https://repo.fusesource.com/nexus/content/groups/ea#id=jbossearlyaccess, http://repository.springsource.com/maven/bundles/release#id=ebrreleases, http://repository.springsource.com/maven/bundles/external#id=ebrexternal, http://10.0.0.124:8081/artifactory/libs-release-local#id=pippo, feature.fabric-web=fabric-web, feature.fabric-git-server=fabric-git-server, repository.apache-camel=mvn:org.apache.camel.karaf/apache-camel/2.13.2/xml/features, bundle.mvn:org.talend.esb.job/org.talend.esb.job.api/5.4.1=mvn:org.talend.esb.job/org.talend.esb.job.api/5.4.1, bundle.mvn:anagrafiche_bloomberg/equities/0.1=mvn:anagrafiche_bloomberg/Equities/0.1, resolve.optional.imports=false, repository.fabric8=mvn:io.fabric8/fabric8-karaf/1.1.0.CR5/xml/features, feature.camel-amq=camel-amq, optional.mvn:org.talend.esb.job/org.talend.esb.job.api/5.4.1=mvn:org.talend.esb.job/org.talend.esb.job.api/5.4.1, bundle.mvn:anagrafiche_bloomberg/equity_option_warrants/0.1=mvn:anagrafiche_bloomberg/Equity_Option_Warrants/0.1, feature.karaf=karaf, feature.camel-ftp=camel-ftp, feature.fabric-agent=fabric-agent, feature.jolokia=jolokia, feature.fabric-git=fabric-git, feature.camel-core=camel-core, feature.fabric-camel=fabric-camel, hash=ProfileImpl[id='Anagrafica_Bloomberg', version='1.0']-328b4e0-9859682-9859682-cbe9c83, bundle.mvn:anagrafiche_bloomberg/commodities/0.1=mvn:anagrafiche_bloomberg/Commodities/0.1, repository.karaf-spring=mvn:org.apache.karaf.assemblies.features/spring/2.3.0.redhat-610379/xml/features, lastrefresh.anagrafica_bloomberg=1410258607347, patch.repositories=https://repo.fusesource.com/nexus/content/repositories/releases, https://repo.fusesource.com/nexus/content/groups/ea, repository.karaf-standard=mvn:org.apache.karaf.assemblies.features/standard/2.3.0.redhat-610379/xml/features, bundle.mvn:anagrafiche_bloomberg/vargov_corp_muni/0.1=mvn:anagrafiche_bloomberg/Vargov_Corp_Muni/0.1, bundle.mvn:it.gestielle/anagrafica_bloomberg/1.0=mvn:it.gestielle/Anagrafica_Bloomberg/1.0}
2014-09-09 12:33:44,157 | ERROR | agent-1-thread-1 | DeploymentAgent | 66 - io.fabric8.fabric-agent - 1.1.0.CR5 | Unable to update agent
org.osgi.service.resolver.ResolutionException: Unable to resolve dummy/0.0.0: missing requirement [dummy/0.0.0] osgi.identity; osgi.identity=anagrafiche_bloomberg.Commodities; type=osgi.bundle; version="[0.1.0,0.1.0]" [caused by: Unable to resolve anagrafiche_bloomberg.Commodities/0.1.0: missing requirement [anagrafiche_bloomberg.Commodities/0.1.0] osgi.wiring.package; filter:="(osgi.wiring.package=org.apache.cxf.management.counters)"]
at org.apache.felix.resolver.Candidates.populateResource(Candidates.java:285)[66:io.fabric8.fabric-agent:1.1.0.CR5]
at org.apache.felix.resolver.Candidates.populate(Candidates.java:153)[66:io.fabric8.fabric-agent:1.1.0.CR5]
at org.apache.felix.resolver.ResolverImpl.resolve(ResolverImpl.java:148)[66:io.fabric8.fabric-agent:1.1.0.CR5]
at io.fabric8.agent.DeploymentBuilder.resolve(DeploymentBuilder.java:224)[66:io.fabric8.fabric-agent:1.1.0.CR5]
at io.fabric8.agent.DeploymentAgent.doUpdate(DeploymentAgent.java:575)[66:io.fabric8.fabric-agent:1.1.0.CR5]
at io.fabric8.agent.DeploymentAgent$2.run(DeploymentAgent.java:301)[66:io.fabric8.fabric-agent:1.1.0.CR5]
at java.util.concurrent.Executors$RunnableAdapter.call(Executors.java:471)[:1.7.0_65]
at java.util.concurrent.FutureTask.run(FutureTask.java:262)[:1.7.0_65]
at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1145)[:1.7.0_65]
at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:615)[:1.7.0_65]
at java.lang.Thread.run(Thread.java:745)[:1.7.0_65]
this is my local repository
http://10.0.0.124:8081/artifactory/libs-release-local#id=pippo
set on org.ops4j.pax.url.mvn.repositories property where the jar are deploy.
Can someone help me ??
Why the dependencies are not resolve?
Thanks
Mirko

dummy/0.0.0.0 is just the entry point -- read the rest of the error where it walks the dependency chain.
You'll see it's requiring anagrafiche_bloomberg.Commodities, which is requiring org.apache.cxf.management.counters. Be sure to install the CXF package.

Related

Spring cloud Eureka error java.lang.IllegalStateException: Error processing condition on when running

Issue/Problem Description : I am not able to start server after spring cloud migration 2021 0 4 and Java 17 . Can any one help me on this
I have tried to upgrade various spring eureka jars . I need to start application without any errors
Error log :
java.lang.IllegalStateException: Error processing condition on org.springframework.cloud.netflix.eureka.config.DiscoveryClientOptionalArgsConfiguration.defaultEurekaClientHttpRequestFactorySupplier\n\t
at org.springframework.boot.autoconfigure.condition.SpringBootCondition.matches(SpringBootCondition.java:60)\n\t
at org.springframework.context.annotation.ConditionEvaluator.shouldSkip(ConditionEvaluator.java:108)\n\t
at org.springframework.context.annotation.ConfigurationClassBeanDefinitionReader.loadBeanDefinitionsForBeanMethod(ConfigurationClassBeanDefinitionReader.java:193)\n\t
at org.springframework.context.annotation.ConfigurationClassBeanDefinitionReader.loadBeanDefinitionsForConfigurationClass(ConfigurationClassBeanDefinitionReader.java:153)\n\t
at org.springframework.context.annotation.ConfigurationClassBeanDefinitionReader.loadBeanDefinitions(ConfigurationClassBeanDefinitionReader.java:129)\n\t
at org.springframework.context.annotation.ConfigurationClassPostProcessor.processConfigBeanDefinitions(ConfigurationClassPostProcessor.java:343)\n\t
at org.springframework.context.annotation.ConfigurationClassPostProcessor.postProcessBeanDefinitionRegistry(ConfigurationClassPostProcessor.java:247)\n\t
at org.springframework.context.support.PostProcessorRegistrationDelegate.invokeBeanDefinitionRegistryPostProcessors(PostProcessorRegistrationDelegate.java:311)\n\t
at org.springframework.context.support.PostProcessorRegistrationDelegate.invokeBeanFactoryPostProcessors(PostProcessorRegistrationDelegate.java:112)\n\t
at org.springframework.context.support.AbstractApplicationContext.invokeBeanFactoryPostProcessors(AbstractApplicationContext.java:746)\n\t
at org.springframework.context.support.AbstractApplicationContext.refresh(AbstractApplicationContext.java:564)\n\t
at org.springframework.boot.web.servlet.context.ServletWebServerApplicationContext.refresh(ServletWebServerApplicationContext.java:147)\n\t
at org.springframework.boot.SpringApplication.refresh(SpringApplication.java:731)\n\t
at org.springframework.boot.SpringApplication.refreshContext(SpringApplication.java:408)\n\t
at org.springframework.boot.SpringApplication.run(SpringApplication.java:307)\n\t
at org.springframework.boot.SpringApplication.run(SpringApplication.java:1303)\n\t
at org.springframework.boot.SpringApplication.run(SpringApplication.java:1292)\n\t
at com.dt.rts.eregusa.discovery.Application.main(Application.java:12)\n\t
at java.base/jdk.internal.reflect.NativeMethodAccessorImpl.invoke0(Native Method)\n\t
at java.base/jdk.internal.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:77)\n\t
at java.base/jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)\n\t
at java.base/java.lang.reflect.Method.invoke(Method.java:568)\n\t
at org.springframework.boot.loader.MainMethodRunner.run(MainMethodRunner.java:49)\n\t
at org.springframework.boot.loader.Launcher.launch(Launcher.java:108)\n\t
at org.springframework.boot.loader.Launcher.launch(Launcher.java:58)\n\t
at org.springframework.boot.loader.JarLauncher.main(JarLauncher.java:65)\n
Caused by: java.lang.IllegalStateException: Failed to introspect Class [org.springframework.cloud.netflix.eureka.config.DiscoveryClientOptionalArgsConfiguration] from ClassLoader [org.springframework.boot.loader.LaunchedURLClassLoader#61af1510]\n\t
at org.springframework.util.ReflectionUtils.getDeclaredMethods(ReflectionUtils.java:485)\n\t
at org.springframework.util.ReflectionUtils.doWithMethods(ReflectionUtils.java:361)\n\t
at org.springframework.util.ReflectionUtils.getUniqueDeclaredMethods(ReflectionUtils.java:418)\n\t
at org.springframework.beans.factory.support.AbstractAutowireCapableBeanFactory.lambda$getTypeForFactoryMethod$2(AbstractAutowireCapableBeanFactory.java:765)\n\t
at java.base/java.util.concurrent.ConcurrentHashMap.computeIfAbsent(ConcurrentHashMap.java:1740)\n\t
at org.springframework.beans.factory.support.AbstractAutowireCapableBeanFactory.getTypeForFactoryMethod(AbstractAutowireCapableBeanFactory.java:764)\n\t
at org.springframework.beans.factory.support.AbstractAutowireCapableBeanFactory.determineTargetType(AbstractAutowireCapableBeanFactory.java:703)\n\t
at org.springframework.beans.factory.support.AbstractAutowireCapableBeanFactory.predictBeanType(AbstractAutowireCapableBeanFactory.java:674)\n\t
at org.springframework.beans.factory.support.AbstractBeanFactory.isFactoryBean(AbstractBeanFactory.java:1684)\n\t
at org.springframework.beans.factory.support.DefaultListableBeanFactory.doGetBeanNamesForType(DefaultListableBeanFactory.java:570)\n\t
at org.springframework.beans.factory.support.DefaultListableBeanFactory.getBeanNamesForType(DefaultListableBeanFactory.java:542)\n\t
at org.springframework.boot.autoconfigure.condition.OnBeanCondition.collectBeanNamesForType(OnBeanCondition.java:250)\n\t
at org.springframework.boot.autoconfigure.condition.OnBeanCondition.getBeanNamesForType(OnBeanCondition.java:243)\n\t
at org.springframework.boot.autoconfigure.condition.OnBeanCondition.getBeanNamesForType(OnBeanCondition.java:233)\n\t
at org.springframework.boot.autoconfigure.condition.OnBeanCondition.getMatchingBeans(OnBeanCondition.java:181)\n\t
at org.springframework.boot.autoconfigure.condition.OnBeanCondition.getMatchOutcome(OnBeanCondition.java:156)\n\t
at org.springframework.boot.autoconfigure.condition.SpringBootCondition.matches(SpringBootCondition.java:47)\n\t...
25 common frames omitted\n
Caused by: java.lang.NoClassDefFoundError: com/netflix/discovery/DiscoveryClient$DiscoveryClientOptionalArgs\n\t
at java.base/java.lang.ClassLoader.defineClass1(Native Method)\n\t
at java.base/java.lang.ClassLoader.defineClass(ClassLoader.java:1012)\n\t
at java.base/java.security.SecureClassLoader.defineClass(SecureClassLoader.java:150)\n\t
at java.base/java.net.URLClassLoader.defineClass(URLClassLoader.java:524)\n\t
at java.base/java.net.URLClassLoader$1.run(URLClassLoader.java:427)\n\t
at java.base/java.net.URLClassLoader$1.run(URLClassLoader.java:421)\n\t

OSGI two dependency chain issue when upgrading cxf jars

I got the following error in felix.log after adding javax.xml.soap:javax.xml.soap-api-1.4.0.jar in osgi/custo directory, how to understand the error message and what could be the possible fix?
WARN - 2022-10-17 14:36:22,821 - FelixDispatchQueue - Could not start bundle: file:custo/cxf-rt-frontend-jaxws-3.4.8.jar
org.osgi.framework.BundleException: Uses constraint violation. Unable to resolve resource org.apache.cxf.cxf-rt-frontend-jaxws [org.apache.cxf.cxf-rt-frontend-jaxws [231](R 231.0)] because it is exposed to package 'javax.xml.soap' from resources javax.xml.soap-api [javax.xml.soap-api [144](R 144.0)] and org.apache.felix.framework [org.apache.felix.framework [0](R 0)] via two dependency chains.
Chain 1:
org.apache.cxf.cxf-rt-frontend-jaxws [org.apache.cxf.cxf-rt-frontend-jaxws [231](R 231.0)]
import: (&(osgi.wiring.package=javax.xml.soap)(version>=1.4.0)(!(version>=2.0.0)))
|
export: osgi.wiring.package: javax.xml.soap
javax.xml.soap-api [javax.xml.soap-api [144](R 144.0)]
Chain 2:
org.apache.cxf.cxf-rt-frontend-jaxws [org.apache.cxf.cxf-rt-frontend-jaxws [231](R 231.0)]
import: (&(osgi.wiring.package=javax.xml.ws.handler.soap)(version>=0.0.0)(!(version>=3.0.0)))
|
export: osgi.wiring.package: javax.xml.ws.handler.soap; uses:=javax.xml.soap
export: osgi.wiring.package=javax.xml.soap
org.apache.felix.framework [org.apache.felix.framework [0](R 0)] Unresolved requirements: [[org.apache.cxf.cxf-rt-frontend-jaxws [231](R 231.0)] osgi.wiring.package; (&(osgi.wiring.package=javax.xml.ws.handler.soap)(version>=0.0.0)(!(version>=3.0.0)))]
at org.apache.felix.framework.Felix.resolveBundleRevision(Felix.java:4398)
at org.apache.felix.framework.Felix.startBundle(Felix.java:2308)
at org.apache.felix.framework.BundleImpl.start(BundleImpl.java:1006)
at com.alcatel.as.service.bundleinstaller.impl.BundleInstallerImpl$DeployedBundle.start(BundleInstallerImpl.java:935)
at com.alcatel.as.service.bundleinstaller.impl.BundleInstallerImpl.startBundles(BundleInstallerImpl.java:624)
at com.alcatel.as.service.bundleinstaller.impl.BundleInstallerImpl.finishInitialisation(BundleInstallerImpl.java:260)
at com.alcatel.as.service.bundleinstaller.impl.BundleInstallerImpl.frameworkEvent(BundleInstallerImpl.java:244)
at org.apache.felix.framework.EventDispatcher.invokeFrameworkListenerCallback(EventDispatcher.java:881)
at org.apache.felix.framework.EventDispatcher.fireEventImmediately(EventDispatcher.java:830)
at org.apache.felix.framework.EventDispatcher.run(EventDispatcher.java:1147)
at org.apache.felix.framework.EventDispatcher.access$000(EventDispatcher.java:54)
at org.apache.felix.framework.EventDispatcher$1.run(EventDispatcher.java:102)
at java.lang.Thread.run(Thread.java:750)
if I don't install javax.xml.soap-api-1.4.0.jar, I will get the following exception with cxf 3.4.8 package:
WARN - 2022-10-19 07:48:04,272 - FelixDispatchQueue - Could not start bundle: file:custo/cxf-rt-frontend-simple-3.4.8.jar
org.osgi.framework.BundleException: Unable to resolve org.apache.cxf.cxf-rt-frontend-simple [25](R 25.0): missing requirement [org.apache.cxf.cxf-rt-frontend-simple [25](R 25.0)] osgi.wiring.package; (&(osgi.wiring.package=org.apache.cxf.binding.soap.interceptor)(version>=3.4.0)(!(version>=4.0.0))) [caused by: Unable to resolve org.apache.cxf.cxf-rt-bindings-soap [112](R 112.0): missing requirement [org.apache.cxf.cxf-rt-bindings-soap [112](R 112.0)] osgi.wiring.package; (&(osgi.wiring.package=javax.xml.soap)(version>=1.4.0)(!(version>=2.0.0)))] Unresolved requirements: [[org.apache.cxf.cxf-rt-frontend-simple [25](R 25.0)] osgi.wiring.package; (&(osgi.wiring.package=org.apache.cxf.binding.soap.interceptor)(version>=3.4.0)(!(version>=4.0.0)))]
at org.apache.felix.framework.Felix.resolveBundleRevision(Felix.java:4398)
at org.apache.felix.framework.Felix.startBundle(Felix.java:2308)
at org.apache.felix.framework.BundleImpl.start(BundleImpl.java:1006)
at com.alcatel.as.service.bundleinstaller.impl.BundleInstallerImpl$DeployedBundle.start(BundleInstallerImpl.java:935)
at com.alcatel.as.service.bundleinstaller.impl.BundleInstallerImpl.startBundles(BundleInstallerImpl.java:624)
at com.alcatel.as.service.bundleinstaller.impl.BundleInstallerImpl.finishInitialisation(BundleInstallerImpl.java:260)
at com.alcatel.as.service.bundleinstaller.impl.BundleInstallerImpl.frameworkEvent(BundleInstallerImpl.java:244)
at org.apache.felix.framework.EventDispatcher.invokeFrameworkListenerCallback(EventDispatcher.java:881)
at org.apache.felix.framework.EventDispatcher.fireEventImmediately(EventDispatcher.java:830)
at org.apache.felix.framework.EventDispatcher.run(EventDispatcher.java:1147)
at org.apache.felix.framework.EventDispatcher.access$000(EventDispatcher.java:54)
at org.apache.felix.framework.EventDispatcher$1.run(EventDispatcher.java:102)
at java.lang.Thread.run(Thread.java:750)
WARN - 2022-10-19 07:48:04,881 - FelixDispatchQueue - Could not start bundle: file:custo/cxf-rt-frontend-jaxws-3.4.8.jar
org.osgi.framework.BundleException: Unable to resolve org.apache.cxf.cxf-rt-frontend-jaxws [84](R 84.0): missing requirement [org.apache.cxf.cxf-rt-frontend-jaxws [84](R 84.0)] osgi.wiring.package; (&(osgi.wiring.package=javax.xml.soap)(version>=1.4.0)(!(version>=2.0.0))) Unresolved requirements: [[org.apache.cxf.cxf-rt-frontend-jaxws [84](R 84.0)] osgi.wiring.package; (&(osgi.wiring.package=javax.xml.soap)(version>=1.4.0)(!(version>=2.0.0)))]
at org.apache.felix.framework.Felix.resolveBundleRevision(Felix.java:4398)
at org.apache.felix.framework.Felix.startBundle(Felix.java:2308)
at org.apache.felix.framework.BundleImpl.start(BundleImpl.java:1006)
at com.alcatel.as.service.bundleinstaller.impl.BundleInstallerImpl$DeployedBundle.start(BundleInstallerImpl.java:935)
at com.alcatel.as.service.bundleinstaller.impl.BundleInstallerImpl.startBundles(BundleInstallerImpl.java:624)
at com.alcatel.as.service.bundleinstaller.impl.BundleInstallerImpl.finishInitialisation(BundleInstallerImpl.java:260)
at com.alcatel.as.service.bundleinstaller.impl.BundleInstallerImpl.frameworkEvent(BundleInstallerImpl.java:244)
at org.apache.felix.framework.EventDispatcher.invokeFrameworkListenerCallback(EventDispatcher.java:881)
at org.apache.felix.framework.EventDispatcher.fireEventImmediately(EventDispatcher.java:830)
at org.apache.felix.framework.EventDispatcher.run(EventDispatcher.java:1147)
at org.apache.felix.framework.EventDispatcher.access$000(EventDispatcher.java:54)
at org.apache.felix.framework.EventDispatcher$1.run(EventDispatcher.java:102)
at java.lang.Thread.run(Thread.java:750)

Writing to cloud DB2 table using pyspark

I have just started my ETL journey using pyspark. My current goal is to write data from a .csv to a dashdb using append mode. However, I ran into an issue I can't seem to solve. Here's what I've done so far:
I've read the .csv and registered it into a temp table to be able to call SQL queries on it. The output of the query is the following:
+--------------------+--------+-------+-----+------+
| Street|District|Area_m2|Rooms| Price|
+--------------------+--------+-------+-----+------+
| Angyalföld| XIII| 105| 2|320000|
| Belváros| V| 70| 2|230000|
| Pozsonyi út| XIII| 89| 2|290000|
| Fecske utca| VIII| 33| 1|130000|
|Margó Tivadar utc...| XVIII| 80| 2|220000|
| Orczy út 46-48| VIII| 44| 2|120000|
| Vaskapu utca| IX| 51|1 + 1|185000|
| Gubacsi út 19| IX| 30| 1|105000|
| Öv utca 133| XIV| 29| 1|150000|
| Mérleg utca| V| 54| 2|190000|
| Szirtes út| I| 160| 4|389000|
| Gubacsi út 19| IX| 50| 2|130000|
| Török utca| II| 53|1 + 1|165000|
| Ferenc tér| IX| 65| 2|235000|
| Kiscelli utca| III| 34| 1|190000|
| Dózsa György út| VII| 47|1 + 1|130000|
| Vadász utca| V| 60|1 + 1|185000|
| István utca 7.| VII| 30| 1|120000|
| Regős utca| XI| 53| 2|180000|
| Országház utca| I| 122| 3|680000|
+--------------------+--------+-------+-----+------+
The columns in the db2 table have the following types: varchar(50), varchar(10), integer, varchar(10), integer.
I stored the output in a variable and used the following code to write to the dashdb:
output.write.jdbc(jdbc_url, table, properties = connection_properties, mode = 'append')
Running this code produces the following error message:
An error occurred while calling o310.jdbc.
: org.apache.spark.SparkException: Job aborted due to stage failure: Task 0 in stage 11.0 failed 1 times, most recent failure: Lost task 0.0 in stage 11.0 (TID 11, localhost, executor driver): com.ibm.db2.jcc.am.BatchUpdateException: [jcc][t4][102][10040][3.62.56] Batch failure. The batch was submitted, but at least one exception occurred on an individual member of the batch.
Use getNextException() to retrieve the exceptions for specific batched elements. ERRORCODE=-4229, SQLSTATE=null
at com.ibm.db2.jcc.am.fd.a(fd.java:404)
at com.ibm.db2.jcc.am.o.a(o.java:381)
at com.ibm.db2.jcc.am.kn.a(kn.java:4523)
at com.ibm.db2.jcc.am.kn.c(kn.java:4294)
at com.ibm.db2.jcc.am.kn.executeBatch(kn.java:2600)
at org.apache.spark.sql.execution.datasources.jdbc.JdbcUtils$.savePartition(JdbcUtils.scala:667)
at org.apache.spark.sql.execution.datasources.jdbc.JdbcUtils$$anonfun$saveTable$1.apply(JdbcUtils.scala:834)
at org.apache.spark.sql.execution.datasources.jdbc.JdbcUtils$$anonfun$saveTable$1.apply(JdbcUtils.scala:834)
at org.apache.spark.rdd.RDD$$anonfun$foreachPartition$1$$anonfun$apply$28.apply(RDD.scala:935)
at org.apache.spark.rdd.RDD$$anonfun$foreachPartition$1$$anonfun$apply$28.apply(RDD.scala:935)
at org.apache.spark.SparkContext$$anonfun$runJob$5.apply(SparkContext.scala:2101)
at org.apache.spark.SparkContext$$anonfun$runJob$5.apply(SparkContext.scala:2101)
at org.apache.spark.scheduler.ResultTask.runTask(ResultTask.scala:90)
at org.apache.spark.scheduler.Task.run(Task.scala:123)
at org.apache.spark.executor.Executor$TaskRunner$$anonfun$10.apply(Executor.scala:408)
at org.apache.spark.util.Utils$.tryWithSafeFinally(Utils.scala:1360)
at org.apache.spark.executor.Executor$TaskRunner.run(Executor.scala:414)
at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1149)
at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:624)
at java.lang.Thread.run(Thread.java:748)
Suppressed: com.ibm.db2.jcc.am.SqlDataException: Error for batch element #66: DB2 SQL Error: SQLCODE=-302, SQLSTATE=22001, SQLERRMC=null, DRIVER=3.62.56
at com.ibm.db2.jcc.am.fd.a(fd.java:668)
at com.ibm.db2.jcc.am.fd.a(fd.java:60)
at com.ibm.db2.jcc.am.fd.a(fd.java:127)
at com.ibm.db2.jcc.t4.cb.a(cb.java:481)
at com.ibm.db2.jcc.t4.cb.a(cb.java:70)
at com.ibm.db2.jcc.t4.q.a(q.java:57)
at com.ibm.db2.jcc.t4.sb.a(sb.java:225)
at com.ibm.db2.jcc.am.kn.a(kn.java:3083)
at com.ibm.db2.jcc.am.kn.d(kn.java:5019)
at com.ibm.db2.jcc.am.kn.a(kn.java:4466)
... 17 more
Driver stacktrace:
at org.apache.spark.scheduler.DAGScheduler.org$apache$spark$scheduler$DAGScheduler$$failJobAndIndependentStages(DAGScheduler.scala:1889)
at org.apache.spark.scheduler.DAGScheduler$$anonfun$abortStage$1.apply(DAGScheduler.scala:1877)
at org.apache.spark.scheduler.DAGScheduler$$anonfun$abortStage$1.apply(DAGScheduler.scala:1876)
at scala.collection.mutable.ResizableArray$class.foreach(ResizableArray.scala:59)
at scala.collection.mutable.ArrayBuffer.foreach(ArrayBuffer.scala:48)
at org.apache.spark.scheduler.DAGScheduler.abortStage(DAGScheduler.scala:1876)
at org.apache.spark.scheduler.DAGScheduler$$anonfun$handleTaskSetFailed$1.apply(DAGScheduler.scala:926)
at org.apache.spark.scheduler.DAGScheduler$$anonfun$handleTaskSetFailed$1.apply(DAGScheduler.scala:926)
at scala.Option.foreach(Option.scala:257)
at org.apache.spark.scheduler.DAGScheduler.handleTaskSetFailed(DAGScheduler.scala:926)
at org.apache.spark.scheduler.DAGSchedulerEventProcessLoop.doOnReceive(DAGScheduler.scala:2110)
at org.apache.spark.scheduler.DAGSchedulerEventProcessLoop.onReceive(DAGScheduler.scala:2059)
at org.apache.spark.scheduler.DAGSchedulerEventProcessLoop.onReceive(DAGScheduler.scala:2048)
at org.apache.spark.util.EventLoop$$anon$1.run(EventLoop.scala:49)
at org.apache.spark.scheduler.DAGScheduler.runJob(DAGScheduler.scala:737)
at org.apache.spark.SparkContext.runJob(SparkContext.scala:2061)
at org.apache.spark.SparkContext.runJob(SparkContext.scala:2082)
at org.apache.spark.SparkContext.runJob(SparkContext.scala:2101)
at org.apache.spark.SparkContext.runJob(SparkContext.scala:2126)
at org.apache.spark.rdd.RDD$$anonfun$foreachPartition$1.apply(RDD.scala:935)
at org.apache.spark.rdd.RDD$$anonfun$foreachPartition$1.apply(RDD.scala:933)
at org.apache.spark.rdd.RDDOperationScope$.withScope(RDDOperationScope.scala:151)
at org.apache.spark.rdd.RDDOperationScope$.withScope(RDDOperationScope.scala:112)
at org.apache.spark.rdd.RDD.withScope(RDD.scala:363)
at org.apache.spark.rdd.RDD.foreachPartition(RDD.scala:933)
at org.apache.spark.sql.execution.datasources.jdbc.JdbcUtils$.saveTable(JdbcUtils.scala:834)
at org.apache.spark.sql.execution.datasources.jdbc.JdbcRelationProvider.createRelation(JdbcRelationProvider.scala:68)
at org.apache.spark.sql.execution.datasources.SaveIntoDataSourceCommand.run(SaveIntoDataSourceCommand.scala:45)
at org.apache.spark.sql.execution.command.ExecutedCommandExec.sideEffectResult$lzycompute(commands.scala:70)
at org.apache.spark.sql.execution.command.ExecutedCommandExec.sideEffectResult(commands.scala:68)
at org.apache.spark.sql.execution.command.ExecutedCommandExec.doExecute(commands.scala:86)
at org.apache.spark.sql.execution.SparkPlan$$anonfun$execute$1.apply(SparkPlan.scala:131)
at org.apache.spark.sql.execution.SparkPlan$$anonfun$execute$1.apply(SparkPlan.scala:127)
at org.apache.spark.sql.execution.SparkPlan$$anonfun$executeQuery$1.apply(SparkPlan.scala:155)
at org.apache.spark.rdd.RDDOperationScope$.withScope(RDDOperationScope.scala:151)
at org.apache.spark.sql.execution.SparkPlan.executeQuery(SparkPlan.scala:152)
at org.apache.spark.sql.execution.SparkPlan.execute(SparkPlan.scala:127)
at org.apache.spark.sql.execution.QueryExecution.toRdd$lzycompute(QueryExecution.scala:80)
at org.apache.spark.sql.execution.QueryExecution.toRdd(QueryExecution.scala:80)
at org.apache.spark.sql.DataFrameWriter$$anonfun$runCommand$1.apply(DataFrameWriter.scala:676)
at org.apache.spark.sql.DataFrameWriter$$anonfun$runCommand$1.apply(DataFrameWriter.scala:676)
at org.apache.spark.sql.execution.SQLExecution$$anonfun$withNewExecutionId$1.apply(SQLExecution.scala:78)
at org.apache.spark.sql.execution.SQLExecution$.withSQLConfPropagated(SQLExecution.scala:125)
at org.apache.spark.sql.execution.SQLExecution$.withNewExecutionId(SQLExecution.scala:73)
at org.apache.spark.sql.DataFrameWriter.runCommand(DataFrameWriter.scala:676)
at org.apache.spark.sql.DataFrameWriter.saveToV1Source(DataFrameWriter.scala:285)
at org.apache.spark.sql.DataFrameWriter.save(DataFrameWriter.scala:271)
at org.apache.spark.sql.DataFrameWriter.jdbc(DataFrameWriter.scala:515)
at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62)
at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
at java.lang.reflect.Method.invoke(Method.java:498)
at py4j.reflection.MethodInvoker.invoke(MethodInvoker.java:244)
at py4j.reflection.ReflectionEngine.invoke(ReflectionEngine.java:357)
at py4j.Gateway.invoke(Gateway.java:282)
at py4j.commands.AbstractCommand.invokeMethod(AbstractCommand.java:132)
at py4j.commands.CallCommand.execute(CallCommand.java:79)
at py4j.GatewayConnection.run(GatewayConnection.java:238)
at java.lang.Thread.run(Thread.java:748)
Caused by: com.ibm.db2.jcc.am.BatchUpdateException: [jcc][t4][102][10040][3.62.56] Batch failure. The batch was submitted, but at least one exception occurred on an individual member of the batch.
Use getNextException() to retrieve the exceptions for specific batched elements. ERRORCODE=-4229, SQLSTATE=null
at com.ibm.db2.jcc.am.fd.a(fd.java:404)
at com.ibm.db2.jcc.am.o.a(o.java:381)
at com.ibm.db2.jcc.am.kn.a(kn.java:4523)
at com.ibm.db2.jcc.am.kn.c(kn.java:4294)
at com.ibm.db2.jcc.am.kn.executeBatch(kn.java:2600)
at org.apache.spark.sql.execution.datasources.jdbc.JdbcUtils$.savePartition(JdbcUtils.scala:667)
at org.apache.spark.sql.execution.datasources.jdbc.JdbcUtils$$anonfun$saveTable$1.apply(JdbcUtils.scala:834)
at org.apache.spark.sql.execution.datasources.jdbc.JdbcUtils$$anonfun$saveTable$1.apply(JdbcUtils.scala:834)
at org.apache.spark.rdd.RDD$$anonfun$foreachPartition$1$$anonfun$apply$28.apply(RDD.scala:935)
at org.apache.spark.rdd.RDD$$anonfun$foreachPartition$1$$anonfun$apply$28.apply(RDD.scala:935)
at org.apache.spark.SparkContext$$anonfun$runJob$5.apply(SparkContext.scala:2101)
at org.apache.spark.SparkContext$$anonfun$runJob$5.apply(SparkContext.scala:2101)
at org.apache.spark.scheduler.ResultTask.runTask(ResultTask.scala:90)
at org.apache.spark.scheduler.Task.run(Task.scala:123)
at org.apache.spark.executor.Executor$TaskRunner$$anonfun$10.apply(Executor.scala:408)
at org.apache.spark.util.Utils$.tryWithSafeFinally(Utils.scala:1360)
at org.apache.spark.executor.Executor$TaskRunner.run(Executor.scala:414)
at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1149)
at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:624)
... 1 more
Suppressed: com.ibm.db2.jcc.am.SqlDataException: Error for batch element #66: DB2 SQL Error: SQLCODE=-302, SQLSTATE=22001, SQLERRMC=null, DRIVER=3.62.56
at com.ibm.db2.jcc.am.fd.a(fd.java:668)
at com.ibm.db2.jcc.am.fd.a(fd.java:60)
at com.ibm.db2.jcc.am.fd.a(fd.java:127)
at com.ibm.db2.jcc.t4.cb.a(cb.java:481)
at com.ibm.db2.jcc.t4.cb.a(cb.java:70)
at com.ibm.db2.jcc.t4.q.a(q.java:57)
at com.ibm.db2.jcc.t4.sb.a(sb.java:225)
at com.ibm.db2.jcc.am.kn.a(kn.java:3083)
at com.ibm.db2.jcc.am.kn.d(kn.java:5019)
at com.ibm.db2.jcc.am.kn.a(kn.java:4466)
... 17 more
Can someone please explain what went wrong and how can the issue be solved?
Thank you in advance!
The issue can be solved by ensuring utf-8 encoding for the .csv and by removing special characters.

How to get the jms quickstart packaged with JBoss-Fuse-6.1.0.redhat-355 to work?

The Karaf output is as under :
JBossFuse:karaf#root> features:addurl mvn:org.jboss.quickstarts.fuse/jms/6.1.0.redhat-355/xml/features
JBossFuse:karaf#root> features:install quickstart-jms
[Fatal Error] :6:3: The element type "hr" must be terminated by the matching end-tag "</hr>".
Error executing command: The element type "hr" must be terminated by the matching end-tag "</hr>".
I am not sure where this html element is being used.
log:display-exceptionreveals :
2016-02-12 07:36:24,571 | INFO | l Console Thread | Console | ?
? | 17 - org.apache.karaf.shell.console - 2.3.0.redhat-610355 | Exception caught while executing command
org.xml.sax.SAXParseException: The element type "hr" must be terminated by the matching end-tag "</hr>".
JBossFuse:karaf#root> log:display-exception
org.xml.sax.SAXParseException: The element type "hr" must be terminated by the matching end-tag "</hr>".
at org.apache.xerces.parsers.DOMParser.parse(Unknown Source)[:]
at org.apache.xerces.jaxp.DocumentBuilderImpl.parse(Unknown Source)[:]
at javax.xml.parsers.DocumentBuilder.parse(Unknown Source)[:2.3.0.redhat-610355]
at org.apache.karaf.features.internal.FeatureValidationUtil.validate(FeatureValidationUtil.java:
52)[23:org.apache.karaf.features.core:2.3.0.redhat-610355]
at org.apache.karaf.features.internal.FeaturesServiceImpl.validateRepository(FeaturesServiceImpl
.java:215)[23:org.apache.karaf.features.core:2.3.0.redhat-610355]
at org.apache.karaf.features.internal.FeaturesServiceImpl.internalAddRepository(FeaturesServiceImpl.java:256)[23:org.apache.karaf.features.core:2.3.0.redhat-610355]
at org.apache.karaf.features.internal.FeaturesServiceImpl.getFeatures(FeaturesServiceImpl.java:1
015)[23:org.apache.karaf.features.core:2.3.0.redhat-610355]
at org.apache.karaf.features.internal.FeaturesServiceImpl.getFeature(FeaturesServiceImpl.java:97
3)[23:org.apache.karaf.features.core:2.3.0.redhat-610355]
at org.apache.karaf.features.internal.FeaturesServiceImpl.installFeature(FeaturesServiceImpl.jav
a:395)[23:org.apache.karaf.features.core:2.3.0.redhat-610355]
at org.apache.karaf.features.command.InstallFeatureCommand.doExecute(InstallFeatureCommand.java:
62)[37:org.apache.karaf.features.command:2.3.0.redhat-610355]
at org.apache.karaf.features.command.FeaturesCommandSupport.doExecute(FeaturesCommandSupport.jav
a:41)[37:org.apache.karaf.features.command:2.3.0.redhat-610355]
at org.apache.karaf.shell.console.OsgiCommandSupport.execute(OsgiCommandSupport.java:39)[17:org.
apache.karaf.shell.console:2.3.0.redhat-610355]
at org.apache.felix.gogo.commands.basic.AbstractCommand.execute(AbstractCommand.java:35)[17:org.
apache.karaf.shell.console:2.3.0.redhat-610355]
at org.apache.felix.gogo.runtime.CommandProxy.execute(CommandProxy.java:78)[18:org.apache.felix.gogo.runtime:0.11.0.redhat-610355] at org.apache.felix.gogo.runtime.Closure.executeCmd(Closure.java:477)[18:org.apache.felix.gogo.runtime:0.11.0.redhat-610355]
at org.apache.felix.gogo.runtime.Closure.executeStatement(Closure.java:403)[18:org.apache.felix.gogo.runtime:0.11.0.redhat-610355]
at org.apache.felix.gogo.runtime.Pipe.run(Pipe.java:108)[18:org.apache.felix.gogo.runtime:0.11.0.redhat-610355]
at org.apache.felix.gogo.runtime.Closure.execute(Closure.java:183)[18:org.apache.felix.gogo.runtime:0.11.0.redhat-610355]
at org.apache.felix.gogo.runtime.Closure.execute(Closure.java:120)[18:org.apache.felix.gogo.runtime:0.11.0.redhat-610355] at org.apache.felix.gogo.runtime.CommandSessionImpl.execute(CommandSessionImpl.java:89)[18:org.apache.felix.gogo.runtime:0.11.0.redhat-610355]
at org.apache.karaf.shell.console.jline.Console.run(Console.java:189)[17:org.apache.karaf.shell.console:2.3.0.redhat-610355]
at org.apache.karaf.shell.console.jline.DelayedStarted.run(DelayedStarted.java:61)[17:org.apache.karaf.shell.console:2.3.0.redhat-610355]
JBossFuse:karaf#root>
If anyone can help me get this working it will help.
TIA
I got most of the quickstarts (incl the JMS one) to work on this version -> jboss-fuse-6.0.0.redhat-024. My JDK was 1.7.0_60 and was set to JAVA_HOME.

OSGI. Waiting for timeout during installation of bundle (tinybundles)

I have a problem running pax exam test under karaf.
During an installation of bundle
2015-05-20 14:16:42,644 | DEBUG | FelixStartLevel | BundleManager | 1 - org.apache.karaf.features.core - 3.0.0 | Checking file:/tmp/ops4j-store-anonymous-568188393605841387/tinybundles_c3beb06a2df3da75e08a1ae1a44c832cfbd034c0.bin
it hangs for some time in
org.apache.karaf.features.internal.BundleManager#waitForUrlHandler
looking for a
(&(objectClass=org.osgi.service.url.URLStreamHandlerService)(url.handler.protocol=file))
here is the thread dump:
at java.lang.Object.wait(Object.java:-1)
at org.osgi.util.tracker.ServiceTracker.waitForService(ServiceTracker.java:499)
at org.apache.karaf.features.internal.BundleManager.waitForUrlHandler(BundleManager.java:229)
at org.apache.karaf.features.internal.BundleManager.getInputStreamForBundle(BundleManager.java:186)
at org.apache.karaf.features.internal.BundleManager.doInstallBundleIfNeeded(BundleManager.java:99)
at org.apache.karaf.features.internal.BundleManager.installBundleIfNeeded(BundleManager.java:90)
at org.apache.karaf.features.internal.FeaturesServiceImpl.doInstallFeature(FeaturesServiceImpl.java:523)
at org.apache.karaf.features.internal.FeaturesServiceImpl.installFeatures(FeaturesServiceImpl.java:383)
at org.apache.karaf.features.internal.BootFeaturesInstaller.installBootFeatures(BootFeaturesInstaller.java:92)
at org.apache.karaf.features.internal.BootFeaturesInstaller.start(BootFeaturesInstaller.java:76)
at sun.reflect.NativeMethodAccessorImpl.invoke0(NativeMethodAccessorImpl.java:-1)
at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:57)
at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
at java.lang.reflect.Method.invoke(Method.java:606)
at org.apache.aries.blueprint.utils.ReflectionUtils.invoke(ReflectionUtils.java:297)
at org.apache.aries.blueprint.container.BeanRecipe.invoke(BeanRecipe.java:958)
at org.apache.aries.blueprint.container.BeanRecipe.runBeanProcInit(BeanRecipe.java:712)
at org.apache.aries.blueprint.container.BeanRecipe.internalCreate2(BeanRecipe.java:824)
at org.apache.aries.blueprint.container.BeanRecipe.internalCreate(BeanRecipe.java:787)
at org.apache.aries.blueprint.di.AbstractRecipe$1.call(AbstractRecipe.java:79)
at java.util.concurrent.FutureTask.run(FutureTask.java:262)
at org.apache.aries.blueprint.di.AbstractRecipe.create(AbstractRecipe.java:88)
at org.apache.aries.blueprint.container.BlueprintRepository.createInstances(BlueprintRepository.java:245)
at org.apache.aries.blueprint.container.BlueprintRepository.createAll(BlueprintRepository.java:183)
at org.apache.aries.blueprint.container.BlueprintContainerImpl.instantiateEagerComponents(BlueprintContainerImpl.java:681)
at org.apache.aries.blueprint.container.BlueprintContainerImpl.doRun(BlueprintContainerImpl.java:378)
at org.apache.aries.blueprint.container.BlueprintContainerImpl.run(BlueprintContainerImpl.java:269)
- locked <0xbb5> (a java.util.concurrent.atomic.AtomicBoolean)
at org.apache.aries.blueprint.container.BlueprintExtender.createContainer(BlueprintExtender.java:276)
at org.apache.aries.blueprint.container.BlueprintExtender.createContainer(BlueprintExtender.java:245)
at org.apache.aries.blueprint.container.BlueprintExtender.modifiedBundle(BlueprintExtender.java:235)
at org.apache.aries.util.tracker.hook.BundleHookBundleTracker$Tracked.customizerModified(BundleHookBundleTracker.java:500)
at org.apache.aries.util.tracker.hook.BundleHookBundleTracker$Tracked.customizerModified(BundleHookBundleTracker.java:433)
at org.apache.aries.util.tracker.hook.BundleHookBundleTracker$AbstractTracked.track(BundleHookBundleTracker.java:725)
at org.apache.aries.util.tracker.hook.BundleHookBundleTracker$Tracked.bundleChanged(BundleHookBundleTracker.java:463)
at org.apache.aries.util.tracker.hook.BundleHookBundleTracker$BundleEventHook.event(BundleHookBundleTracker.java:422)
at org.apache.felix.framework.util.SecureAction.invokeBundleEventHook(SecureAction.java:1103)
at org.apache.felix.framework.util.EventDispatcher.createWhitelistFromHooks(EventDispatcher.java:695)
at org.apache.felix.framework.util.EventDispatcher.fireBundleEvent(EventDispatcher.java:483)
at org.apache.felix.framework.Felix.fireBundleEvent(Felix.java:4403)
at org.apache.felix.framework.Felix.startBundle(Felix.java:2092)
at org.apache.felix.framework.Felix.setActiveStartLevel(Felix.java:1291)
at org.apache.felix.framework.FrameworkStartLevelImpl.run(FrameworkStartLevelImpl.java:304)
at java.lang.Thread.run(Thread.java:745)
I didn't find any bundle that provides that service => so my test hangs for 30sec each time installing tinybundle's folder.
I've started a paxexam's karaf manually from the target folder and I ran service:list to see what bundles can provide URLStreamHandlerService service but there wasn't any with url.handler.protocol=file
service.id = 2
Provided by :
System Bundle (0)
Used by:
Apache Felix File Install (17)
Apache Aries JMX Core (225)
[org.osgi.service.url.URLStreamHandlerService]
service.id = 51
url.handler.protocol = feature
Provided by :
Apache Karaf :: Deployer :: Features (12)
[org.osgi.service.url.URLStreamHandlerService]
service.id = 15
url.handler.protocol = jardir
Provided by :
Apache Felix File Install (17)
[org.osgi.service.url.URLStreamHandlerService]
service.id = 42
url.handler.protocol = spring
Provided by :
Apache Karaf :: Deployer :: Spring (27)
[org.osgi.service.url.URLStreamHandlerService]
service.id = 45
url.handler.protocol = blueprint
Provided by :
Apache Karaf :: Deployer :: Blueprint (30)
[org.osgi.service.url.URLStreamHandlerService]
service.id = 4
url.handler.protocol = [wrap]
Provided by :
OPS4J Pax Url - wrap: (37)
[org.osgi.service.url.URLStreamHandlerService]
So the question is: how can I avoid waiting for a timeout. Do we really need URLStreamHandlerService service, if yes how can I provide it ?

Resources