After executing pip install elasticsearch-dbapi, when I am trying to connect Superset to Elasticsearch, it gives the error:
ERROR: {"error": "Connection failed!\n\nThe error message returned was:\nCan't load plugin: sqlalchemy.dialects:elasticsearch.http", "stacktrace": null}
Not sure the reason for the error.
I had the same issue post the installation of elasticsearch-dbapi. Then I have restarted the instance, its started working fine.
Related
I am trying to install cloudflow 2.0.25 version in the eks cluster using helm. But the pod goes to CrashLoopBackOff status with below error:
ERROR [ActorSystemImpl] - Unexpected error starting cloudflow operator, terminating.
io.fabric8.kubernetes.client.KubernetesClientException: Failure executing: POST at: https://172.20.0.1/apis/apiextensions.k8s.io/v1beta1/customresourcedefinitions. Message: the server could not find the requested resource. Received status: Status(apiVersion=v1, code=404, details=StatusDetails(causes=[], group=null, kind=null, name=null, retryAfterSeconds=null, uid=null, additionalProperties={}), kind=Status, message=the server could not find the requested resource, metadata=ListMeta(_continue=null, remainingItemCount=null, resourceVersion=null, selfLink=null, additionalProperties={}), reason=NotFound, status=Failure, additionalProperties={}).
at io.fabric8.kubernetes.client.dsl.base.OperationSupport.requestFailure(OperationSupport.java:570)
at io.fabric8.kubernetes.client.dsl.base.OperationSupport.assertResponseCode(OperationSupport.java:509)
at io.fabric8.kubernetes.client.dsl.base.OperationSupport.handleResponse(OperationSupport.java:474)
at io.fabric8.kubernetes.client.dsl.base.OperationSupport.handleResponse(OperationSupport.java:435)
at io.fabric8.kubernetes.client.dsl.base.OperationSupport.handleCreate(OperationSupport.java:250)
at io.fabric8.kubernetes.client.dsl.base.BaseOperation.handleCreate(BaseOperation.java:871)
at io.fabric8.kubernetes.client.dsl.base.BaseOperation.create(BaseOperation.java:366)
at io.fabric8.kubernetes.client.dsl.base.BaseOperation.create(BaseOperation.java:85)
at cloudflow.operator.Main$.checkCRD(Main.scala:140)
at cloudflow.operator.Main$.main(Main.scala:61)
at cloudflow.operator.Main.main(Main.scala)
Please help how to resolve the issue ?
I tried creating crd and then also created custom_roles and rolebindings but it didn't work.
My python version is 3.7, and after I ran pip3 install happybase, I started the command hbase thrift start and tried to write a brief .py file as following:
import happybase
connection = happybase.Connection('master')
table = connection.table('jmlr') #'jmlr' is a table in hbase
for i in table.scan():
print(i)
table.put('001', {'title':'dasds'}) #error here
connection.close()
When it's about to run table.put(), it reported such an error:
thriftpy2.transport.base.TTransportException: TTransportException(type=4, message='TSocket read 0 bytes')
And at the same time, the thrift reported an error:
ERROR [thrift-worker-1] thrift.TBoundedThreadPoolServer: Error occurred during processing of message. java.lang.IllegalArgumentException: Invalid famAndQf provided.
But just now I ran this python file again, it gave me a different error in thrift:
thrift.TBoundedThreadPoolServer: Thrift error occurred during processing of message.
org.apache.thrift.protocol.TProtocolException: Bad version in readMessageBegin
I have tried to add parameters like protocol='compact', transport='framed', but this didn't work, even the table.scan() failed.
Everything in the hbase shell is OK, so I can't figure out what went wrong, I'm about to collapse.
I ran into the same issue and found this sollution. You need to add even empty Column Qualifier ( ':' symbol as delimiter between Column Family and Column Qualifier) into put() method:
table.put('001:', {'title':'dasds'})
Also, you have a different error message after second run of script because thrift server is already failed.
I hope it will help you.
I am having problems trying to deploy the rest server with persistence storage (mongo db) I ge this message when deploying the rest server:
docker logs -f rest
[2018-03-12 00:01:13] PM2 log: Launching in no daemon mode
[2018-03-12 00:01:14] PM2 log: Starting execution sequence in -fork mode- for app name:composer-rest-server id:0
[2018-03-12 00:01:14] PM2 log: App name:composer-rest-server id:0 online
WARNING: NODE_APP_INSTANCE value of '0' did not match any instance config file names.
WARNING: See https://github.com/lorenwest/node-config/wiki/Strict-Mode
Discovering types from business network definition ...
Connection fails: Error: Error trying to ping. Error: Error trying to query business network. Error: REQUEST_TIMEOUT
It will be retried for the next request.
Exception: Error: Error trying to ping. Error: Error trying to query business network. Error: REQUEST_TIMEOUT
Error: Error trying to ping. Error: Error trying to query business network. Error: REQUEST_TIMEOUT
at _checkRuntimeVersions.then.catch (/home/composer/.npm-global/lib/node_modules/composer-rest-server/node_modules/composer-connector-hlfv1/lib/hlfconnection.js:699:34)
at
at process._tickDomainCallback (internal/process/next_tick.js:228:7)
[2018-03-12 00:07:03] PM2 log: App [composer-rest-server] with id [0] and pid [14], exited with code [1] via signal [SIGINT]
I am using composer v.18
is there any work around to fix this error. when I run the composer-rest- server alone it work. but if I try to add the mongodb it does not work
I am trying to install Graph-Aided Search to integrate Neo4j with ElasticSearch (2.3.1) as shown here. But when I try this command line:
plugin install com.graphaware.es/graph-aided-search/2.3.2.0
I get errors:
plugin install com.graphaware.es/graph-aided-search/2.3.2.0
-> Installing com.graphaware.es/graph-aided-search/2.3.2.0...
Trying https://download.elastic.co/com.graphaware.es/graph-aided-search/graph-aided-search-2.3.2.0.zip ...
Trying https://search.maven.org/remotecontent?filepath=com/graphaware/es/graph-aided-search/2.3.2.0/graph-aided-search-2.3.2.0.zip ...
Trying https://oss.sonatype.org/service/local/repositories/releases/content/com/graphaware/es/graph-aided-search/2.3.2.0/graph-aided-search-2.3.2.0.zip ...
Trying https://github.com/com.graphaware.es/graph-aided-search/archive/2.3.2.0.zip ...
Trying https://github.com/com.graphaware.es/graph-aided-search/archive/master.zip ...
ERROR: failed to download out of all possible locations..., use --verbose to get detailed information
And this when I add --verbose for more details:
plugin install com.graphaware.es/graph-aided-search/2.3.2.0 --verbose
-> Installing com.graphaware.es/graph-aided-search/2.3.2.0...
Trying https://download.elastic.co/com.graphaware.es/graph-aided-search/graph-aided-search-2.3.2.0.zip ...
Failed: SocketTimeoutException[connect timed out]
Trying https://search.maven.org/remotecontent?filepath=com/graphaware/es/graph-aided-search/2.3.2.0/graph-aided-search-2.3.2.0.zip ...
Failed: SocketTimeoutException[connect timed out]
Trying https://oss.sonatype.org/service/local/repositories/releases/content/com/graphaware/es/graph-aided-search/2.3.2.0/graph-aided-search-2.3.2.0.zip ...
Failed: SocketTimeoutException[connect timed out]
Trying https://github.com/com.graphaware.es/graph-aided-search/archive/2.3.2.0.zip ...
Failed: SocketTimeoutException[connect timed out]
Trying https://github.com/com.graphaware.es/graph-aided-search/archive/master.zip ...
Failed: SocketTimeoutException[connect timed out]
ERROR: failed to download out of all possible locations..., use --verbose to get detailed information
I looked for that error on google and I found that it might be because of the proxy, and since I am working behind a proxy I tried something that basically worked for others. This is the command line that I tried to install the plugin through proxy:
C:\dev\elasticsearch-2.3.1\bin> plugin -Dhttps.proxyHost=http://example.test.fr -Dhttps.proxyPort=3128 -Dhttps.proxyUser=SomeUser -Dhttps.proxyPassword=Password install com.graphaware.es/graph-aided-search/2.3.2.0
But still it didn't work, and I am still getting the same error. I forced the authentification to the proxy on Internet Explorer, something that usually works for me when I install packages with Maven. However, in that case it didn't work.
I am struggling to integrate my Neo4j database with ElasticSearch, I have been trying that for 4days now and it is taking all my time and cannot work without making this integration with success.
I really appreciate any help or clarification to resolve this error. Thank you.
[UPDATE]
When I try to replace the path, it is still not working. I've tried this command line:
plugin install C:\dev\graph-aided-search-master\target\releases\graph-aided-search-2.4.1.4-SNAPSHOT.zip
and I got this error message:
-> Installing C:\dev\graph-aided-search-master\target\releases\graph-aided-search-2.4.1.4-SNAPSHOT.zip...
ERROR: Invalid prefix or suffix
I put the graph-aided-search Zip file in the same folder as plugin file, so in C:\dev\elasticsearch-2.3.1\bin. Then I tried with the relative path, and I got this message:
> plugin install graph-aided-search-2.4.1.4-SNAPSHOT.zip --verbose
-> Installing graph-aided-search-2.4.1.4-SNAPSHOT.zip...
Trying https://download.elastic.co/elasticsearch/release/org/elasticsearch/plugin/graph-aided-search-2.4.1.4-SNAPSHOT.zip/2.3.1/graph-aided-search-2.4.1.4-SNAPSHOT.zip-2.3.1.zip ...
Failed: SocketTimeoutException[connect timed out]
ERROR: failed to download out of all possible locations..., use --verbose to get detailed information
You can download the plugin manually from the MavenRepository
https://mvnrepository.com/artifact/com.graphaware.es/graph-aided-search
The correct syntax is this one:
plugin install com.graphaware.es/graph-aided-search/2.3.2.0 -DproxyHost=exampleHost -DproxyPort=portNumber
It is working now!
when I try to connect to database from admintools I am getting following error:
Error: Unable to connect to database
Hint: Username or password could be invalid
I have found in the logs following error:
Apr 20 08:08:29 [24291] [vsql.connect spawn] Exception: Error! pty.fork() failed: out of pty devices
Do you know what is the problem?
Your node might be down
Check logs at
/opt/vertica/log
or at
/opt/vertica/config/admintools.conf
check restart policy section is right
[Database:mydb] host = 11.11.11.11
restartpolicy = ksafe