Hadoop Hue: Counter results greyed out - hadoop

I have created a Solr dashboard with analytics enabled. The Counter button is now shown, but its greyed out. Is there any additional flag in Hue or solr which I should enable?
Thanks
I am using Solr 7 + Hue 4

This is a bug https://issues.cloudera.org/browse/HUE-7337 that will be fixed soon. A workaround is to add a numeric field in your collection.

Related

Getting error when we are trying to download more than 2500 rows from obiee12c analytics

We are getting the following error message when we are trying to download data more than 2500 rows from the obiee12c analytics.
There was an error processing your download.
Please check with your administrator
OBIEE version: 12.2.0.4
To resolve this issue, InputStreamLimitInKB setting in OBIEE will need to be increased.
To increase the InputStreamLimitInKB limit:
Browse to $OBIEE_Domain/config/fmwconfig/biconfig/OBIJH
Open the config.xml
Find the following lines:
<XMLP>
<InputStreamLimitInKB>8192</InputStreamLimitInKB>
<ReadRequestBeforeProcessing>true</ReadRequestBeforeProcessing>
</XMLP>
Change the InputStreamLimitInKB value from 8192 to 1024000.
Save and close the file.
Then restart the OBIEE12c services.
The download of rows is controlled by several different settings in addition to the one mentioned above in the other response. Way too many to list in the post, but they are shown and explained in detail here:
https://docs.oracle.com/middleware/1221/biee/BIESG/answersconfigset.htm#BIESG3278
Those settings go hand in hand with increasing the InputStreamLimitInKB.

how to use tilemap in kibana with key & token

Map visualizations not working anymore in kibana. Looks like since 11 july the map visualizations are not working anymore. Only error messages are displayed. This is caused by a change of rules by MapQuest.
direct tile access of mapQuest has been discontinued.
so how can i use mapQuest in Kibana visualizations Chart with API Key & Token
attaching here snapshot of that error.
You can't.
Just wait for update from elasticsearch team
You can follow here
The solution is now released here

Setting default index in Kibana 4

I can't define default index in Kibana 4.0.2, as you can see on the image below. It doesn't saves if I enter it manually in "Advanced" tab, it only glitches when I click on "Set as default index" button but it doesn't make needed changes.
I'm using Couchbase 3.0.3, transport-couchbase plugin 2.0, ElasticSearch 1.5.2.
I've tried reinstalling everything (except couchbase), one by one, always the same.
I've googled it whole day, nothing. Someone has any idea?
it seems your not the only one having this issue with couchbase => https://github.com/elastic/kibana/issues/3331#issuecomment-84942136
After changing couchbase_template.json it's needed to set
curl -XPUT http://localhost:9200/_template/couchbase -d #/usr/share/elasticsearch/plugins/transport-couchbase/couchbase_template.json
and now it works.
Thnx juliendangers for directions.

Kibana 4 'Discover' search error

I indexed a dataset of geo-data records in ElasticSearch for analysis in Kibana. My issue is that the 'Discover' tab doesn't pick up the data but instead displays the error message
Discover: An error occurred with your request. Reset your inputs and try again.
In 'Settings', I could configure my data index just fine, and Kibana is picking up all the mapping fields with correct type/analysis/indexing metadata. 'Visualize' works fine, too. I can create my charts, add them to the dashboard, drill down - everything. Just the 'Discover' tab is broken for me.
I'm running ElasticSearch 1.5.2, and tried with Kibana 4.0.1, 4.0.2 and 4.1-snapshot now (on Ubuntu 14.04), all with the same results.
Another effect I'm noticing: the sidebar is not showing any 'Available Fields'. Only if I unfold the field settings and untick 'Hide Missing Fields' I'll get my list of schema fields. (These are greyed out as they are considered 'missing' by Kibana. But interestingly, clicking on 'Visualize' on one of them to chart their distribution works, again, perfectly fine.)
My only suspicion is: my data doesn't have a timestamp field, so maybe that's what's messing things up. Although judging from the docs I'd assume that non-timeseries data should be supported.
Any hints appreciated!
In my case, the cause was that I had indexed malformed JSON into elasticsearch. It was valid Javascript, but not valid JSON. In particular I neglected to quote the keys in the objects
I had inserted my (test) data using curl, e.g.
curl -X PUT http://localhost:9200/foo/doc/1 -d '{ts: "2015-06-24T01:07:00.000Z", employeeId: 105, action: "PICK", quantity: 8}'
Note that ts: should have been "ts":, etc.
Seems like elasticsearch tolerates such things, but Kibana does not. Once I fixed that, Discover worked fine.
Note that the error you are seeing is generated client side when an error arises. If you open your client debugger (e.g. Firefox) you will see the error in the console log. In my case, the error message was
Error: Unable to parse/serialize body
If your error is different, it will be a different cause.
It was my fault for entering bad JSON to begin with. Odd that elasticsearch is more tolerant than Kibana.
It happened to me as well. I tried all...:
Deleting all the indices (.kibana, my own, etc) didn't work
Restarting the ES, Kibana and LS services didn't help.
I didn't have the Request Timeout problem in kibana.yml either.
My problem was that the timestamp field was using an incorrect time format. I changed it to this format and it worked: "date": "2015-05-13T00:00:00"
I had the same problem. None of the suggested solutions helped. I finally found the problem while comparing a working version with a non-working version in Wireshark.
Don't emit a UTF8 byte order mark in front of your JSON. Somehow, my serializer was set up to do that... ElasticSearch is fine with it, but Kibana cannot handle it on the Discover page.

Kibana dashboard - error saving to ElasticSearch

I have a logstash-elasticsearch-kibana local setup and I have a problem when it comes to save Kibana dashboards.
Selecting the "Save" option I get the following error: "Save failed Dashboard could not be saved to Elasicsearch"
I'm using the logstash dashboard that comes with Kibana and after some modifications I tried to save it getting this error.
As far as I understand dashboards loaded from templates (json files located in kibana3/app/dashboards) cannot be saved to Elasticsearch (as stated in kibana templates). But I haven't been able to figure out how to create a new dashboard for logstash and save it to Elasticsearch, nor find instructions to do that. I would like to have different dashboards and be able to modify them and load them as needed.
I have exported the dashboard schema and successfully load it back, which works as far as saving a dashboard after all customization is done. But I would prefer to save them to elasticsearch rather than to template files.
Communication between ES and Kibana works fine (no errors show up in logs and information is retrieved and showed in Kibana).
Someone who could tell me what I'm missing here?
Thanks!
I got the error when I had a '/' (slash) in the name of the dashboard. Changing this to '-' solved the problem. See the following issue on GitHub: https://github.com/elasticsearch/kibana/issues/837

Resources