I tried to set-up a elasticsearch on my windows 7 os pc. Installed elasticsearch & sense and it's working as the loacahost:9200 is working fine.
Now i am strugging to search in a file located at c:\user\rajesh\default.json.
indexing of a data i.e
Put test\te\2
{
---datas
}
is working fine, but when i try to reference to file i.e POST test\te\2-d #default.json it's giving error as Unexpected '<'
I Installed KIBANA but not able to do anything with it's UI. when trying to search anything it's giving error:No index found at http://localhost:9200/INDEX_MISSING/_mapping/field/
I have edited config.js file with elasticsearch: "http://"+"localhost"+":9200", but it is not able to use the index which i created using sense to ES.
Thanks in Advance
First try this:
http://localhost:9200/_search?pretty
If you get no data returned (no indices), then you have some error either in elistcsearch or logstash.
Additionally, I recomend you try to access kopf from your browser.
http://lmenezes.com/elasticsearch-kopf/?location=http://localhost:9200
If you have data in the port, it will show you the indexes.
Regards
Ari
Related
When I use the elastic api, via curl or the developer tool. I cannot got data from some old indice.
But if I use the discover feature of Kibana, it can get those data.
Moreover, if I go from discover, got some datas, the with the inspect option I click on "Open in dev tools", the the copied request don't work.
Have you any idea?
I also try to get /_cout api, it says count:0. But with /_stats I have some counts
it works
not working
I'm using 7.8.1 version
I'm trying to upload csv data
I hit the link of Upload data from log file Import a CSV, NDJSON, or log file
And it takes me to /app/ml#/filedatavisualizer and shows the following error
{"statusCode":404,"error":"Not Found","message":"Not Found"}
There's discussion, at https://github.com/elastic/kibana/pull/24260 , they say it's been resolved, but it's not..
I found the answer,
you need to add the machine_learning_user or machine_learning_admin role to the user.
https://discuss.elastic.co/t/filedatavisualizer-is-only-reachable-when-using-a-user-assigned-to-the-original-super-admin-role-like-elastic/221270/2
I am facing a weird issue with elasticsearch.
I am working on a Laravel project and need to integrate elastic search with it.
I am using https://github.com/babenkoivan/scout-elasticsearch-driver along with Laravel Scout.
Initially everything worked fine and I was also able to search. When it worked fine, I wanted to drop my index and reindex everything from scratch. But after that when I ran my seeder, I got an "Index not found exception". Then to test it I created index and type. eg: TestIndex/department
After that the seeder ran fine and I didn't get the "index not found exception". However, it didn't use the index which I created. It created a new index of the form "TestIndex_write/department"
Can someone please explain what am I doing wrong? why is new index with a different name being created automatically?
I am using ElasticSearch > 6.0
Although 2 months past,try update index configurator :
php artisan elastic:update-index App\MyIndexConfigurator
I indexed a dataset of geo-data records in ElasticSearch for analysis in Kibana. My issue is that the 'Discover' tab doesn't pick up the data but instead displays the error message
Discover: An error occurred with your request. Reset your inputs and try again.
In 'Settings', I could configure my data index just fine, and Kibana is picking up all the mapping fields with correct type/analysis/indexing metadata. 'Visualize' works fine, too. I can create my charts, add them to the dashboard, drill down - everything. Just the 'Discover' tab is broken for me.
I'm running ElasticSearch 1.5.2, and tried with Kibana 4.0.1, 4.0.2 and 4.1-snapshot now (on Ubuntu 14.04), all with the same results.
Another effect I'm noticing: the sidebar is not showing any 'Available Fields'. Only if I unfold the field settings and untick 'Hide Missing Fields' I'll get my list of schema fields. (These are greyed out as they are considered 'missing' by Kibana. But interestingly, clicking on 'Visualize' on one of them to chart their distribution works, again, perfectly fine.)
My only suspicion is: my data doesn't have a timestamp field, so maybe that's what's messing things up. Although judging from the docs I'd assume that non-timeseries data should be supported.
Any hints appreciated!
In my case, the cause was that I had indexed malformed JSON into elasticsearch. It was valid Javascript, but not valid JSON. In particular I neglected to quote the keys in the objects
I had inserted my (test) data using curl, e.g.
curl -X PUT http://localhost:9200/foo/doc/1 -d '{ts: "2015-06-24T01:07:00.000Z", employeeId: 105, action: "PICK", quantity: 8}'
Note that ts: should have been "ts":, etc.
Seems like elasticsearch tolerates such things, but Kibana does not. Once I fixed that, Discover worked fine.
Note that the error you are seeing is generated client side when an error arises. If you open your client debugger (e.g. Firefox) you will see the error in the console log. In my case, the error message was
Error: Unable to parse/serialize body
If your error is different, it will be a different cause.
It was my fault for entering bad JSON to begin with. Odd that elasticsearch is more tolerant than Kibana.
It happened to me as well. I tried all...:
Deleting all the indices (.kibana, my own, etc) didn't work
Restarting the ES, Kibana and LS services didn't help.
I didn't have the Request Timeout problem in kibana.yml either.
My problem was that the timestamp field was using an incorrect time format. I changed it to this format and it worked: "date": "2015-05-13T00:00:00"
I had the same problem. None of the suggested solutions helped. I finally found the problem while comparing a working version with a non-working version in Wireshark.
Don't emit a UTF8 byte order mark in front of your JSON. Somehow, my serializer was set up to do that... ElasticSearch is fine with it, but Kibana cannot handle it on the Discover page.
I have a logstash-elasticsearch-kibana local setup and I have a problem when it comes to save Kibana dashboards.
Selecting the "Save" option I get the following error: "Save failed Dashboard could not be saved to Elasicsearch"
I'm using the logstash dashboard that comes with Kibana and after some modifications I tried to save it getting this error.
As far as I understand dashboards loaded from templates (json files located in kibana3/app/dashboards) cannot be saved to Elasticsearch (as stated in kibana templates). But I haven't been able to figure out how to create a new dashboard for logstash and save it to Elasticsearch, nor find instructions to do that. I would like to have different dashboards and be able to modify them and load them as needed.
I have exported the dashboard schema and successfully load it back, which works as far as saving a dashboard after all customization is done. But I would prefer to save them to elasticsearch rather than to template files.
Communication between ES and Kibana works fine (no errors show up in logs and information is retrieved and showed in Kibana).
Someone who could tell me what I'm missing here?
Thanks!
I got the error when I had a '/' (slash) in the name of the dashboard. Changing this to '-' solved the problem. See the following issue on GitHub: https://github.com/elasticsearch/kibana/issues/837