I have a requirement of creating multi-content document in FileNet. e.g. Let's say that I need to have one document which consists of a word document, A JPEG file and a MP4 video file as its content.
Any help is appreciated. Thanks!
The content of a Document is stored as ContentElement. Document can have multiple Content Elements – with their own MIME-type, name, etc.
Documents / Document Content
You can do this in 2 way . 1) Through Workplace/XT - while adding document you need to select Compound Document as Yes and need to followup the document add wizard to get the all docuemnts added in filenet .
2) In FEM lunch New Document wizard , on bottom check Compound Document and complete the wizard by adding all documents .
Related
I am new to stormcrawler and elasticsearch in general. I am currently using stormcrawler 2.0 to index website data (including non-HTML items such as PDF's and Word Documents) into elasticsearch. In some cases, the metadata of PDF's or Word documents do not contain a title so the field is stored blank/null in elasticsearch. This is unfortunately causing issues in the webapp I am using to display search results (search-ui). Is there a way I can have stormcrawler insert a default value of "Untitled" into the title field if none exists in the metadata?
I understand that elasticsearch has a null_value field parameter, but if I understand correctly that parameter cannot be used for text fields and only helps with searching.
Thanks!
One option would be to write a custom ParseFilter to give an arbitrary value to any missing key or a key with an empty value. The StormCrawler code has quite a few examples of ParseFilters, see also the WIKI.
The same could be done as a custom Bolt placed between the parser and the indexer; grab the metadata and normalise to your heart's content.
I would like to create a one to many relashanship for the purpose of aggregations.
The "join" will be according to a field called "common_id":
When I create the first document belonging to the same group I would like to use it's flakeId (it's _id) as the common_id.
When adding other document belonging to the same group I would like to explicitly set the common_id to have the same value as the first document I added. This can be done by my app since my application will know the common_id of the first element.
My problem is with the first document:
How can i tell elasticsearch to copy the _id into common_id in a single call to elastic (I know I can do it using update script, or using two calls one for index and one for update... but this requires two requests instead of one).
I would like a simple syntax for this.
thanks
I'm using kibana-4. Following the documentation here I should be able to create an index by putting this in my elasticsearch.yaml file:
PUT .kibana
{
"index.mapper.dynamic": true
}
I'm not sure I understand how to do this, because a yaml file should not take values formatted like the above block, right?
I noticed that .kibana was a default index, so after inputting it into the kibana console, I was asked to input a time field for the default index. However, the input HTML element is a dropdown that contained no options. Without selecting a time-field option I am not allowed to create a default index. What am I supposed to do? Has anyone else run into a similar problem?
I understand the problem faced by you. Even i faced the same while using Kibana 4 for first time.
Here are 2 possible solutions to your problem:-
1. Input data into elasticsearch which contains a timestamped field. So upon inputting data that field will be directly recognized by Kibana & would be showed to you in the dropdown menu (where you are currently seeing empty).
It is empty because Kibana couldn't recognize the timestamped field from the data inserted by you in elasticsearch.
2. Untick the option of Index contains time-based events which will allow you to just enter your index name & access Kibana.
Note:- while using Option 2 & specifying index name as .kibana you would notice that it doesn't contain any field or data because .kibana doesnt store any data.
I would suggest you to create an index using curl command and insert data in it with or without timestamped field. If inserted data without timestamped field use Option 2 otherwise use Option 1.
In Access 2013...
How can I search all the queries for a particular table name?
How can I find all the queries that include the UPDATE command?
How can I search all the tables for a particular field name?
Thanks.
You should give this a try. Essentially, you can export your objects into a text file, and do a search there.
I had indexed the document and I am able to search content of document.
But I want to find the type of document is indexed, author of document , name of document, size of document basically properties of file.
How it can be achieved with the help of elasticsearch.
Thank You.
Use elastic search head to take a look at the index tika generated in ElasticSearch for you
http://mobz.github.io/elasticsearch-head/
Then you can search on those fields by using queries like author:Lav (assuming you found a field named author in your index)