Type Descriptor File for StanfordCore NLP for Apache UIMA RUTA - stanford-nlp

I am trying to annotate a text in German literature in Apache UIMA by writing rules in RUTA. I am using DKPro Core as well. I am very new to this and figuring out how to do stuff.
I am unable to get few annotations which are not mentioned in TypeDescriptor files as generated by the example German Novels (https://github.com/apache/uima-ruta/tree/trunk/example-projects/GermanNovels).
For example: ADJA tag in Partofspeech tagset that is available in Standford NLP POS tagger.
I searched for Typedescriptor file for StanfordCoreNLP but couldn't find on the net.
How can I generate these files?

Related

Convert multiple YAMLs to pojos using maven

How do i convert multiple OpenAPI based YAML files POJO using maven plugin using java (not springboot)
Example: https://github.com/jdegre/5GC_APIs
Some reference code shall be of great help

How do I get the enhanced++ dependencies of CoreNLP using the Stanza python wrapper in google colab?

Stanza with Basic Dependencies
I need to get the enhanced++ dependencies instead of the above basic dependencies.
I have used the code from here: https://stanfordnlp.github.io/stanza/depparse.html
The Stanza depparse model only produces basic dependencies. However, there is a universal converter using a pipe to a Java process from CoreNLP which may help:
https://stanfordnlp.github.io/stanza/client_regex.html#universal-enhancements
If you have any issues using that, I suggest posting an issue on the stanza github

What features Apache OpenNLP does use by default while running its named entity recognition (NER) models?

I know Apache OpenNLP uses MaxEnt model for its NER tagger. But what features Apache OpenNLP does use (by default) while running its named entity recognition (NER) models? and also how can we incorporate/customize new features in OpenNLP (Java implementation)?
In Apache OpenNLP NER, it allows users to define features via XML file. The default XML is this:
https://github.com/apache/opennlp/blob/master/opennlp-tools/src/main/resources/opennlp/tools/namefind/ner-default-features.xml
If you want to customize it, use -featuregen option when you train the model:
$ opennlp TokenNameFinderTrainer -featuregen your-features-definition.xml -model my-model.bin ...
You don't need to specify your customized feature XML file when you execute TokenNameFinder as the model file includes the information of your features.

Want to use #Value (reading the properties from property file) in UIMA framework

I have a property file like myProperties.properties. I want to read one property like MAX_YEARS using spring annotation #value as like below in UIMA JCasAnnotator_ImplBase extending class.
private #Value("${REQUIRED_COLUMNS}") String requiredColumns;
Or any alternatives for reading properties from property file in UIMA framework.
Thanks in advance.
Narasimha.
UIMA does not support value injection via Java annotations (from Spring or any other DI frameworks) at this time. It does support External Configuration Parameter Overrides, though.
uimaFIT offers annotations like #ConfigurationParameter to inject UIMA parameters into fields. These parameter values can come from descriptors automatically generated by uimaFIT using reflection, or they can come from pre-built XML descriptors.
When using pre-built XML descriptors, it should be possible to employ the External Configuration Parameter Overrides mechanism in conjunction with uimaFIT - but I am not sure if this has already been tried by anybody.
It may even be possible to employ the External Configuration Parameter Overrides mechanism with the descriptors internally generated by uimaFIT.
Disclosure: I am a developer on the UIMA project, focussing on uimaFIT.

Stanford Ner API Details

Iam trying to find the Stanford Ner API
i could not find Stanford Ner API details
in Stanford NLP website.
please can any body help to find out those
API
Where I can get Stanford Ner API.
Stanford NER is included in the Stanford NLP javadocs. Or you can download release from the Stanford NER page and make javadocs from the sources.
Stanford NER is library, you must download it, import for your project, all API is declared in file NERDemo.java.
You can get a CoreLabel object from result.

Resources