You are viewing a plain text version of this content. The canonical link for it is here.
Posted to commits@stanbol.apache.org by fc...@apache.org on 2012/06/25 13:41:04 UTC

svn commit: r1353471 - in /incubator/stanbol/site/trunk/content/stanbol/docs/trunk: customvocabulary.mdtext enhancementusage.mdtext multilingual.mdtext

Author: fchrist
Date: Mon Jun 25 11:41:03 2012
New Revision: 1353471

URL: http://svn.apache.org/viewvc?rev=1353471&view=rev
Log:
Review - minor fixes

Modified:
    incubator/stanbol/site/trunk/content/stanbol/docs/trunk/customvocabulary.mdtext
    incubator/stanbol/site/trunk/content/stanbol/docs/trunk/enhancementusage.mdtext
    incubator/stanbol/site/trunk/content/stanbol/docs/trunk/multilingual.mdtext

Modified: incubator/stanbol/site/trunk/content/stanbol/docs/trunk/customvocabulary.mdtext
URL: http://svn.apache.org/viewvc/incubator/stanbol/site/trunk/content/stanbol/docs/trunk/customvocabulary.mdtext?rev=1353471&r1=1353470&r2=1353471&view=diff
==============================================================================
--- incubator/stanbol/site/trunk/content/stanbol/docs/trunk/customvocabulary.mdtext (original)
+++ incubator/stanbol/site/trunk/content/stanbol/docs/trunk/customvocabulary.mdtext Mon Jun 25 11:41:03 2012
@@ -27,12 +27,14 @@ To build the indexing tool from source -
 In case you downloaded or "svn co" the source to {stanbol-source} and successfully build the source as described in the [Tutorial](tutorial.html) you still need to assembly the indexing tool by
  
     :::bash
-    {stanbol-source}/entityhub/indexing/genericrdf/ $ mvn assembly:single
+    $ cd {stanbol-source}/entityhub/indexing/genericrdf/
+    $ mvn assembly:single
 
 and move the assembled indexing tool from
 
     :::bash
-    {stanbol-source}/entityhub/indexing/genericrdf/target/org.apache.stanbol.entityhub.indexing.genericrdf-*-jar-with-dependencies.jar
+    {stanbol-source}/entityhub/indexing/genericrdf/target/ _
+    org.apache.stanbol.entityhub.indexing.genericrdf-*-jar-with-dependencies.jar
 
 into a the directory you plan to use for the indexing process. We will refer to this directory as {indexing-working-dir}.
 
@@ -42,7 +44,8 @@ into a the directory you plan to use for
 Initialize the tool with
 
     :::bash
-    {indexing-working-dir} $ java -jar org.apache.stanbol.entityhub.indexing.genericrdf-*-jar-with-dependencies.jar init
+    $ cd {indexing-working-dir}
+    $ java -jar org.apache.stanbol.entityhub.indexing.genericrdf-*-jar-with-dependencies.jar init
 
 This will create/initialize the default configuration for the Indexing Tool including (relative to {indexing-working-dir}):
 
@@ -63,7 +66,8 @@ For more details about possible configur
 Once all source files are in place, you can start the index process by running
 
     :::bash
-    {indexing-working-dir} $ java -Xmx1024m -jar org.apache.stanbol.entityhub.indexing.genericrdf-*-jar-with-dependencies.jar index
+    $ cd {indexing-working-dir}
+    $ java -Xmx1024m -jar org.apache.stanbol.entityhub.indexing.genericrdf-*-jar-with-dependencies.jar index
 
 Depending on your hardware and on complexity and size of your sources, it may take several hours to built the index. As a result, you will get an archive of an [Apache Solr](http://lucene.apache.org/solr/) index together with an OSGI bundle to work with the index in Stanbol.
 
@@ -90,93 +94,96 @@ After the installation your data will be
 
     http://{stanbol-instance}/entityhub/site/{name}
 
-You can use the Web UI of the Stanbol Enhancer to explore your vocabulary. Note that in case of big vocabulary it might take some time until the site becomes functional.
+You can use the Web UI of the Stanbol Enhancer to explore your vocabulary. Note, that in case of big vocabulary it might take some time until the site becomes functional.
 
-## B. Configure and use the index with the Stanbol Enhancer
+## B. Configure and use the index with the Apache Stanbol Enhancer
 
-This section covers how to configure the Stanbol Enhancer to recognize and link Entities of your custom vocabulary with processed Documents.
+This section covers how to configure the Apache Stanbol Enhancer to recognize and link entities of your custom vocabulary with processed documents.
 
-Generally there are two possible ways you can use to recognize Entities of your Vocabulary:
+Generally there are two possible ways you can use to recognize entities of your vocabulary:
 
-1. __Named Entity Linking__: This first uses Named Entity Recoqunition (NER) for spotting "Named Entities" in the text and second try to link those "Named Entities" with Entities defined in your Vocabulary. This approach is limited to Entities with the type Person, Organization and Places. So if your vocabulary contains entities of other types they will not be recognized. In addition it also requires the availability of NER for the language(s) of the processed documents.
-2. __Keyword Linking__: This uses the labels of Entities in your vocabulary for the recognition and linking process. Natural Language Processing (NLP) techniques such as Part-of-Speach (POS) detection can be used to improve performance and results but this works also without NLP support. As extraction and linking is based on labels mentioned in the analyzed content this method has no restrictions regarding the types of your Entities.
+1. __Named Entity Linking__: This first uses Named Entity Recoqunition (NER) for spotting "named entities" in the text and second try to link those named entities with entities defined in your vocabulary. This approach is limited to entities with the type person, organization and places. So if your vocabulary contains entities of other types, they will not be recognized. In addition it also requires the availability of NER for the language(s) of the processed documents.
+2. __Keyword Linking__: This uses the labels of entities in your vocabulary for the recognition and linking process. Natural Language Processing (NLP) techniques such as part-of-speach (POS) detection can be used to improve performance and results but this works also without NLP support. As extraction and linking is based on labels mentioned in the analyzed content this method has no restrictions regarding the types of your entities.
 
-For more information about this you might also have a look at the introduction of the [multi lingual](multilingual) Usage scenario.
+For more information about this you might also have a look at the introduction of the [multi lingual](multilingual) usage scenario.
 
 _TIP_: If you are unsure about what to use you can also start with configuring both options to give it a try. 
 
-Depending on if you want to use Named Entity Linking or Keyword Linking the configuration of the [Enhancement Chain](enhancer/chains) and the [Enhancement Engine](enhancer/engines) making use of your vocabulary will be different.
+Depending on if you want to use named entity linking or keyword linking the configuration of the [enhancement chain](enhancer/chains) and the [enhancement engine](enhancer/engines) making use of your vocabulary will be different.
 
 ### Configuring Named Entity Linking
 
-In case "Named Entity Linking" is used the linking with the custom vocabulary is done by the [Named Entity Tagging Engine](enhancer/engines/namedentitytaggingengine.html).
+In case named entity linking is used the linking with the custom vocabulary is done by the [Named Entity Tagging Engine](enhancer/engines/namedentitytaggingengine.html).
 For the configuration of this engine you need to provide the following parameters
 
-1. The "name" of the enhancement engine. It is recommended to use "{name}Linking" - where {name} is the name of your vocabulary as used in part A. of this scenario
-2. The name of the Referenced Site holding your vocabulary. Here you have to configure the {name}
-3. Enable/Disable Persons, Organizations and Places and if enabled configure the <code>rdf:type</code> used by your vocabulary for those type. If you do not want to restrict the type you can also leave the type field empty.
-4. Define the property used to match against the "Named Entities" detected by the used NER engine(s)
+1. The "name" of the enhancement engine. It is recommended to use "{name}Linking" - where {name} is the name of your vocabulary as used in part A. of this scenario.
+2. The name of the referenced site holding your vocabulary. Here you have to configure the {name}.
+3. Enable/disable persons, organizations and places and if enabled configure the <code>rdf:type</code> used by your vocabulary for those type. If you do not want to restrict the type, you can also leave the type field empty.
+4. Define the property used to match against the named entities detected by the used NER engine(s).
 
-For detailed information please see the documentation of the [Named Entity Tagging Engine](enhancer/engines/namedentitytaggingengine.html).
+For more detailed information please see the documentation of the [Named Entity Tagging Engine](enhancer/engines/namedentitytaggingengine.html).
 
-Note that for using Named Entity Linking you need also ensure that an Enhancement Engine that provides NER is available in the [Enhancement Chain](enhancer/chains). By default Apache Stanbol includes three different Engines that provide this feature: (1) [Named Entity Extraction Enhancement Engine](enhancer/engines/namedentityextractionengine.html) based on [OpenNLP](http://opennlp.apache.org), (2) CELI NER engine based on the [linguagrid.org](http://Linguagrid.org) service and (3) [OpenCalais Enhancement Engine](enhancer/engines/opencalaisengine.html) based on [OpenCalais](http://opencalais.com). Note that the later two options will require to send your content to the according services.
+Note, that for using named entity linking you need also ensure that an enhancement engine that provides NER is available in the [enhancement chain](enhancer/chains). By default Apache Stanbol includes three different engines that provide this feature: (1) [Named Entity Extraction Enhancement Engine](enhancer/engines/namedentityextractionengine.html) based on [OpenNLP](http://opennlp.apache.org), (2) CELI NER engine based on the [linguagrid.org](http://Linguagrid.org) service and (3) [OpenCalais Enhancement Engine](enhancer/engines/opencalaisengine.html) based on [OpenCalais](http://opencalais.com). Note that the later two options will require to send your content to the according services that are not part of your local Apache Stanbol instance.
 
-So a typical [Enhancement Chain](enhancer/chains) for Named Entity Linking with your vocabulary might look like
+A typical [enhancement chain](enhancer/chains) for named entity linking with your custom vocabulary might look like
 
-* "langid" - [Language Identification Engine](enhancer/engines/langidengine.html) - to detect the language of the parsed content - a pre requirement of all NER engines
+* "langid" - [Language Identification Engine](enhancer/engines/langidengine.html) - to detect the language of the parsed content - a pre-requirement of all NER engines
 * "ner" - for NER support in English, Spanish and Dutch via the [Named Entity Extraction Enhancement Engine](enhancer/engines/namedentityextractionengine.html)
 * "celiNer" - for NER support in French and Italien via the CELI NER engine
 * "{name}Linking - the [Named Entity Tagging Engine](enhancer/engines/namedentitytaggingengine.html) for your vocabulary as configured above.
 
-Both the [Weighted Chain](enhancer/chains/weightedchain.html) and the [List Chain](enhancer/chains/listchain.html) can be used for the configuration of such a chain.
+Both the [weighted chain](enhancer/chains/weightedchain.html) and the [list chain](enhancer/chains/listchain.html) can be used for the configuration of such a chain.
 
 ### Configure Keyword Linking
 
-In case you want to use "Keyword Linking" to extract and link Entities of your vocabulary you will need to configure the [Keyword Linking Engine](enhancer/engines/keywordlinkingengine.html) accordingly.
+In case you want to use keyword linking to extract and link entities of your vocabulary you will need to configure the [Keyword Linking Engine](enhancer/engines/keywordlinkingengine.html) accordingly.
 
-Here are the most important configuration options provided by the Keyword Linking Engine when configured via the [Configuration Tab](http://localhost:8080/system/console/configMgr) of the Apache Felix WebConsole - http://{host}:{port}/system/console/configMgr. For the full list and detailed information please see the [documentation](enhancer/engines/keywordlinkingengine.html)).
+Here are the most important configuration options provided by the Keyword Linking Engine when configured via the [configuration tab](http://localhost:8080/system/console/configMgr) of the Apache Felix web console - http://{host}:{port}/system/console/configMgr. For the full list and detailed information please see the [documentation](enhancer/engines/keywordlinkingengine.html)).
 
 1. The "Name" of the enhancement engine. It is recommended to use "{name}Keyword" - where {name} is the name of your vocabulary as used in part A. of this scenario
 2. The name of the "Referenced Site" holding your vocabulary. Here you have to configure the {name}
 3. The "Label Field" is the URI of the property in your vocabulary providing the labels used for matching. You can only use a single field. If you want to use values of several fields you have two options: (1) to adapt your indexing configuration to copy the values of those fields to a single one (e.g. the values of "skos:prefLabel" and "skos:altLabel" are copied to "rdfs:label" in the default configuration of the Entityhub indexing tool (see {indexing-working-dir}/indexing/config/mappings.txt) (2) to configure multiple Keyword Linking Engine(s) - one for each label field. Option (1) is preferable as long as you do not need to use different configurations for the different labels.
-4. The "Type Mappings" might be interesting for you if your vocabulary contains custom types as those mappings can be used to map 'rdf:type's of Entities in your Vocabulary to 'dc:type's used for 'fise:TextAnnotation's - created by the Stanbol Enhancer to annotate occurrences of extracted Entities in the parsed text. See the [Type Mapping Syntax](enhancer/engines/keywordlinkingengine.html#type-mappings-syntax) and the [Usage Scenario for the Stanbol Enhancement Structure](enhancementusage.html#entity-tagging-with-disambiguation-support) for details.
+4. The "Type Mappings" might be interesting for you if your vocabulary contains custom types as those mappings can be used to map 'rdf:type's of entities in your vocabulary to 'dc:type's used for 'fise:TextAnnotation's - created by the Apache Stanbol Enhancer to annotate occurrences of extracted entities in the parsed text. See the [type mapping syntax](enhancer/engines/keywordlinkingengine.html#type-mappings-syntax) and the [usage scenario for the Apache Stanbol Enhancement Structure](enhancementusage.html#entity-tagging-with-disambiguation-support) for details.
 
-A typical [Enhancement Chain](enhancer/chains) for Named Entity Linking with your vocabulary might look like
+A typical [enhancement chain](enhancer/chains) for named entity linking with your vocabulary might look like
 
-* "langid" - [Language Identification Engine](enhancer/engines/langidengine.html) - to detect the language of the parsed content - a pre requirement of the Keyword Linking Engine.
+* "langid" - [Language Identification Engine](enhancer/engines/langidengine.html) - to detect the language of the parsed content - a pre-requirement of the Keyword Linking Engine.
 * "{name}Keyword - the [Keyword Linking Engine](enhancer/engines/keywordlinkingengine.html) for your vocabulary as configured above.
 
-Both the [Weighted Chain](enhancer/chains/weightedchain.html) and the [List Chain](enhancer/chains/listchain.html) can be used for the configuration of such a chain.
+Both the [weighted chain](enhancer/chains/weightedchain.html) and the [list chain](enhancer/chains/listchain.html) can be used for the configuration of such a chain.
 
-### How to use EnhancementChains
+### How to use enhancement chains
 
-In the default configuration the Stanbol Enhancer provides two Enhancement chains: (1) a "default" chain that includes all currently active [Enhancement Engine](enhancer/engines)s and (2) the "language" Chain that is intended to be used to detect the language of parsed content.
+In the default configuration the Apache Stanbol Enhancer provides two enhancement chains:
 
-As soon as Stanbol users start to add own vocabularies to the Stanbol Entityhub and configure [Named Entity Tagging Engine](enhancer/engines/namedentitytaggingengine.html) or [Keyword Linking Engine](enhancer/engines/keywordlinkingengine.html) for them the default chain - that includes all active engines - becomes unprintable very soon. This is because those users might most likely want to deactivate the "default" chain and configure their own - as described above. This section provides more information on how to do that.
+1) a "default" chain that includes all currently active [enhancement engines](enhancer/engines) and 
+2) the "language" chain that is intended to be used to detect the language of parsed content.
 
-__Deactivate the Chain of all active Enhancement Engines__
+As soon as Apache Stanbol users start to add own vocabularies to the Apache Stanbol Entityhub and configure [Named Entity Tagging Engine](enhancer/engines/namedentitytaggingengine.html) or [Keyword Linking Engine](enhancer/engines/keywordlinkingengine.html), the default chain, which includes all active engines, may become unusable. Most likely users want to deactivate the "default" chain and configure their own. This section provides more information on how to do that.
 
-Users that add additional EnhancementEngines might need to deactivate the Enhancement Chain that includes all active engines. This can be done in the configuration tab of the Felic Webconsole - [http://{stabol-host}/system/console/configMgr](http://localhost:8080/system/console/configMgr). Open the configuration dialog of the "Apache Stanbol Enhancer Chain: Default Chain" component and deactivate it.
+__Deactivate the chain of all active enhancement engines__
 
-__Change the Enhancement Chain bound to "/enhancer"__
+Users that add additional enhancement engines might need to deactivate the enhancement chain that includes all active engines. This can be done in the configuration tab of the Apache Felix web console - [http://{stabol-host}/system/console/configMgr](http://localhost:8080/system/console/configMgr). Open the configuration dialog of the "Apache Stanbol Enhancer Chain: Default Chain" component and deactivate it.
 
-The Enhancement Chain bound to 
+__Change the enhancement chain bound to "/enhancer"__
+
+The enhancement chain bound to 
 
     http://{stanbol-host}/enhancer
 
 is determined by the following rules
 
-1. the Chain with the name "default". If more than one Chain is present with that name, than the above rules for resolving name conflicts apply. If none,
-2. the Chain with the highest "service.ranking". If several have the same ranking,
-3. the Chain with the lowest "service.id".
+1. the chain with the name "default". If more than one chain is present with that name, than the above rules for resolving name conflicts apply. If none,
+2. the chain with the highest "service.ranking". If several have the same ranking,
+3. the chain with the lowest "service.id".
 
-So you can change this by configuring the names and/or the "service.ranking" of the Enhancements Chians. NOTE also that (2) and (3) are also used to resolve name conflicts of chains. So if you configure two Enhancement Chains with the same name only the one with the highest "service.ranking" and lowest "service.id" will be accessible via the restful API. 
+You can change this by configuring the names and/or the "service.ranking" of the enhancements chains. Note, that (2) and (3) are also used to resolve name conflicts of chains. If you configure two enhancement chains with the same name, only the one with the highest "service.ranking" and lowest "service.id" will be accessible via the restful API. 
 
 ## Examples
 
 While this usage scenario provides the basic information about how to index/use custom vocabularies there are a lot of additional possibilities to configure the indexing process and the enhancement engines.
 
-If you are interested in the more advanced options the following resources/examples might be of interest to you.
+If you are interested in the more advanced options, the following resources/examples might be of interest to you.
 
 * [Readme](https://github.com/apache/stanbol/tree/17db70cb87ae5bbb905c1dbe76fbe4c0ca1dd90d/entityhub/indexing/genericrdf) of the generic RDF indexing tool (see also "{stanbol-source-root}/entityhub/indexing/genericrdf" if you have obtained the source code of Apache Stanbol).
 * [eHealth](https://github.com/apache/stanbol/tree/17db70cb87ae5bbb905c1dbe76fbe4c0ca1dd90d/demos/ehealth) example: This provides an indexing and enhancement engine configuration for 4 datasets of the life science domain. It goes into some of the details - such as customized Solr schema.xml configuration for the Apache Stanbol Entityhub; Keyword Linking Engine configurations optimized for extracting alpha-numeric IDs; using LD-Path to merge information of different datasets by following owl:sameAs relations; ... (see also "{stanbol-trunk}/demo/ehealth" if you have checked out the trunk Apache Stanbol). In addition this example may also provide some information on how to automate some of the steps described here by using shell scripts and maven. 

Modified: incubator/stanbol/site/trunk/content/stanbol/docs/trunk/enhancementusage.mdtext
URL: http://svn.apache.org/viewvc/incubator/stanbol/site/trunk/content/stanbol/docs/trunk/enhancementusage.mdtext?rev=1353471&r1=1353470&r2=1353471&view=diff
==============================================================================
--- incubator/stanbol/site/trunk/content/stanbol/docs/trunk/enhancementusage.mdtext (original)
+++ incubator/stanbol/site/trunk/content/stanbol/docs/trunk/enhancementusage.mdtext Mon Jun 25 11:41:03 2012
@@ -101,7 +101,7 @@ The following information of the [TextAn
 
 ## Entity checker - inline editing of content enhancements
 
-This describes a user interface similar to one of a spell/grammar checker. Instead of marking misspelled words entities recognized within the text are suggested to the user. The following figure shows such an interface as implemented by the [hallo.js](http://hallojs.org) combined with the [annotate.js](https://github.com/szabyg/annotate.js) plugin (see the [Demo here](http://hallojs.org/annotate.html) <small>(last accessed 2012-05-30)</small> - click in the Text and press the "annotate" button).
+This describes a user interface similar to one of a spell/grammar checker. Instead of marking misspelled words entities recognized within the text are suggested to the user. The following figure shows such an interface as implemented by the [hallo.js](http://hallojs.org) combined with the [annotate.js](https://github.com/szabyg/annotate.js) plugin (see the [demo here](http://hallojs.org/annotate.html) <small>(last accessed 2012-05-30)</small> - click in the Text and press the "annotate" button).
 
 ![Occurrence based Annotation UI](enhancer/hallo-annotate_scrrenshot.png "hallo.js with the annotate.js plugin used to implement an text occurrence based annotation UI")
 
@@ -148,7 +148,7 @@ _Tips and Tricks:_
 
 ### Interact with suggested entities
 
-This section explains how users mitt want to interact with extracted/suggested Entities. Extracted entities are represented by '[fise:EntityAnnotation](/enhancer/enhancementstructure.html#fiseentityannotation)'s. Those EntityAnnotations are linked with the [TextAnnotation](/enhancer/enhancementstructure.html#fisetextannotation) (occurrences) and to the entity of the used knowledge base. The following figure shows an example for an EntityAnnotation that suggests the entity ['dbpedia:Bob_Marley'](http:dbpedia.org/resource/Bob_Marley) for the TextAnnotation used in the example of the previous section.
+This section explains how users mitt want to interact with extracted/suggested entities. Extracted entities are represented by '[fise:EntityAnnotation](/enhancer/enhancementstructure.html#fiseentityannotation)'s. Those EntityAnnotations are linked with the [TextAnnotation](/enhancer/enhancementstructure.html#fisetextannotation) (occurrences) and to the entity of the used knowledge base. The following figure shows an example for an EntityAnnotation that suggests the entity ['dbpedia:Bob_Marley'](http:dbpedia.org/resource/Bob_Marley) for the TextAnnotation used in the example of the previous section.
 
 !['fise:EntityAnnotation' example](enhancer/es_entityannotation.png "This example shown an EntityAnnotation that suggests the entity 'dbpedia:Bob_Marley' for the TextAnnotation")
 

Modified: incubator/stanbol/site/trunk/content/stanbol/docs/trunk/multilingual.mdtext
URL: http://svn.apache.org/viewvc/incubator/stanbol/site/trunk/content/stanbol/docs/trunk/multilingual.mdtext?rev=1353471&r1=1353470&r2=1353471&view=diff
==============================================================================
--- incubator/stanbol/site/trunk/content/stanbol/docs/trunk/multilingual.mdtext (original)
+++ incubator/stanbol/site/trunk/content/stanbol/docs/trunk/multilingual.mdtext Mon Jun 25 11:41:03 2012
@@ -3,10 +3,10 @@ Title: Configure Apache Stanbol to work 
 
 To understand multi lingual support with Apache Stanbol one needs to consider that Stanbol supports two different workflows for extracting Entities from parsed text:
 
-1. __Named Entity Linking__: This first uses Named Entity Recoqunition (NER) for spotting Entities and second linked found Named Entities with Entities defined by the Controlled Vocabulary (e.g. DBpedia.or). For the NER step the [NamedEntityExtraction](enhancer/engines/namedentityextractionengine.html), the CELI NER engine - using the [linguagrid.org](http://linguagrid.org) service or the [OpenCalais](enhancer/engines/opencalaisengine.html) /work/workspace/stanbol-website/content/stanbol/docs/trunk/enhancer/engines/refactorengine.mdtextcan be used. The linking functionality is implemented by the [NamedEntityTaggingEngine](enhancer/engines/namedentitytaggingengine.html). Multi lingual support depends on the availability of NER models for a language. Note also that separate models are required for each Entity type. Typical supported types are Persons, Organizations and Places.
-2. __Keyword Linking__: Entity label based spotting and linking of Entities as implemented by the [KeywordLinkingEngine](enhancer/engines/keywordlinkingengine.html). Natural Language Processing (NLP) techniques such as Part-of-Speach (POS) processing are used to improve performance and result of the extraction process but are not a absolute requirement. As extraction only requires a label this method is also independent of the types of the Entities.
+1. __Named Entity Linking__: This first uses Named Entity Recoqunition (NER) for spotting Entities and second linked found Named Entities with Entities defined by the Controlled Vocabulary (e.g. DBpedia.or). For the NER step the [NamedEntityExtraction](enhancer/engines/namedentityextractionengine.html), the CELI NER engine - using the [linguagrid.org](http://linguagrid.org) service or the [OpenCalais](enhancer/engines/opencalaisengine.html) /work/workspace/stanbol-website/content/stanbol/docs/trunk/enhancer/engines/refactorengine.mdtextcan be used. The linking functionality is implemented by the [NamedEntityTaggingEngine](enhancer/engines/namedentitytaggingengine.html). Multi lingual support depends on the availability of NER models for a language. Note also that separate models are required for each entity type. Typical supported types are Persons, Organizations and Places.
+2. __Keyword Linking__: entity label based spotting and linking of Entities as implemented by the [KeywordLinkingEngine](enhancer/engines/keywordlinkingengine.html). Natural Language Processing (NLP) techniques such as Part-of-Speach (POS) processing are used to improve performance and result of the extraction process but are not a absolute requirement. As extraction only requires a label this method is also independent of the types of the Entities.
 
-The following Languages are supported for NER - and can therefore be used for Named Entity Linking:
+The following Languages are supported for NER - and can therefore be used for Named entity Linking:
 
 * __English__ (via [NamedEntityTaggingEngine](enhancer/engines/namedentitytaggingengine.html), [OpenCalais](enhancer/engines/opencalaisengine.html))
 * __Spansh__ (via [NamedEntityTaggingEngine](enhancer/engines/namedentitytaggingengine.html), [OpenCalais](enhancer/engines/opencalaisengine.html))
@@ -101,7 +101,7 @@ The following describe the different Opt
 - **Case Sensitivity:** Allows to enable case sensitive matching of labels. This allows to work around problems with suggesting abbreviations like "AND" for mentions of the english stop word "and". 
 - **Type Field:** The field used to retrieve the types of matched Entities. The values of this field are added to the 'fise:entity-type' property of created 'fise:EntityAnnotation's. 
 - **Redirect Field:** Entities may define redirects to other Entities (e.g. "USA"(http://dbpedia.org/resource/USA) -> "United States"(http://dbpedia.org/resource/United_States). Values of this field are expected to link to other entities part of the controlled vocabulary
-- **Redirect Mode:** Defines how to process redirects of Entities mentioned in the parsed content.. Three modes to deal with such links are supported: Ignore redirects; Add values from redirected Entities to extracted; Follow Redirects and suggest the redirected Entity instead of the extracted. 
+- **Redirect Mode:** Defines how to process redirects of Entities mentioned in the parsed content.. Three modes to deal with such links are supported: Ignore redirects; Add values from redirected Entities to extracted; Follow Redirects and suggest the redirected entity instead of the extracted. 
 - **Min Token Length:**	The minimum length of Tokens used to lookup Entities within the Controlled Vocabulary. This parameter is ignored in case a certain POS (Part of Speech) tag is available.
 - **Keyword Tokenizer:** Forces the use of a word tokenizer that is optimized for Alpha numeric keys such as ISBN numbers, product codes ... 
 - **Suggestions:** The maximal number of suggestions returned for a single mention. 
@@ -132,13 +132,13 @@ A Enhancement Chain configuration consis
 - **Engines:** The name of the engine(s) used by this Chain. The order of the configuration is only used by the "List Chain". "{engine-name};optional" can be used to specify that this Chain can still be used if this engine is currently not available or fails to process a content item.
 - **Ranking:** If there are two Enhancement Chains with the same name, than the one with the higher ranking will be executed
 
-See also [the documentation](enhancer/chains) for details on Enhancement chains.
+See also [the documentation](enhancer/chains) for details on enhancement chains.
 
 ##Results
 
 Extracted Entities will be formally describend in the RDF enhancement results of the Stanbol Enhancer by
 
-* fise:TextAnnotation: The occurrence of the extracted Entity within the Text. Also providing the general nature - value of the 'dc:type' property - of the Entity. In case of Named Entity Linking TextAnnotations represent the Named Entities extracted by the used NER engine(s)
+* fise:TextAnnotation: The occurrence of the extracted entity within the Text. Also providing the general nature - value of the 'dc:type' property - of the entity. In case of Named Entity Linking TextAnnotations represent the Named Entities extracted by the used NER engine(s)
 * fise:EntityAnnotation: Entities of the configured controlled vocabulary suggested for one or more 'fise:TextAnnotation's - value(s) of the 'dc:relation' property.
 
 The following figure provides an overview about the knowledge structure.