You are viewing a plain text version of this content. The canonical link for it is here.
Posted to dev@rya.apache.org by mi...@apache.org on 2015/11/24 20:31:24 UTC

[1/3] incubator-rya git commit: RYA-3 Changing Manual to Markdown; linking README

Repository: incubator-rya
Updated Branches:
  refs/heads/develop 2a6f73088 -> ce4a10ff5


http://git-wip-us.apache.org/repos/asf/incubator-rya/blob/ce4a10ff/extras/rya.manual/src/site/markdown/_index.md
----------------------------------------------------------------------
diff --git a/extras/rya.manual/src/site/markdown/_index.md b/extras/rya.manual/src/site/markdown/_index.md
new file mode 100644
index 0000000..184b94f
--- /dev/null
+++ b/extras/rya.manual/src/site/markdown/_index.md
@@ -0,0 +1,23 @@
+
+# Rya
+- [Overview](overview.md)
+- [Quick Start](quickstart.md)
+- [Load Data](loaddata.md)
+- [Query Data](querydata.md)
+- [Evaluation Table](eval.md)
+- [Pre-computed Joins](loadPrecomputedJoin.md)
+- [Inferencing](infer.md)
+
+# Samples
+- [Typical First Steps](sm-firststeps.md)
+- [Simple Add/Query/Remove Statements](sm-simpleaqr.md)
+- [Sparql query](sm-sparqlquery.md)
+- [Adding Authentication](sm-addauth.md)
+- [Inferencing](sm-infer.md)
+- [Named Graph](sm-namedgraph.md)
+- [Update data](sm-updatedata.md)
+- [Alx](alx.md)
+
+# Development
+- [Building From Source](build-source.md)
+- [LTS Maven Settings XML](maven-settings.md)

http://git-wip-us.apache.org/repos/asf/incubator-rya/blob/ce4a10ff/extras/rya.manual/src/site/markdown/alx.md
----------------------------------------------------------------------
diff --git a/extras/rya.manual/src/site/markdown/alx.md b/extras/rya.manual/src/site/markdown/alx.md
new file mode 100644
index 0000000..78a4c8e
--- /dev/null
+++ b/extras/rya.manual/src/site/markdown/alx.md
@@ -0,0 +1,61 @@
+# Alx Rya Integration
+
+Alx is a modular framework for developing applications. Rya has mechanisms to integrate directly into Alx to provide other modules access to queries.
+
+Currently, the Alx Rya extension only allows interacting with an Accumulo store.
+
+## Prerequisites
+
+- Alx 1.0.5+ (we will refer to it at the ALX_HOME directory from now on)
+- alx.rya features xml (can be found in maven at `mvn:mvm.rya/alx.rya/<version>/xml/features`)
+
+## Steps
+
+1. Start up Alx
+2. features:addurl alx.rya features xml
+3. features:install alx.rya
+4. (optional) features:install alx.rya.console
+
+That's it. To make sure, run `ls <alx.rya bundle id>` and make sure something like this pops up:
+
+```
+mvm.rya.alx.rya (99) provides:
+------------------------------
+Bundle-SymbolicName = mvm.rya.alx.rya
+Bundle-Version = 3.0.4.SNAPSHOT
+objectClass = org.osgi.service.cm.ManagedService
+service.id = 226
+service.pid = mvm.rya.alx
+----
+...
+```
+
+## Using
+
+The bundle registers a Sail Repository, so you can interact with it directly as in the other code examples. Here is a quick groovy example of the usage:
+
+``` JAVA
+import org.springframework.osgi.extensions.annotation.*;
+import org.openrdf.repository.*;
+import org.openrdf.model.ValueFactory;
+import static mvm.rya.api.RdfCloudTripleStoreConstants.*;
+
+class TstRepo {
+
+	@ServiceReference
+	public void setRepo(Repository repo) {
+		println repo
+		RepositoryConnection conn = repo.getConnection();
+		ValueFactory vf = VALUE_FACTORY;
+        def statements = conn.getStatements(vf.createURI("http://www.Department0.University0.edu"), null, null, true);
+        while(statements.hasNext()) {
+            System.out.println(statements.next());
+        }
+        statements.close();
+        conn.close();
+	}
+
+}
+```
+
+The bundle also registers a RyaDAO, so you can interact with the RyaDAO interface directly

http://git-wip-us.apache.org/repos/asf/incubator-rya/blob/ce4a10ff/extras/rya.manual/src/site/markdown/build-source.md
----------------------------------------------------------------------
diff --git a/extras/rya.manual/src/site/markdown/build-source.md b/extras/rya.manual/src/site/markdown/build-source.md
new file mode 100644
index 0000000..e811622
--- /dev/null
+++ b/extras/rya.manual/src/site/markdown/build-source.md
@@ -0,0 +1,15 @@
+# Building from Source
+
+## Prerequisites
+
+* Rya code
+* Maven 2.2 +
+
+## Building
+
+Using Git, pull down the latest code from the url above.
+
+Run the command to build the code `mvn clean install`
+
+If all goes well, here are the artifacts that you will be interested in:
+* Rya-WAR : web/web-rya/target/web.rya.war

http://git-wip-us.apache.org/repos/asf/incubator-rya/blob/ce4a10ff/extras/rya.manual/src/site/markdown/eval.md
----------------------------------------------------------------------
diff --git a/extras/rya.manual/src/site/markdown/eval.md b/extras/rya.manual/src/site/markdown/eval.md
new file mode 100644
index 0000000..8a40389
--- /dev/null
+++ b/extras/rya.manual/src/site/markdown/eval.md
@@ -0,0 +1,58 @@
+# Prospects Table
+
+The Prospects Table provides statistics on the number of subject/predicate/object data found in the triple store. It is currently a
+Map Reduce job that will run against the Rya store and save all the statistics in the prosepcts table.
+
+## Build
+
+[Build the mmrts.git repo](build-source.md)
+
+## Run
+
+Deploy the `extras/rya.prospector/target/rya.prospector-<version>-shade.jar` file to the hadoop cluster.
+
+The prospector also requires a configuration file that defines where Accumulo is, which Rya table (has to be the SPO table) to read from, and
+which table to output to. (Note: Make sure you follow the same schema as the Rya tables (prospects table name: tableprefix_prospects)
+
+A sample configuration file might look like the following:
+
+``` XML
+<?xml version="1.0"?>
+<?xml-stylesheet type="text/xsl" href="configuration.xsl"?>
+<configuration>
+    <property>
+        <name>prospector.intable</name>
+        <value>triplestore_spo</value>
+    </property>
+    <property>
+        <name>prospector.outtable</name>
+        <value>triplestore_prospects</value>
+    </property>
+    <property>
+        <name>prospector.auths</name>
+        <value>U,FOUO</value>
+    </property>
+    <property>
+        <name>instance</name>
+        <value>accumulo</value>
+    </property>
+    <property>
+        <name>zookeepers</name>
+        <value>localhost:2181</value>
+    </property>
+    <property>
+        <name>username</name>
+        <value>root</value>
+    </property>
+    <property>
+        <name>password</name>
+        <value>secret</value>
+    </property>
+</configuration>
+```
+
+Run the command, filling in the correct information.
+
+```
+hadoop jar rya.prospector-3.0.4-SNAPSHOT-shade.jar mvm.rya.prospector.mr.Prospector /tmp/prospectorConf.xml
+```
\ No newline at end of file

http://git-wip-us.apache.org/repos/asf/incubator-rya/blob/ce4a10ff/extras/rya.manual/src/site/markdown/index.md
----------------------------------------------------------------------
diff --git a/extras/rya.manual/src/site/markdown/index.md b/extras/rya.manual/src/site/markdown/index.md
new file mode 100644
index 0000000..aa49e3b
--- /dev/null
+++ b/extras/rya.manual/src/site/markdown/index.md
@@ -0,0 +1,24 @@
+# Rya
+
+This project contains documentation about the Rya, a scalable RDF triple store on top of Accumulo.
+
+- [Overview](overview.md)
+- [Quick Start](quickstart.md)
+- [Load Data](loaddata.md)
+- [Query Data](querydata.md)
+- [Evaluation Table](eval.md)
+- [Pre-computed Joins](loadPrecomputedJoin.md)
+- [Inferencing](infer.md)
+
+# Samples
+- [Typical First Steps](sm-firststeps.md)
+- [Simple Add/Query/Remove Statements](sm-simpleaqr.md)
+- [Sparql query](sm-sparqlquery.md)
+- [Adding Authentication](sm-addauth.md)
+- [Inferencing](sm-infer.md)
+- [Named Graph](sm-namedgraph.md)
+- [Update data](sm-updatedata.md)
+- [Alx](alx.md)
+
+# Development
+- [Building From Source](build-source.md)

http://git-wip-us.apache.org/repos/asf/incubator-rya/blob/ce4a10ff/extras/rya.manual/src/site/markdown/infer.md
----------------------------------------------------------------------
diff --git a/extras/rya.manual/src/site/markdown/infer.md b/extras/rya.manual/src/site/markdown/infer.md
new file mode 100644
index 0000000..ee769c5
--- /dev/null
+++ b/extras/rya.manual/src/site/markdown/infer.md
@@ -0,0 +1,14 @@
+# Inferencing
+
+The current inferencing set supported includes:
+
+* rdfs:subClassOf
+* rdfs:subPropertyOf
+* owl:equivalentProperty
+* owl:inverseOf
+* owl:SymmetricProperty
+* owl:TransitiveProperty (* This is implemented, but probably not fully. Still in testing)
+
+Nothing special has to be done outside of making sure that the RdfCloudTripleStore object has the InferencingEngine object set on it and properly configured. This is usually done by default. See the [Query Data Section](querydata.md) for a simple example.
+
+Also, the inferencing engine is set to pull down the latest model every 5 minutes currently (which is configurable). So if you load a new model, a previous RepositoryConnection may not pick up these changes into the Inferencing Engine yet. Getting the InferencingEngine object from the RdfCloudTripleStore and running the `refreshGraph` method can refresh the inferred graph immediately.
\ No newline at end of file

http://git-wip-us.apache.org/repos/asf/incubator-rya/blob/ce4a10ff/extras/rya.manual/src/site/markdown/loadPrecomputedJoin.md
----------------------------------------------------------------------
diff --git a/extras/rya.manual/src/site/markdown/loadPrecomputedJoin.md b/extras/rya.manual/src/site/markdown/loadPrecomputedJoin.md
new file mode 100644
index 0000000..472b409
--- /dev/null
+++ b/extras/rya.manual/src/site/markdown/loadPrecomputedJoin.md
@@ -0,0 +1,28 @@
+# Load Pre-computed Join
+
+A tool has been created to load a pre-computed join.  This tool will generate an index to support a pre-computed join on a user provided SPARQL query, and then register that query within Rya.
+
+
+## Registering a pre-computed join
+
+Generating a pre-computed join is done using Pig to execute a series of Map Reduce jobs.  The index (pre-computed join) is associated with a user defined SPARQL query.  
+  
+To execute the indexing tool, compile and run `mvm.rya.accumulo.pig.IndexWritingTool` 
+with the following seven input arguments: `[hdfsSaveLocation] [sparqlFile] [instance] [cbzk] [user] [password] [rdfTablePrefix]`
+
+
+Options:
+
+* hdfsSaveLocation: a working directory on hdfs for storing interim results
+* sparqlFile: the query to generate a precomputed join for
+* instance: the accumulo instance name
+* cbzk: the accumulo zookeeper name
+* user: the accumulo username
+* password:  the accumulo password for the supplied user
+* rdfTablePrefix : The tables (spo, po, osp) are prefixed with this qualifier. The tables become: (rdf.tablePrefix)spo,(rdf.tablePrefix)po,(rdf.tablePrefix)osp
+
+
+# Using a Pre-computed Join
+
+An example of using a pre-computed join can be referenced in 
+`mvm.rya.indexing.external.ExternalSailExample`

http://git-wip-us.apache.org/repos/asf/incubator-rya/blob/ce4a10ff/extras/rya.manual/src/site/markdown/loaddata.md
----------------------------------------------------------------------
diff --git a/extras/rya.manual/src/site/markdown/loaddata.md b/extras/rya.manual/src/site/markdown/loaddata.md
new file mode 100644
index 0000000..3a66d6a
--- /dev/null
+++ b/extras/rya.manual/src/site/markdown/loaddata.md
@@ -0,0 +1,121 @@
+# Load Data
+
+There are a few mechanisms to load data
+
+## Web REST endpoint
+
+The War sets up a Web REST endpoint at `http://server/web.rya/loadrdf` that allows POST data to get loaded into the Rdf Store. This short tutorial will use Java code to post data.
+
+First, you will need data to load and will need to figure out what format that data is in.
+
+For this sample, we will use the following N-Triples:
+
+```
+<http://mynamespace/ProductType1> <http://www.w3.org/1999/02/22-rdf-syntax-ns#type> <http://mynamespace/ProductType> .
+<http://mynamespace/ProductType1> <http://www.w3.org/2000/01/rdf-schema#label> "Thing" .
+<http://mynamespace/ProductType1> <http://purl.org/dc/elements/1.1/publisher> <http://mynamespace/Publisher1> .
+```
+
+Save this file somewhere `$RDF_DATA`
+
+Second, use the following Java code to load data to the REST endpoint:
+
+``` JAVA
+import java.io.BufferedReader;
+import java.io.InputStream;
+import java.io.InputStreamReader;
+import java.io.OutputStream;
+import java.net.URL;
+import java.net.URLConnection;
+
+public class LoadDataServletRun {
+
+    public static void main(String[] args) {
+        try {
+            final InputStream resourceAsStream = Thread.currentThread().getContextClassLoader()
+                    .getResourceAsStream("$RDF_DATA");
+            URL url = new URL("http://server/web.rya/loadrdf" +
+                    "?format=N-Triples" +
+                    "");
+            URLConnection urlConnection = url.openConnection();
+            urlConnection.setRequestProperty("Content-Type", "text/plain");
+            urlConnection.setDoOutput(true);
+
+            final OutputStream os = urlConnection.getOutputStream();
+
+            int read;
+            while((read = resourceAsStream.read()) >= 0) {
+                os.write(read);
+            }
+            resourceAsStream.close();
+            os.flush();
+
+            BufferedReader rd = new BufferedReader(new InputStreamReader(
+                    urlConnection.getInputStream()));
+            String line;
+            while ((line = rd.readLine()) != null) {
+                System.out.println(line);
+            }
+            rd.close();
+            os.close();
+        } catch (Exception e) {
+            e.printStackTrace();
+        }
+    }
+}
+```
+
+Compile and run this code above, changing the references for $RDF_DATA and the url that your Rdf War is running at.
+
+The default "format" is RDF/XML, but these formats are supported : RDFXML, NTRIPLES, TURTLE, N3, TRIX, TRIG.
+
+## Bulk Loading data
+
+Bulk loading data is done through Map Reduce jobs
+
+### Bulk Load RDF data
+
+This Map Reduce job will read a full file into memory and parse it into statements. The statements are saved into the store. Here is an example for storing in Accumulo:
+
+```
+hadoop jar target/accumulo.rya-3.0.4-SNAPSHOT-shaded.jar mvm.rya.accumulo.mr.fileinput.BulkNtripsInputTool -Dac.zk=localhost:2181 -Dac.instance=accumulo -Dac.username=root -Dac.pwd=secret -Drdf.tablePrefix=triplestore_ -Dio.sort.mb=64 /tmp/temp.ntrips
+```
+
+Options:
+
+- rdf.tablePrefix : The tables (spo, po, osp) are prefixed with this qualifier. The tables become: (rdf.tablePrefix)spo,(rdf.tablePrefix)po,(rdf.tablePrefix)osp
+- ac.* : Accumulo connection parameters
+- rdf.format : See RDFFormat from openrdf, samples include (Trig, N-Triples, RDF/XML)
+- io.sort.mb : Higher the value, the faster the job goes. Just remember that you will need this much ram at least per mapper
+
+The argument is the directory/file to load. This file needs to be loaded into HDFS before running.
+
+## Direct OpenRDF API
+
+Here is some sample code to load data directly through the OpenRDF API. (Loading N-Triples data)
+You will need at least `accumulo.rya-<version>`, `rya.api`, `rya.sail.impl` on the classpath and transitive dependencies. I find that Maven is the easiest way to get a project dependency tree set up.
+
+``` JAVA
+final RdfCloudTripleStore store = new RdfCloudTripleStore();
+AccumuloRdfConfiguration conf = new AccumuloRdfConfiguration();
+AccumuloRyaDAO dao = new AccumuloRdfDAO();
+Connector connector = new ZooKeeperInstance("instance", "zoo1,zoo2,zoo3").getConnector("user", "password");
+dao.setConnector(connector);
+conf.setTablePrefix("rya_");
+dao.setConf(conf);
+store.setRdfDao(dao);
+
+Repository myRepository = new RyaSailRepository(store);
+myRepository.initialize();
+RepositoryConnection conn = myRepository.getConnection();
+
+//load data from file
+final File file = new File("ntriples.ntrips");
+conn.add(new FileInputStream(file), file.getName(),
+        RDFFormat.NTRIPLES, new Resource[]{});
+
+conn.commit();
+
+conn.close();
+myRepository.shutDown();
+```
\ No newline at end of file

http://git-wip-us.apache.org/repos/asf/incubator-rya/blob/ce4a10ff/extras/rya.manual/src/site/markdown/overview.md
----------------------------------------------------------------------
diff --git a/extras/rya.manual/src/site/markdown/overview.md b/extras/rya.manual/src/site/markdown/overview.md
new file mode 100644
index 0000000..546530f
--- /dev/null
+++ b/extras/rya.manual/src/site/markdown/overview.md
@@ -0,0 +1,5 @@
+# Overview
+
+RYA is a scalable RDF Store that is built on top of a Columnar Index Store (such as Accumulo). It is implemented as an extension to OpenRdf to provide easy query mechanisms (SPARQL, SERQL, etc) and Rdf data storage (RDF/XML, NTriples, etc).
+
+RYA stands for RDF y(and) Accumulo.

http://git-wip-us.apache.org/repos/asf/incubator-rya/blob/ce4a10ff/extras/rya.manual/src/site/markdown/querydata.md
----------------------------------------------------------------------
diff --git a/extras/rya.manual/src/site/markdown/querydata.md b/extras/rya.manual/src/site/markdown/querydata.md
new file mode 100644
index 0000000..70f3045
--- /dev/null
+++ b/extras/rya.manual/src/site/markdown/querydata.md
@@ -0,0 +1,116 @@
+# Query Data
+
+There are a few mechanisms to query data
+
+## Web JSP endpoint
+
+Open a url to `http://server/web.rya/sparqlQuery.jsp`. This simple form can run Sparql.
+
+## Web REST endpoint
+
+The War sets up a Web REST endpoint at `http://server/web.rya/queryrdf` that allows GET requests with queries.
+
+For this sample, we will assume you already loaded data from the [Load Data](loaddata.md) tutorial
+
+Save this file somewhere $RDF_DATA
+
+Second, use the following Java code to load data to the REST endpoint:
+
+``` JAVA
+import java.io.BufferedReader;
+import java.io.InputStreamReader;
+import java.net.URL;
+import java.net.URLConnection;
+import java.net.URLEncoder;
+
+public class QueryDataServletRun {
+
+    public static void main(String[] args) {
+        try {
+            String query = "select * where {\n" +
+                                "<http://mynamespace/ProductType1> ?p ?o.\n" +
+                                "}";
+
+            String queryenc = URLEncoder.encode(query, "UTF-8");
+
+            URL url = new URL("http://server/rdfTripleStore/queryrdf?query=" + queryenc);
+            URLConnection urlConnection = url.openConnection();
+            urlConnection.setDoOutput(true);
+
+            BufferedReader rd = new BufferedReader(new InputStreamReader(
+                    urlConnection.getInputStream()));
+            String line;
+            while ((line = rd.readLine()) != null) {
+                System.out.println(line);
+            }
+            rd.close();
+        } catch (Exception e) {
+            e.printStackTrace();
+        }
+    }
+}
+```
+
+Compile and run this code above, changing the url that your Rdf War is running at.
+
+## Direct Code
+
+Here is a code snippet for directly running against Accumulo with the code. You will need at least accumulo.rya.jar, rya.api, rya.sail.impl on the classpath and transitive dependencies. I find that Maven is the easiest way to get a project dependency tree set up.
+
+``` JAVA
+Connector connector = new ZooKeeperInstance("instance", "zoo1,zoo2,zoo3").getConnector("user", "password");
+
+final RdfCloudTripleStore store = new RdfCloudTripleStore();
+AccumuloRyaDAO crdfdao = new AccumuloRyaDAO();
+crdfdao.setConnector(connector);
+
+AccumuloRdfConfiguration conf = new AccumuloRdfConfiguration();
+conf.setTablePrefix("rts_");
+conf.setDisplayQueryPlan(true);
+crdfdao.setConf(conf);
+store.setRdfDao(crdfdao);
+
+ProspectorServiceEvalStatsDAO evalDao = new ProspectorServiceEvalStatsDAO(connector, conf);
+evalDao.init();
+store.setRdfEvalStatsDAO(evalDao);
+
+InferenceEngine inferenceEngine = new InferenceEngine();
+inferenceEngine.setRdfDao(crdfdao);
+inferenceEngine.setConf(conf);
+store.setInferenceEngine(inferenceEngine);
+
+Repository myRepository = new RyaSailRepository(store);
+myRepository.initialize();
+
+String query = "select * where {\n" +
+        "<http://mynamespace/ProductType1> ?p ?o.\n" +
+        "}";
+RepositoryConnection conn = myRepository.getConnection();
+System.out.println(query);
+TupleQuery tupleQuery = conn.prepareTupleQuery(
+        QueryLanguage.SPARQL, query);
+ValueFactory vf = ValueFactoryImpl.getInstance();
+
+TupleQueryResultHandler writer = new SPARQLResultsXMLWriter(System.out);
+tupleQuery.evaluate(new TupleQueryResultHandler() {
+
+    int count = 0;
+
+    @Override
+    public void startQueryResult(List<String> strings) throws TupleQueryResultHandlerException {
+    }
+
+    @Override
+    public void endQueryResult() throws TupleQueryResultHandlerException {
+    }
+
+    @Override
+    public void handleSolution(BindingSet bindingSet) throws TupleQueryResultHandlerException {
+        System.out.println(bindingSet);
+    }
+});
+
+conn.close();
+myRepository.shutDown();
+```
+

http://git-wip-us.apache.org/repos/asf/incubator-rya/blob/ce4a10ff/extras/rya.manual/src/site/markdown/quickstart.md
----------------------------------------------------------------------
diff --git a/extras/rya.manual/src/site/markdown/quickstart.md b/extras/rya.manual/src/site/markdown/quickstart.md
new file mode 100644
index 0000000..52bc111
--- /dev/null
+++ b/extras/rya.manual/src/site/markdown/quickstart.md
@@ -0,0 +1,41 @@
+# Quick Start
+
+This tutorial will outline the steps needed to get quickly started with the Rya store using the web based endpoint.
+
+## Prerequisites
+
+* Columnar Store (Accumulo)
+* Rya code (Git: git://git.apache.org/incubator-rya.git)
+* Maven 3.0 +
+
+## Building from Source
+
+Using Git, pull down the latest code from the url above.
+
+Run the command to build the code `mvn clean install`
+
+If all goes well, the build should be successful and a war should be produced in `web/web.rya/target/web.rya.war`
+
+## Deployment Using Tomcat
+
+Unwar the above war into the webapps directory.
+
+To point the web.rya war to the appropriate Accumulo instance, make a properties file `environment.properties` and put it in the classpath. Here is an example:
+
+```
+instance.name=accumulo  #Accumulo instance name
+instance.zk=localhost:2181  #Accumulo Zookeepers
+instance.username=root  #Accumulo username
+instance.password=secret  #Accumulo pwd
+rya.tableprefix=triplestore_  #Rya Table Prefix
+rya.displayqueryplan=true  #To display the query plan
+```
+
+Start the Tomcat server. `./bin/startup.sh`
+
+## Usage
+
+First, we need to load data. See the [Load Data Section] (loaddata.md)
+
+Second, we need to query that data. See the [Query Data Section](querydata.md)
+

http://git-wip-us.apache.org/repos/asf/incubator-rya/blob/ce4a10ff/extras/rya.manual/src/site/markdown/sm-addauth.md
----------------------------------------------------------------------
diff --git a/extras/rya.manual/src/site/markdown/sm-addauth.md b/extras/rya.manual/src/site/markdown/sm-addauth.md
new file mode 100644
index 0000000..aadef07
--- /dev/null
+++ b/extras/rya.manual/src/site/markdown/sm-addauth.md
@@ -0,0 +1,98 @@
+# Add Authentication
+
+This tutorial will give a few examples on how to load and query data with authentication.
+
+This is only available for accumulo and Accumulo because they provide the security filters necessary to do row level authentication and visibility.
+
+## Load Data with Visibilities
+
+During the Load process, there are a few ways to set the Column Visibility you want set on each of the corresponding rdf rows.
+
+### Global Visibility
+
+You can set the Column Visibility globally on the RdfCloudTripleStore, and it will use that particular value for every row saved.
+
+To do this, once you create and set up the RdfCloudTripleStore, just set the property on the store configuration:
+
+``` JAVA
+//setup
+final RdfCloudTripleStore store = new RdfCloudTripleStore();
+AccumuloRyaDAO crdfdao = new AccumuloRyaDAO();
+crdfdao.setConnector(connector);
+
+AccumuloRdfConfiguration conf = new AccumuloRdfConfiguration();
+conf.setTablePrefix("rts_");
+conf.setDisplayQueryPlan(true);
+
+//set global column Visibility
+conf.setCv("AUTH1|AUTH2");
+
+crdfdao.setConf(conf);
+store.setRdfDao(crdfdao);
+```
+
+The format is simply the same as the Column Visibility format.
+
+### Per triple or document based Visibility
+
+TODO: Not available as of yet
+
+## Query Data with Authentication
+
+Attaching an Authentication to the query process is very simple. It requires just adding the property `RdfCloudTripleStoreConfiguration.CONF_QUERY_AUTH` to the query `BindingSet`
+Example:
+
+``` JAVA
+//setup
+Connector connector = new ZooKeeperInstance("instance", "zoo1,zoo2,zoo3").getConnector("user", "password");
+final RdfCloudTripleStore store = new RdfCloudTripleStore();
+AccumuloRyaDAO crdfdao = new AccumuloRyaDAO();
+crdfdao.setConnector(connector);
+
+AccumuloRdfConfiguration conf = new AccumuloRdfConfiguration();
+conf.setTablePrefix("rts_");
+conf.setDisplayQueryPlan(true);
+crdfdao.setConf(conf);
+//set global column Visibility
+conf.setCv("1|2");
+store.setRdfDao(crdfdao);
+
+InferenceEngine inferenceEngine = new InferenceEngine();
+inferenceEngine.setRdfDao(crdfdao);
+inferenceEngine.setConf(conf);
+store.setInferenceEngine(inferenceEngine);
+
+Repository myRepository = new RyaSailRepository(store);
+myRepository.initialize();
+RepositoryConnection conn = myRepository.getConnection();
+
+//define and add statement
+String litdupsNS = "urn:test:litdups#";
+URI cpu = vf.createURI(litdupsNS, "cpu");
+URI loadPerc = vf.createURI(litdupsNS, "loadPerc");
+URI uri1 = vf.createURI(litdupsNS, "uri1");
+conn.add(cpu, loadPerc, uri1);
+conn.commit();
+
+//query with auth
+String query = "select * where {" +
+                "<" + cpu.toString() + "> ?p ?o1." +
+                "}";
+TupleQuery tupleQuery = conn.prepareTupleQuery(QueryLanguage.SPARQL, query);
+tupleQuery.setBinding(RdfCloudTripleStoreConfiguration.CONF_QUERY_AUTH, vf.createLiteral("2"));
+TupleQueryResult result = tupleQuery.evaluate();
+while(result.hasNext()) {
+    System.out.println(result.next());
+}
+result.close();
+
+//close
+conn.close();
+myRepository.shutDown();
+```
+
+Or you can set a global auth using the configuration:
+
+``` JAVA
+conf.setAuth("2")
+```
\ No newline at end of file

http://git-wip-us.apache.org/repos/asf/incubator-rya/blob/ce4a10ff/extras/rya.manual/src/site/markdown/sm-firststeps.md
----------------------------------------------------------------------
diff --git a/extras/rya.manual/src/site/markdown/sm-firststeps.md b/extras/rya.manual/src/site/markdown/sm-firststeps.md
new file mode 100644
index 0000000..c08c035
--- /dev/null
+++ b/extras/rya.manual/src/site/markdown/sm-firststeps.md
@@ -0,0 +1,59 @@
+# Typical First Steps
+
+In this tutorial, I will give you a quick overview of some of the first steps I perform to get data loaded and read for query.
+
+## Prerequisites
+
+ We are assuming Accumulo 1.5+ usage here.
+
+ * Rya Source Code `web.rya.war`)
+ * Accumulo on top of Hadoop 0.20+
+ * RDF Data (in N-Triples format, this format is the easiest to bulk load)
+
+## Building Source
+
+Skip this section if you already have the Map Reduce artifact and the WAR
+
+See the [Build From Source Section](build-source.md) to get the appropriate artifacts built
+
+## Load Data
+
+I find that the best way to load the data is through the Bulk Load Map Reduce job.
+
+* Save the RDF Data above onto HDFS. From now on we will refer to this location as `<RDF_HDFS_LOCATION>`
+* Move the `accumulo.rya-<version>-job.jar` onto the hadoop cluster
+* Bulk load the data. Here is a sample command line:
+
+```
+hadoop jar ../accumulo.rya-2.0.0-SNAPSHOT-job.jar BulkNtripsInputTool -Drdf.tablePrefix=lubm_ -Dcb.username=user -Dcb.pwd=cbpwd -Dcb.instance=instance -Dcb.zk=zookeeperLocation -Drdf.format=N-Triples <RDF_HDFS_LOCATION>
+```
+
+Once the data is loaded, it is actually a good practice to compact your tables. You can do this by opening the accumulo shell `shell` and running the `compact` command on the generated tables. Remember the generated tables will be prefixed by the `rdf.tablePrefix` property you assigned above. The default tablePrefix is `rts`.
+
+Here is a sample accumulo shell command:
+
+```
+compact -p lubm_(.*)
+```
+
+See the [Load Data Section](loaddata.md) for more options on loading rdf data
+
+## Run the Statistics Optimizer
+
+For the best query performance, it is recommended to run the Statistics Optimizer to create the Evaluation Statistics table. This job will read through your data and gather statistics on the distribution of the dataset. This table is then queried before query execution to reorder queries based on the data distribution.
+
+See the [Evaluation Statistics Table Section](eval.md) on how to do this.
+
+## Query data
+
+I find the easiest way to query is just to use the WAR. Load the WAR into your favorite web application container and go to the sparqlQuery.jsp page. Example:
+
+```
+http://localhost:8080/web.rya/sparqlQuery.jsp
+```
+
+This page provides a very simple text box for running queries against the store and getting data back. (SPARQL queries)
+
+Remember to update the connection information in the WAR: `WEB-INF/spring/spring-accumulo.xml`
+
+See the [Query data section](querydata.md) for more information.
\ No newline at end of file

http://git-wip-us.apache.org/repos/asf/incubator-rya/blob/ce4a10ff/extras/rya.manual/src/site/markdown/sm-infer.md
----------------------------------------------------------------------
diff --git a/extras/rya.manual/src/site/markdown/sm-infer.md b/extras/rya.manual/src/site/markdown/sm-infer.md
new file mode 100644
index 0000000..712af78
--- /dev/null
+++ b/extras/rya.manual/src/site/markdown/sm-infer.md
@@ -0,0 +1,332 @@
+# Inferencing
+
+Rya currently provides simple inferencing. The supported list of inferred relationships include:
+
+- rdfs:subClassOf
+- rdfs:subPropertyOf
+- owl:EquivalentProperty
+- owl:inverseOf
+- owl:SymmetricProperty
+- owl:TransitiveProperty (This is currently in beta and will not work for every case)
+- owl:sameAs
+
+## Setup
+
+The Inferencing Engine is a scheduled job that runs by default every 5 minutes, this is configurable, to query the relationships in the store and develop the inferred graphs necessary to answer inferencing questions.
+
+This also means that if you load a model into the store, it could take up to 5 minutes for the inferred relationships to be available.
+
+As usual you will need to set up your `RdfCloudTripleStore` with the correct DAO, notice we add an `InferencingEngine` as well to the store. If this is not added, then no inferencing will be done on the queries:
+
+``` JAVA
+//setup
+Connector connector = new ZooKeeperInstance("instance", "zoo1,zoo2,zoo3").getConnector("user", "password");
+final RdfCloudTripleStore store = new RdfCloudTripleStore();
+AccumuloRyaDAO crdfdao = new AccumuloRyaDAO();
+crdfdao.setConnector(connector);
+
+AccumuloRdfConfiguration conf = new AccumuloRdfConfiguration();
+conf.setTablePrefix("rts_");
+conf.setDisplayQueryPlan(true);
+crdfdao.setConf(conf);
+store.setRdfDao(crdfdao);
+
+ProspectorServiceEvalStatsDAO evalDao = new ProspectorServiceEvalStatsDAO(connector, conf);
+evalDao.init();
+store.setRdfEvalStatsDAO(evalDao);
+
+InferenceEngine inferenceEngine = new InferenceEngine();
+inferenceEngine.setRdfDao(crdfdao);
+inferenceEngine.setConf(conf);
+store.setInferenceEngine(inferenceEngine);
+
+Repository myRepository = new RyaSailRepository(store);
+myRepository.initialize();
+RepositoryConnection conn = myRepository.getConnection();
+
+//query code goes here
+
+//close
+conn.close();
+myRepository.shutDown();
+```
+
+## Samples
+
+We will go through some quick samples on loading inferred relationships, seeing and diagnosing the query plan, and checking the data
+
+### Rdfs:SubClassOf
+
+First the code, which will load the following subclassof relationship: `UndergraduateStudent subclassof Student subclassof Person`. Then we will load into the tables three triples defining `UgradA rdf:type UndergraduateStudent, StudentB rdf:type Student, PersonC rdf:type Person`
+
+``` JAVA
+conn.add(new StatementImpl(vf.createURI(litdupsNS, "UndergraduateStudent"), RDFS.SUBCLASSOF, vf.createURI(litdupsNS, "Student")));
+conn.add(new StatementImpl(vf.createURI(litdupsNS, "Student"), RDFS.SUBCLASSOF, vf.createURI(litdupsNS, "Person")));
+conn.add(new StatementImpl(vf.createURI(litdupsNS, "UgradA"), RDF.TYPE, vf.createURI(litdupsNS, "UndergraduateStudent")));
+conn.add(new StatementImpl(vf.createURI(litdupsNS, "StudentB"), RDF.TYPE, vf.createURI(litdupsNS, "Student")));
+conn.add(new StatementImpl(vf.createURI(litdupsNS, "PersonC"), RDF.TYPE, vf.createURI(litdupsNS, "Person")));
+conn.commit();
+```
+
+Remember that once the model is committed, it may take up to 5 minutes for the inferred relationships to be ready. Though you can override this property in the `InferencingEngine`.
+
+We shall run the following query:
+
+```
+PREFIX rdfs: <http://www.w3.org/2000/01/rdf-schema#>
+PREFIX rdf: <http://www.w3.org/1999/02/22-rdf-syntax-ns#>
+PREFIX lit: <urn:test:litdups#>
+select * where {?s rdf:type lit:Person.}
+```
+
+And should get back the following results:
+
+```
+[s=urn:test:litdups#StudentB]
+[s=urn:test:litdups#PersonC]
+[s=urn:test:litdups#UgradA]
+```
+
+#### How it works
+
+Let us look at the query plan:
+
+```
+QueryRoot
+   Projection
+      ProjectionElemList
+         ProjectionElem "s"
+      Join
+         FixedStatementPattern
+            Var (name=79f261ee-e930-4af1-bc09-e637cc0affef)
+            Var (name=c-79f261ee-e930-4af1-bc09-e637cc0affef, value=http://www.w3.org/2000/01/rdf-schema#subClassOf)
+            Var (name=-const-2, value=urn:test:litdups#Person, anonymous)
+         DoNotExpandSP
+            Var (name=s)
+            Var (name=-const-1, value=http://www.w3.org/1999/02/22-rdf-syntax-ns#type, anonymous)
+            Var (name=79f261ee-e930-4af1-bc09-e637cc0affef)
+```
+
+Basically, we first find out (through the InferencingEngine) what triples have subclassof with Person. The InferencingEngine will do the graph analysis to find the both Student and UndergraduateStudent are Person classes.
+Then this information is joined with the statement pattern `(?s rdf:type ?inf)` where `?inf` is the results from the InferencingEngine.
+
+### Rdfs:SubPropertyOf
+
+SubPropertyOf defines that a property can be an instance of another property. For example, a `gradDegreeFrom subPropertyOf degreeFrom`.
+
+Also, EquivalentProperty can be thought of as specialized SubPropertyOf relationship where if `propA equivalentProperty propB` then that means that `propA subPropertyOf propB AND propB subPropertyOf propA`
+
+Sample Code:
+
+``` JAVA
+conn.add(new StatementImpl(vf.createURI(litdupsNS, "undergradDegreeFrom"), RDFS.SUBPROPERTYOF, vf.createURI(litdupsNS, "degreeFrom")));
+conn.add(new StatementImpl(vf.createURI(litdupsNS, "gradDegreeFrom"), RDFS.SUBPROPERTYOF, vf.createURI(litdupsNS, "degreeFrom")));
+conn.add(new StatementImpl(vf.createURI(litdupsNS, "degreeFrom"), RDFS.SUBPROPERTYOF, vf.createURI(litdupsNS, "memberOf")));
+conn.add(new StatementImpl(vf.createURI(litdupsNS, "memberOf"), RDFS.SUBPROPERTYOF, vf.createURI(litdupsNS, "associatedWith")));
+conn.add(new StatementImpl(vf.createURI(litdupsNS, "UgradA"), vf.createURI(litdupsNS, "undergradDegreeFrom"), vf.createURI(litdupsNS, "Harvard")));
+conn.add(new StatementImpl(vf.createURI(litdupsNS, "GradB"), vf.createURI(litdupsNS, "gradDegreeFrom"), vf.createURI(litdupsNS, "Yale")));
+conn.add(new StatementImpl(vf.createURI(litdupsNS, "ProfessorC"), vf.createURI(litdupsNS, "memberOf"), vf.createURI(litdupsNS, "Harvard")));
+conn.commit();
+```
+
+With query:
+
+```
+PREFIX rdfs: <http://www.w3.org/2000/01/rdf-schema#>
+PREFIX rdf: <http://www.w3.org/1999/02/22-rdf-syntax-ns#>
+PREFIX lit: <urn:test:litdups#>
+select * where {?s lit:memberOf lit:Harvard.}
+```
+
+Will return results:
+
+```
+[s=urn:test:litdups#UgradA]
+[s=urn:test:litdups#ProfessorC]
+```
+
+Since UgradA has undergraduateDegreeFrom Harvard and ProfessorC is memberOf Harvard.
+
+#### How it works
+
+This is very similar to the subClassOf relationship above. Basically the InferencingEngine provides what properties are subPropertyOf relationships with memberOf, and the second part of the Join checks to see if those properties are predicates with object "Harvard".
+
+Query Plan:
+
+```
+QueryRoot
+   Projection
+      ProjectionElemList
+         ProjectionElem "s"
+      Join
+         FixedStatementPattern
+            Var (name=0bad69f3-4769-4293-8318-e828b23dc52a)
+            Var (name=c-0bad69f3-4769-4293-8318-e828b23dc52a, value=http://www.w3.org/2000/01/rdf-schema#subPropertyOf)
+            Var (name=-const-1, value=urn:test:litdups#memberOf, anonymous)
+         DoNotExpandSP
+            Var (name=s)
+            Var (name=0bad69f3-4769-4293-8318-e828b23dc52a)
+            Var (name=-const-2, value=urn:test:litdups#Harvard, anonymous)
+```
+
+### InverseOf
+
+InverseOf defines a property that is an inverse relation of another property. For example, a student who has a `degreeFrom` a University also means that the University `hasAlumnus` student.
+
+Code:
+
+``` JAVA
+conn.add(new StatementImpl(vf.createURI(litdupsNS, "degreeFrom"), OWL.INVERSEOF, vf.createURI(litdupsNS, "hasAlumnus")));
+conn.add(new StatementImpl(vf.createURI(litdupsNS, "UgradA"), vf.createURI(litdupsNS, "degreeFrom"), vf.createURI(litdupsNS, "Harvard")));
+conn.add(new StatementImpl(vf.createURI(litdupsNS, "GradB"), vf.createURI(litdupsNS, "degreeFrom"), vf.createURI(litdupsNS, "Harvard")));
+conn.add(new StatementImpl(vf.createURI(litdupsNS, "Harvard"), vf.createURI(litdupsNS, "hasAlumnus"), vf.createURI(litdupsNS, "AlumC")));
+conn.commit();
+```
+
+Query:
+
+```
+PREFIX rdfs: <http://www.w3.org/2000/01/rdf-schema#>
+PREFIX rdf: <http://www.w3.org/1999/02/22-rdf-syntax-ns#>
+PREFIX lit: <urn:test:litdups#>
+select * where {lit:Harvard lit:hasAlumnus ?s.}
+```
+
+Result:
+
+```
+[s=urn:test:litdups#AlumC]
+[s=urn:test:litdups#GradB]
+[s=urn:test:litdups#UgradA]
+```
+
+#### How it works
+
+The query planner will expand the statement pattern `Harvard hasAlumnus ?s` to a Union between `Harvard hasAlumnus ?s. and ?s degreeFrom Harvard`
+
+As a caveat, it is important to note that in general Union queries do not have the best performance, so having a property that has an inverseOf and subPropertyOf, could cause a query plan that might take long depending on how the query planner orders the joins.
+
+Query Plan
+
+```
+QueryRoot
+   Projection
+      ProjectionElemList
+         ProjectionElem "s"
+      InferUnion
+         StatementPattern
+            Var (name=-const-1, value=urn:test:litdups#Harvard, anonymous)
+            Var (name=-const-2, value=urn:test:litdups#hasAlumnus, anonymous)
+            Var (name=s)
+         StatementPattern
+            Var (name=s)
+            Var (name=-const-2, value=urn:test:litdups#degreeFrom)
+            Var (name=-const-1, value=urn:test:litdups#Harvard, anonymous)
+```
+
+### SymmetricProperty
+
+SymmetricProperty defines a relationship where, for example, if Bob is a friendOf Jeff, then Jeff is a friendOf Bob. (Hopefully)
+
+Code:
+
+``` JAVA
+conn.add(new StatementImpl(vf.createURI(litdupsNS, "friendOf"), RDF.TYPE, OWL.SYMMETRICPROPERTY));
+conn.add(new StatementImpl(vf.createURI(litdupsNS, "Bob"), vf.createURI(litdupsNS, "friendOf"), vf.createURI(litdupsNS, "Jeff")));
+conn.add(new StatementImpl(vf.createURI(litdupsNS, "James"), vf.createURI(litdupsNS, "friendOf"), vf.createURI(litdupsNS, "Jeff")));
+conn.commit();
+```
+
+Query:
+
+```
+PREFIX rdfs: <http://www.w3.org/2000/01/rdf-schema#>
+PREFIX rdf: <http://www.w3.org/1999/02/22-rdf-syntax-ns#>
+PREFIX lit: <urn:test:litdups#>
+select * where {?s lit:friendOf lit:Bob.}
+```
+
+Results:
+
+```
+[s=urn:test:litdups#Jeff]
+```
+
+#### How it works
+
+The query planner will recognize that `friendOf` is a SymmetricProperty and devise a Union to find the specified relationship and inverse.
+
+Query Plan:
+
+```
+QueryRoot
+   Projection
+      ProjectionElemList
+         ProjectionElem "s"
+      InferUnion
+         StatementPattern
+            Var (name=s)
+            Var (name=-const-1, value=urn:test:litdups#friendOf, anonymous)
+            Var (name=-const-2, value=urn:test:litdups#Bob, anonymous)
+         StatementPattern
+            Var (name=-const-2, value=urn:test:litdups#Bob, anonymous)
+            Var (name=-const-1, value=urn:test:litdups#friendOf, anonymous)
+            Var (name=s)
+```
+
+### TransitiveProperty
+
+TransitiveProperty provides a transitive relationship between resources. For example, if Queens is subRegionOf NYC and NYC is subRegionOf NY, then Queens is transitively a subRegionOf NY.
+
+Code:
+
+``` JAVA
+conn.add(new StatementImpl(vf.createURI(litdupsNS, "subRegionOf"), RDF.TYPE, OWL.TRANSITIVEPROPERTY));
+conn.add(new StatementImpl(vf.createURI(litdupsNS, "Queens"), vf.createURI(litdupsNS, "subRegionOf"), vf.createURI(litdupsNS, "NYC")));
+conn.add(new StatementImpl(vf.createURI(litdupsNS, "NYC"), vf.createURI(litdupsNS, "subRegionOf"), vf.createURI(litdupsNS, "NY")));
+conn.add(new StatementImpl(vf.createURI(litdupsNS, "NY"), vf.createURI(litdupsNS, "subRegionOf"), vf.createURI(litdupsNS, "US")));
+conn.add(new StatementImpl(vf.createURI(litdupsNS, "US"), vf.createURI(litdupsNS, "subRegionOf"), vf.createURI(litdupsNS, "NorthAmerica")));
+conn.add(new StatementImpl(vf.createURI(litdupsNS, "NorthAmerica"), vf.createURI(litdupsNS, "subRegionOf"), vf.createURI(litdupsNS, "World")));
+conn.commit();
+```
+
+Query:
+
+```
+PREFIX rdfs: <http://www.w3.org/2000/01/rdf-schema#>
+PREFIX rdf: <http://www.w3.org/1999/02/22-rdf-syntax-ns#>
+PREFIX lit: <urn:test:litdups#>
+select * where {?s lit:subRegionOf lit:NorthAmerica.}
+```
+
+Results:
+
+```
+[s=urn:test:litdups#Queens]
+[s=urn:test:litdups#NYC]
+[s=urn:test:litdups#NY]
+[s=urn:test:litdups#US]
+```
+
+#### How it works
+
+The TransitiveProperty relationship works by running recursive queries till all the results are returned.
+
+It is important to note that certain TransitiveProperty relationships will not work:
+* Open ended property: ?s subRegionOf ?o (At least one of the properties must be filled or will be filled as the query gets answered)
+* Closed property: Queens subRegionOf NY (At least one of the properties must be empty)
+
+We are working on fixing these issues.
+
+Query Plan:
+
+```
+QueryRoot
+   Projection
+      ProjectionElemList
+         ProjectionElem "s"
+      TransitivePropertySP
+         Var (name=s)
+         Var (name=-const-1, value=urn:test:litdups#subRegionOf, anonymous)
+         Var (name=-const-2, value=urn:test:litdups#NorthAmerica, anonymous)
+```
\ No newline at end of file

http://git-wip-us.apache.org/repos/asf/incubator-rya/blob/ce4a10ff/extras/rya.manual/src/site/markdown/sm-namedgraph.md
----------------------------------------------------------------------
diff --git a/extras/rya.manual/src/site/markdown/sm-namedgraph.md b/extras/rya.manual/src/site/markdown/sm-namedgraph.md
new file mode 100644
index 0000000..f869adc
--- /dev/null
+++ b/extras/rya.manual/src/site/markdown/sm-namedgraph.md
@@ -0,0 +1,136 @@
+# Named Graphs
+
+Named graphs are supported simply in the Rdf Store in a few ways. OpenRdf supports sending `contexts` as each triple is saved.
+
+## Simple Named Graph Load and Query
+
+Here is a very simple example of using the API to Insert data in named graphs and querying with Sparql
+
+First we will define a Trig document to load
+Trig document
+
+```
+@prefix rdf: <http://www.w3.org/1999/02/22-rdf-syntax-ns#> .
+@prefix xsd: <http://www.w3.org/2001/XMLSchema#> .
+@prefix swp: <http://www.w3.org/2004/03/trix/swp-1/> .
+@prefix dc: <http://purl.org/dc/elements/1.1/> .
+@prefix ex: <http://www.example.org/vocabulary#> .
+@prefix : <http://www.example.org/exampleDocument#> .
+:G1 { :Monica ex:name "Monica Murphy" .
+      :Monica ex:homepage <http://www.monicamurphy.org> .
+      :Monica ex:email <ma...@monicamurphy.org> .
+      :Monica ex:hasSkill ex:Management }
+
+:G2 { :Monica rdf:type ex:Person .
+      :Monica ex:hasSkill ex:Programming }
+
+:G4 { :Phobe ex:name "Phobe Buffet" }
+
+:G3 { :G1 swp:assertedBy _:w1 .
+      _:w1 swp:authority :Chris .
+      _:w1 dc:date "2003-10-02"^^xsd:date .
+      :G2 swp:quotedBy _:w2 .
+      :G4 swp:assertedBy _:w2 .
+      _:w2 dc:date "2003-09-03"^^xsd:date .
+      _:w2 swp:authority :Tom .
+      :Chris rdf:type ex:Person .
+      :Chris ex:email <ma...@bizer.de>.
+      :Tom rdf:type ex:Person .
+      :Tom ex:email <ma...@bizer.de>}
+```
+
+We will assume that this file is saved on your classpath somewhere at `<TRIG_FILE>`
+
+Load data through API:
+
+``` JAVA
+InputStream stream = Thread.currentThread().getContextClassLoader().getResourceAsStream("namedgraphs.trig");
+RepositoryConnection conn = repository.getConnection();
+conn.add(stream, "", RDFFormat.TRIG);
+conn.commit();
+```
+
+Now that the data is loaded we can easily query it. For example, we will query to find what `hasSkill` is defined in graph G2, and relate that to someone defined in G1.
+
+**Query:**
+
+```
+PREFIX  ex:  <http://www.example.org/exampleDocument#>
+PREFIX  voc:  <http://www.example.org/vocabulary#>
+PREFIX  foaf:  <http://xmlns.com/foaf/0.1/>
+PREFIX  rdfs:  <http://www.w3.org/2000/01/rdf-schema#>
+
+SELECT *
+WHERE
+{
+  GRAPH ex:G1
+  {
+    ?m voc:name ?name ;
+       voc:homepage ?hp .
+  } .
+ GRAPH ex:G2
+  {
+    ?m voc:hasSkill ?skill .
+  } .
+}
+```
+
+**Results:**
+
+```
+[hp=http://www.monicamurphy.org;m=http://www.example.org/exampleDocument#Monica;skill=http://www.example.org/vocabulary#Programming;name="Monica Murphy"]
+```
+
+**Here is the Query Plan as well:**
+
+```
+QueryRoot
+   Projection
+      ProjectionElemList
+         ProjectionElem "m"
+         ProjectionElem "name"
+         ProjectionElem "hp"
+         ProjectionElem "skill"
+      Join
+         Join
+            StatementPattern FROM NAMED CONTEXT
+               Var (name=m)
+               Var (name=-const-2, value=http://www.example.org/vocabulary#name, anonymous)
+               Var (name=name)
+               Var (name=-const-1, value=http://www.example.org/exampleDocument#G1, anonymous)
+            StatementPattern FROM NAMED CONTEXT
+               Var (name=m)
+               Var (name=-const-3, value=http://www.example.org/vocabulary#homepage, anonymous)
+               Var (name=hp)
+               Var (name=-const-1, value=http://www.example.org/exampleDocument#G1, anonymous)
+         StatementPattern FROM NAMED CONTEXT
+            Var (name=m)
+            Var (name=-const-5, value=http://www.example.org/vocabulary#hasSkill, anonymous)
+            Var (name=skill)
+            Var (name=-const-4, value=http://www.example.org/exampleDocument#G2, anonymous)
+```
+
+## Inserting named graph data through Sparql
+
+The new Sparql update standard provides another way to insert data, even into named graphs.
+
+First the insert update:
+
+```
+PREFIX dc: <http://purl.org/dc/elements/1.1/>
+PREFIX ex: <http://example/addresses#>
+INSERT DATA
+{
+    GRAPH ex:G1 {
+        <http://example/book3> dc:title    "A new book" ;
+                               dc:creator  "A.N.Other" .
+    }
+}
+```
+
+To perform this update, it requires different code than querying the data directly:
+
+```
+Update update = conn.prepareUpdate(QueryLanguage.SPARQL, insert);
+update.execute();
+```
\ No newline at end of file

http://git-wip-us.apache.org/repos/asf/incubator-rya/blob/ce4a10ff/extras/rya.manual/src/site/markdown/sm-simpleaqr.md
----------------------------------------------------------------------
diff --git a/extras/rya.manual/src/site/markdown/sm-simpleaqr.md b/extras/rya.manual/src/site/markdown/sm-simpleaqr.md
new file mode 100644
index 0000000..a509195
--- /dev/null
+++ b/extras/rya.manual/src/site/markdown/sm-simpleaqr.md
@@ -0,0 +1,54 @@
+# Simple Add Query and Remove of Statements
+
+This quick tutorial will give a small example on how to add, query, and remove statements from Rya
+
+## Code
+
+``` JAVA
+//setup
+Connector connector = new ZooKeeperInstance("instance", "zoo1,zoo2,zoo3").getConnector("user", "password");
+final RdfCloudTripleStore store = new RdfCloudTripleStore();
+AccumuloRyaDAO crdfdao = new AccumuloRyaDAO();
+crdfdao.setConnector(connector);
+
+AccumuloRdfConfiguration conf = new AccumuloRdfConfiguration();
+conf.setTablePrefix("rts_");
+conf.setDisplayQueryPlan(true);
+crdfdao.setConf(conf);
+store.setRdfDao(crdfdao);
+
+ProspectorServiceEvalStatsDAO evalDao = new ProspectorServiceEvalStatsDAO(connector, conf);
+evalDao.init();
+store.setRdfEvalStatsDAO(evalDao);
+
+InferenceEngine inferenceEngine = new InferenceEngine();
+inferenceEngine.setRdfDao(crdfdao);
+inferenceEngine.setConf(conf);
+store.setInferenceEngine(inferenceEngine);
+
+Repository myRepository = new RyaSailRepository(store);
+myRepository.initialize();
+RepositoryConnection conn = myRepository.getConnection();
+
+//define and add statement
+String litdupsNS = "urn:test:litdups#";
+URI cpu = vf.createURI(litdupsNS, "cpu");
+URI loadPerc = vf.createURI(litdupsNS, "loadPerc");
+URI uri1 = vf.createURI(litdupsNS, "uri1");
+conn.add(cpu, loadPerc, uri1);
+conn.commit();
+
+//query for all statements that have subject=cpu and pred=loadPerc (wildcard object)
+RepositoryResult<Statement> result = conn.getStatements(cpu, loadPerc, null, true)
+while(result.hasNext()) {
+    System.out.println(result.next());
+}
+result.close();
+
+//remove statement
+conn.remove(cpu, loadPerc, uri1);
+
+//close
+conn.close();
+myRepository.shutDown();
+```
\ No newline at end of file

http://git-wip-us.apache.org/repos/asf/incubator-rya/blob/ce4a10ff/extras/rya.manual/src/site/markdown/sm-sparqlquery.md
----------------------------------------------------------------------
diff --git a/extras/rya.manual/src/site/markdown/sm-sparqlquery.md b/extras/rya.manual/src/site/markdown/sm-sparqlquery.md
new file mode 100644
index 0000000..cff732a
--- /dev/null
+++ b/extras/rya.manual/src/site/markdown/sm-sparqlquery.md
@@ -0,0 +1,58 @@
+# Simple Add Query and Remove of Statements
+
+This quick tutorial will give a small example on how to query data with SPARQL
+
+## Code
+
+``` JAVA
+//setup
+Connector connector = new ZooKeeperInstance("instance", "zoo1,zoo2,zoo3").getConnector("user", "password");
+final RdfCloudTripleStore store = new RdfCloudTripleStore();
+AccumuloRyaDAO crdfdao = new AccumuloRyaDAO();
+crdfdao.setConnector(connector);
+
+AccumuloRdfConfiguration conf = new AccumuloRdfConfiguration();
+conf.setTablePrefix("rts_");
+conf.setDisplayQueryPlan(true);
+crdfdao.setConf(conf);
+store.setRdfDao(crdfdao);
+
+ProspectorServiceEvalStatsDAO evalDao = new ProspectorServiceEvalStatsDAO(connector, conf);
+evalDao.init();
+store.setRdfEvalStatsDAO(evalDao);
+
+InferenceEngine inferenceEngine = new InferenceEngine();
+inferenceEngine.setRdfDao(crdfdao);
+inferenceEngine.setConf(conf);
+store.setInferenceEngine(inferenceEngine);
+
+Repository myRepository = new RyaSailRepository(store);
+myRepository.initialize();
+RepositoryConnection conn = myRepository.getConnection();
+
+//define and add statements
+String litdupsNS = "urn:test:litdups#";
+URI cpu = vf.createURI(litdupsNS, "cpu");
+URI loadPerc = vf.createURI(litdupsNS, "loadPerc");
+URI uri1 = vf.createURI(litdupsNS, "uri1");
+URI pred2 = vf.createURI(litdupsNS, "pred2");
+URI uri2 = vf.createURI(litdupsNS, "uri2");
+conn.add(cpu, loadPerc, uri1);
+conn.commit();
+
+//query using sparql
+String query = "select * where {" +
+                "?x <" + loadPerc.stringValue() + "> ?o1." +
+                "?x <" + pred2.stringValue() + "> ?o2." +
+                "}";
+TupleQuery tupleQuery = conn.prepareTupleQuery(QueryLanguage.SPARQL, query);
+TupleQueryResult result = tupleQuery.evaluate();
+while(result.hasNext()) {
+      System.out.println(result.next());
+}
+result.close();
+
+//close
+conn.close();
+myRepository.shutDown();
+```
\ No newline at end of file

http://git-wip-us.apache.org/repos/asf/incubator-rya/blob/ce4a10ff/extras/rya.manual/src/site/markdown/sm-updatedata.md
----------------------------------------------------------------------
diff --git a/extras/rya.manual/src/site/markdown/sm-updatedata.md b/extras/rya.manual/src/site/markdown/sm-updatedata.md
new file mode 100644
index 0000000..7227f54
--- /dev/null
+++ b/extras/rya.manual/src/site/markdown/sm-updatedata.md
@@ -0,0 +1,62 @@
+# Sparql Update
+
+OpenRDF supports the Sparql Update functionality. Here are a few samples:
+
+Remember, you have to use `RepositoryConnection.prepareUpdate(..)` to perform these queries
+
+**Insert:**
+
+```
+PREFIX dc: <http://purl.org/dc/elements/1.1/>
+INSERT DATA
+{ <http://example/book3> dc:title    "A new book" ;
+                         dc:creator  "A.N.Other" .
+}
+```
+
+**Delete:**
+
+```
+PREFIX dc: <http://purl.org/dc/elements/1.1/>
+DELETE DATA
+{ <http://example/book3> dc:title    "A new book" ;
+                         dc:creator  "A.N.Other" .
+}
+```
+
+**Update:**
+
+```
+PREFIX dc: <http://purl.org/dc/elements/1.1/>
+DELETE { ?book dc:title ?title }
+INSERT { ?book dc:title "A newer book".         ?book dc:add "Additional Info" }
+WHERE
+  { ?book dc:creator "A.N.Other" .
+  }
+```
+
+**Insert Named Graph:**
+
+```
+PREFIX dc: <http://purl.org/dc/elements/1.1/>
+PREFIX ex: <http://example/addresses#>
+INSERT DATA
+{ GRAPH ex:G1 {
+<http://example/book3> dc:title    "A new book" ;
+                         dc:creator  "A.N.Other" .
+}
+}
+```
+
+**Update Named Graph:**
+
+```
+PREFIX dc: <http://purl.org/dc/elements/1.1/>
+WITH <http://example/addresses#G1>
+DELETE { ?book dc:title ?title }
+INSERT { ?book dc:title "A newer book".
+         ?book dc:add "Additional Info" }
+WHERE
+  { ?book dc:creator "A.N.Other" .
+  }
+```

http://git-wip-us.apache.org/repos/asf/incubator-rya/blob/ce4a10ff/extras/rya.manual/src/site/resources/js/fixmarkdownlinks.js
----------------------------------------------------------------------
diff --git a/extras/rya.manual/src/site/resources/js/fixmarkdownlinks.js b/extras/rya.manual/src/site/resources/js/fixmarkdownlinks.js
new file mode 100644
index 0000000..484c5d3
--- /dev/null
+++ b/extras/rya.manual/src/site/resources/js/fixmarkdownlinks.js
@@ -0,0 +1,6 @@
+window.onload = function() {
+    var anchors = document.getElementsByTagName("a");
+        for (var i = 0; i < anchors.length; i++) {
+            anchors[i].href = anchors[i].href.replace(/\.md$/,'\.html');
+        }
+    }
\ No newline at end of file

http://git-wip-us.apache.org/repos/asf/incubator-rya/blob/ce4a10ff/extras/rya.manual/src/site/site.xml
----------------------------------------------------------------------
diff --git a/extras/rya.manual/src/site/site.xml b/extras/rya.manual/src/site/site.xml
new file mode 100644
index 0000000..19078a0
--- /dev/null
+++ b/extras/rya.manual/src/site/site.xml
@@ -0,0 +1,45 @@
+<?xml version="1.0" encoding="ISO-8859-1"?>
+<project name="Maven">
+  <skin>
+    <groupId>org.apache.maven.skins</groupId>
+    <artifactId>maven-fluido-skin</artifactId>
+    <version>1.4</version>
+  </skin>
+  <custom>
+    <fluidoSkin>
+      <topBarEnabled>false</topBarEnabled>
+      <sideBarEnabled>true</sideBarEnabled>
+     <sourceLineNumbersEnabled>false</sourceLineNumbersEnabled>
+    </fluidoSkin>
+  </custom>
+  
+  <body>
+    <head>
+        <script src="./js/fixmarkdownlinks.js" type="text/javascript"></script>  
+    </head>
+    <menu name="Rya">
+        <item name="Overview" href="overview.html"/>
+        <item name="Quick Start" href="quickstart.html"/>
+        <item name="Load Data" href="loaddata.html"/>
+        <item name="Query Data" href="querydata.html"/>
+        <item name="Evaluation Table" href="eval.html"/>
+        <item name="Pre-computed Joins" href="loadPrecomputedJoin.html"/>
+        <item name="Inferencing" href="infer.html"/>
+    </menu>
+
+    <menu name="Samples">
+        <item name="Typical First Steps" href="sm-firststeps.html"/>
+        <item name="Simple Add/Query/Remove Statements" href="sm-simpleaqr.html"/>
+        <item name="Sparql query" href="sm-sparqlquery.html"/>
+        <item name="Adding Authentication" href="sm-addauth.html"/>
+        <item name="Inferencing" href="sm-infer.html"/>
+        <item name="Named Graph" href="sm-namedgraph.html"/>
+        <item name="Update data" href="sm-updatedata.html"/>
+        <item name="Alx" href="alx.html"/>
+    </menu>
+
+    <menu name="Development">
+        <item name="Building From Source" href="build-source.html"/>
+    </menu>
+  </body>
+</project>
\ No newline at end of file


[2/3] incubator-rya git commit: RYA-3 Changing Manual to Markdown; linking README

Posted by mi...@apache.org.
http://git-wip-us.apache.org/repos/asf/incubator-rya/blob/ce4a10ff/extras/rya.manual/src/main/webapp/css/style.css
----------------------------------------------------------------------
diff --git a/extras/rya.manual/src/main/webapp/css/style.css b/extras/rya.manual/src/main/webapp/css/style.css
deleted file mode 100644
index 496901e..0000000
--- a/extras/rya.manual/src/main/webapp/css/style.css
+++ /dev/null
@@ -1,1588 +0,0 @@
-/**
- * Copyright (C) 2009-2010 the original author or authors.
- * See the notice.md file distributed with this work for additional
- * information regarding copyright ownership.
- *
- * Licensed under the Apache License, Version 2.0 (the "License");
- * you may not use this file except in compliance with the License.
- * You may obtain a copy of the License at
- *
- *     http://www.apache.org/licenses/LICENSE-2.0
- *
- * Unless required by applicable law or agreed to in writing, software
- * distributed under the License is distributed on an "AS IS" BASIS,
- * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
- * See the License for the specific language governing permissions and
- * limitations under the License.
- */
-body {
-  margin: 0;
-  padding: 0;
-}
-
-#cell-0-0 { background: white url( '../images/bg02-white-left-nogr.png' ) repeat-y right top; }
-#cell-0-1 { height: 1em; width: 85%; }
-#cell-0-2 { background: transparent url( '../images/bg02-white-right-nogr.png' ) repeat-y left top; }
-#cell-1-0 { background: transparent url( '../images/bg.png' ) repeat-x top; }
-#cell-1-1 { background: transparent url( '../images/bg02-blue-left.png' ) no-repeat right top; width:13px; }
-#cell-1-2 { background: transparent url( '../images/bg.png' ) repeat-x top; width:85%; }
-#banner { height: 80px; }
-#cell-1-3 { background: transparent url( '../images/bg02-blue-right.png' ) no-repeat left top; width:13px; }
-#cell-1-4 { background: transparent url( '../images/bg.png' ) repeat-x top; }
-#cell-2-0 { background: white url( '../images/bg02-white-left-nogr.png' ) repeat-y right top; }
-#cell-2-1 { height: 1em; width: 85%; }
-#cell-2-2 { background: white url( '../images/bg02-white-right-nogr.png' ) repeat-y left top; }
-#cell-3-0 { background: transparent url( '../images/bg.png' ) repeat-x top; }
-#cell-3-1 { background: transparent url( '../images/bg02-blue-left.png' ) no-repeat right top; width:13px; }
-#cell-3-2 { background: transparent url( '../images/bg.png' ) repeat-x top; width:85%; }
-#cell-3-3 { background: transparent url( '../images/bg02-blue-right.png' ) no-repeat left top; width:13px; }
-#cell-3-4 { background: transparent url( '../images/bg.png' ) repeat-x top; }
-#cell-4-0 { background: white url( '../images/bg02-white-left-nogr.png' ) repeat-y right top; }
-#cell-4-1 { height: 1em; width: 85%; }
-#cell-4-2 { background: transparent url( '../images/bg02-white-right-nogr.png' ) repeat-y left top; }
-
-#banner {
-  margin: 0px 20px;
-  padding: 0px 50px 0px 50px;
-  font-size: 200%;
-  text-align: center;
-  background: transparent;
-}
-#top-menu {
-  margin: 0px 25px;
-  padding: 2px;
-  line-height: 100%;
-  font-family: Verdana, arial, sans-serif;
-  font-weight: bold;
-  font-size: 11px;
-  color: white;
-  line-height: 13px;
-  overflow: none;
-}
-#top-menu a:link, #top-menu a:visited {
-  color: #ffffff;
-  text-decoration: none;
-}
-#top-menu a:active, #top-menu a:hover {
-  color: #5b8fbe;
-  text-decoration: underline;
-}
-#top-menu sup img {
-  height: 0px;
-  width: 0px;
-}
-
-#footer {
-  padding: 0px 50px 0px 50px;
-  min-height: 0em;
-  text-align: center;
-  color: white;
-  font-size: 65%;
-  font-family: Verdana, arial, sans-serif;
-}
-#site-footer {
-  margin: 5px;
-}
-#footer a:link, #footer a:visited {
-  color: white;
-  text-decoration: none;
-}
-#footer a:hover {
-  color: white;
-  text-decoration: underline;
-}
-
-#wrapper-menu-page-right {
-  background: transparent url( '../images/left-box-right.png' ) repeat-y right;
-}
-
-#wrapper-menu-page-bottom {
-  background: transparent url( '../images/left-box-bottom.png' ) no-repeat right bottom;
-}
-
-#wrapper-menu-page-top {
-  background: transparent url( '../images/left-box-top.png' ) no-repeat right top;
-}
-
-#menu-page {
-  width: 200px;
-  color: #5b8fbe;
-  font-family: Verdana, arial, sans-serif;
-  font-weight: bold;
-  margin: 0px;
-  padding: 17px 11px 17px 0px;
-  text-align: left;
-  margin-left: 0px;
-  line-height: 1em;
-}
-
-#menu-page ul {
-  margin: 0;
-  padding: 0;
-  padding-left: 35px;
-  font-size: 75%;
-}
-
-#menu-page h3 {
-  font-size: 75%;
-  text-transform: uppercase;
-  margin: 1em 0 0.3em 0;
-  color: #5b8fbe;
-  font-weight: normal;
-  letter-spacing: 0.15em;
-  background: url('../images/big-bullet.png') no-repeat 0pt;
-  background-position:left center;
-  margin-left:15px;
-  padding-left: 20px;
-  font-family: Verdana, arial, sans-serif;
-  font-weight: bold;
-}
-
-#menu-page a:link, #menu-page a:visited {
-  color: black;
-}
-
-#menu-page a:hover {
-  color: #5b8fbe;
-}
-
-#content {
-  padding: 20px 0px 0px 0px;
-  display: block;
-  text-align: left;
-  height: 100%;
-  width: 100%;
-  color: black;
-}
-
-h2 {
-  font-size: 110%;
-}
-
-h3 {
-  font-size: 100%;
-}
-
-a:link, a:visited {
-  color: #5b8fbe;
-  text-decoration: none;
-}
-
-a:hover {
-  color: #666;
-  text-decoration: none;
-}
-
-
-table.align {
-  padding: 0px;
-  border: none;
-}
-
-td.align {
-  padding: 0px;
-  border: none;
-}
-
-/*********************************************************************
- CSS Elements that are part of the confluence rendered xhtml
- *********************************************************************/
-.wiki-content {
-	font-family: Verdana, arial, sans-serif;
-	font-size: 11px;
-	line-height: 16px;
-	color: #000000;
-	font-weight: normal;
-	padding-top: 10px;
-	padding-bottom: 10px;
-}
-
-/*
-** when this stylesheet is used for the Tiny MCE Wysiwyg editor's edit area, we can't
-** use an id=PageContent or class=wiki-content, so we must
-** set the body style to that used for PageContent, and p to that used for wiki-content.
-*/
-
-.monospaceInput {
-    font:12px monospace
-}
-
-.wiki-content p, .commentblock p {
-    margin: 16px 0px 16px 0px;
-    padding: 0px;
-}
-
-.wiki-content-preview {
-    padding: 5px;
-    border-left: 1px solid #3c78b5;
-    border-right: 1px solid #3c78b5;
-}
-
-ul, ol {
-    margin-top: 2px;
-    margin-bottom: 2px;
-    padding-top: 0px;
-    padding-bottom: 0px;
-}
-
-pre {
-    padding: 0px;
-    margin-top: 5px;
-    margin-left: 15px;
-    margin-bottom: 5px;
-    margin-right: 5px;
-    text-align: left;
-}
-
-.helpheading {
-    font-weight: bold;
-    background-color: #D0D9BD;
-        border-bottom: 1px solid #3c78b5;
-        padding: 4px 4px 4px 4px;
-        margin: 0px;
-        margin-top: 10px;
-}
-.helpcontent {
-        padding: 4px 4px 20px 4px;
-    background-color: #f5f7f1;
-}
-
-.code {
- 	border: 1px dashed #3c78b5;
-    font-size: 11px;
-	font-family: Courier;
-    margin: 10px;
-	line-height: 13px;
-}
-
-.focusedComment {
-    background: #ffffce;
-}
-
-.commentBox, .focusedComment {
-    padding: 10px;
-    margin: 5px 0 5px 0;
-    border: 1px #bbb solid;
-}
-
-.codeHeader {
-    background-color: #f0f0f0;
- 	border-bottom: 1px dashed #3c78b5;
-    padding: 3px;
-	text-align: center;
-}
-
-.codeContent {
-    text-align: left;
-    background-color: #f0f0f0;
-    padding: 3px;
-}
-
-.preformatted {
- 	border: 1px dashed #3c78b5;
-    font-size: 11px;
-	font-family: Courier;
-    margin: 10px;
-	line-height: 13px;
-}
-
-.preformattedHeader {
-    background-color: #f0f0f0;
- 	border-bottom: 1px dashed #3c78b5;
-    padding: 3px;
-	text-align: center;
-}
-
-.preformattedContent {
-    background-color: #f0f0f0;
-    padding: 3px;
-}
-
-.panel {
- 	border: 1px dashed #3c78b5;
-    margin: 10px;
-    margin-top: 0px;
-}
-
-.panelHeader {
-    background-color: #f0f0f0;
- 	border-bottom: 1px dashed #3c78b5;
-    padding: 3px;
-	text-align: center;
-}
-
-.panelContent {
-    background-color: #f0f0f0;
-    padding: 5px;
-}
-
-.anonymousAlert {
-    background-color: #f0f0f0;
- 	border: 1px dashed red;
-    font-size: 11px;
-    padding: 10px 5px 10px 5px;
-    margin: 4px;
-	line-height: 13px;
-}
-
-.lockAlert {
-    background-color: #f0f0f0;
-    width: 50%;
- 	border: 1px dashed red;
-    font-size: 11px;
-    padding: 10px 5px 10px 5px;
-    margin: 4px;
-	line-height: 13px;
-}
-
-
-.code-keyword {
-  color: #000091;
-  background-color: inherit;
-}
-
-.code-object {
-  color: #910091;
-  background-color: inherit;
-}
-
-.code-quote {
-  color: #009100;
-  background-color: inherit;
-}
-
-.code-comment {
-  color: #808080;
-  background-color: inherit;
-}
-
-
-.code-xml .code-keyword {
-  color: inherit;
-  font-weight: bold;
-}
-
-.code-tag {
-  color: #000091;
-  background-color: inherit;
-}
-
-.breadcrumbs {
-    background-color: #f0f0f0;
- 	border-color: #3c78b5;
-	border-width: 1px 0px 1px 0px;
-	border-style: solid;
-    font-size: 11px;
-    padding: 3px 0px 3px 0px;
-}
-
-.navmenu {
-    border: 1px solid #ccc;
-}
-
-.menuheading {
-    font-weight: bold;
-    background-color: #f0f0f0;
- 	border-bottom: 1px solid #3c78b5;
-	padding: 4px 4px 2px 4px;
-}
-
-.menuitems {
-	padding: 4px 4px 20px 4px;
-}
-
-.rightpanel {
-    border-left: 1px solid #ccc;
-    border-bottom: 1px solid #ccc;
-}
-
-#helpheading {
-    text-align: left;
-    font-weight: bold;
-    background-color: #D0D9BD;
- 	border-bottom: 1px solid #3c78b5;
-	padding: 4px 4px 4px 4px;
-	margin: 0px;
-}
-#helpcontent {
-	padding: 4px 4px 4px 4px;
-    background-color: #f5f7f1;
-}
-.helptab-unselected {
-    font-weight: bold;
-	padding: 5px;
-    background-color: #f5f7f1;
-}
-.helptab-selected {
-    font-weight: bold;
-    background-color: #D0D9BD;
-	padding: 5px;
-}
-.helptabs {
-    margin: 0px;
-    background-color: #f5f7f1;
-	padding: 5px;
-}
-.infopanel-heading {
-    font-weight: bold;
-	padding: 4px 0px 2px 0px;
-}
-
-.pagebody {
-}
-
-.pageheader {
-	padding: 5px 5px 5px 0px;
- 	border-bottom: 1px solid #3c78b5;
-}
-
-.steptitle {
-	font-size: 18px;
-	font-weight: bold;
-	font-family: Arial, sans-serif;
-	color: #003366;
-	margin-bottom: 7px;
-}
-
-.substeptitle {
-    font-size: 12px;
-    font-weight: bold;
-    font-family: Arial, sans-serif;
-    color: #003366;
-    margin: 2px 4px 4px 4px;
-    padding: 2px 4px 1px 4px;
-}
-
-.stepdesc {
-    font-family: Verdana, arial, sans-serif;
-	font-size: 11px;
-	line-height: 16px;
-	font-weight: normal;
-    color: #666666;
-    margin-top: 7px;
-    margin-bottom: 7px;
-}
-
-.steplabel {
-    font-weight: bold;
-    margin-right: 4px;
-    color: black;
-    float: left;
-    width: 15%;
-    text-align: right;
-}
-
-.stepfield {
-    background: #f0f0f0;
-    padding: 5px;
-}
-
-.submitButtons{
-    margin-top:5px;
-    text-align:right;
-}
-
-.formtitle {
-	font-size: 12px;
-	font-weight: bold;
-	font-family: Arial, sans-serif;
-	color: #003366;
-}
-
-.sectionbottom {
-    border-bottom: 1px solid #3c78b5;
-}
-
-.topRow {
-    border-top: 2px solid #3c78b5;
-}
-
-.tabletitle {
-	font-size: 14px;
-	font-weight: bold;
-	font-family: Arial, sans-serif;
-    padding: 3px 0px 2px 0px;
-    margin: 8px 4px 2px 0px;
-	color: #003366;
-	border-bottom: 2px solid #3c78b5;
-}
-.pagesubheading {
-    color: #666666;
-    font-size: 10px;
-    padding: 0px 0px 5px 0px;
-}
-
-HR {
-	color: 3c78b5;
-	height: 1;
-}
-
-A:link, A:visited, A:active, A:hover {
-  color: #5b8fbe;
-}
-
-A:hover {
-  text-decoration: underline;
-}
-
-h1 A:link, h1 A:visited, h1 A:active {
-	text-decoration: none;
-}
-
-h1 A:hover {
-    border-bottom: 1px dotted #003366;
-}
-
-.wiki-content > :first-child, .commentblock > :first-child {
-    margin-top: 3px;
-}
-
-.logocell {
-    padding: 10px;
-}
-
-input {
-	font-family: verdana, geneva, arial, sans-serif;
-	font-size: 11px;
-	color: #000000;
-}
-
-textarea, textarea.editor {
-	font-family: verdana, geneva, arial, sans-serif;
-	font-size: 11px;
-	color: #333333;
-}
-
-/* use logoSpaceLink instead.
-.spacenametitle {
-	font: 21px/31px Impact, Arial, Helvetica;
-    font-weight: 100;
-    color: #999999;
-	margin: 0px;
-}
-.spacenametitle img {
-  margin: 0 0 -4px 0;
-}
-.spacenametitle a {
-    text-decoration: none;
-    color: #999999;
-}
-.spacenametitle a:visited {
-    text-decoration: none;
-    color: #999999;
-}*/
-
-.spacenametitle-printable {
-	font: 20px/25px Impact, Arial, Helvetica;
-    font-weight: 100;
-    color: #999999;
-	margin: 0px;
-}
-.spacenametitle-printable a {
-    text-decoration: none;
-    color: #999999;
-}
-.spacenametitle-printable a:visited {
-    text-decoration: none;
-    color: #999999;
-}
-
-.blogDate {
-	font-weight: bold;
-	text-decoration: none;
-	color: black;
-}
-
-.blogSurtitle {
-    background: #f0f0f0;
- 	border: 1px solid #ddd;
-	padding: 3px;
-	margin: 1px 1px 10px 1px;
-}
-
-.blogHeading {
-    font-size: 20px;
-    line-height: normal;
-    font-weight: bold;
-    padding: 0px;
-    margin: 0px;
-}
-
-.blogHeading a {
-   text-decoration: none;
-   color: black;
-}
-
-.endsection {
-	align: right;
-	color: #666666;
-	margin-top: 10px;
-}
-.endsectionleftnav {
-	align: right;
-	color: #666666;
-	margin-top: 10px;
-}
-
-h1 {
-	font-size: 24px;
-	line-height: normal;
-	font-weight: bold;
-	background-color: #f0f0f0;
-	color: #003366;
- 	border-bottom: 1px solid #3c78b5;
-	padding: 2px;
-	margin: 36px 0px 4px 0px;
-}
-
-h2 {
-	font-size: 18px;
-	line-height: normal;
-	font-weight: bold;
-	background-color: #f0f0f0;
- 	border-bottom: 1px solid #3c78b5;
-	padding: 2px;
-	margin: 27px 0px 4px 0px;
-}
-
-h3 {
-	font-size: 14px;
-	line-height: normal;
-	font-weight: bold;
-	background-color: #f0f0f0;
-	padding: 2px;
-	margin: 21px 0px 4px 0px;
-}
-
-h4 {
-	font-size: 12px;
-	line-height: normal;
-	font-weight: bold;
-	background-color: #f0f0f0;
-	padding: 2px;
-	margin: 18px 0px 4px 0px;
-}
-
-h4.search {
-	font-size: 12px;
-	line-height: normal;
-	font-weight: normal;
-	background-color: #f0f0f0;
-	padding: 4px;
-	margin: 18px 0px 4px 0px;
-}
-
-h5 {
-	font-size: 10px;
-	line-height: normal;
-	font-weight: bold;
-	background-color: #f0f0f0;
-	padding: 2px;
-	margin: 14px 0px 4px 0px;
-}
-
-h6 {
-	font-size: 8px;
-	line-height: normal;
-	font-weight: bold;
-	background-color: #f0f0f0;
-	padding: 2px;
-	margin: 14px 0px 4px 0px;
-}
-
-.smallfont {
-    font-size: 10px;
-}
-.descfont {
-    font-size: 10px;
-    color: #666666;
-}
-.smallerfont {
-    font-size: 9px;
-}
-.smalltext {
-    color: #666666;
-    font-size: 10px;
-}
-.smalltext a {
-    color: #666666;
-}
-.smalltext-blue {
-    color: #3c78b5;
-    font-size: 10px;
-}
-.surtitle {
-    margin-left: 1px;
-    margin-bottom: 5px;
-    font-size: 14px;
-    color: #666666;
-}
-
-/* css hack found here:  http://www.fo3nix.pwp.blueyonder.co.uk/tutorials/css/hacks/ */
-.navItemOver { font-size: 10px; font-weight: bold; color: #ffffff; background-color: #003366; cursor: hand; voice-family: '\'}\''; voice-family:inherit; cursor: pointer;}
-.navItemOver a { color: #ffffff; background-color:#003366; text-decoration: none; }
-.navItemOver a:visited { color: #ffffff; background-color:#003366; text-decoration: none; }
-.navItemOver a:hover { color: #ffffff; background-color:#003366; text-decoration: none; }
-.navItem { font-size: 10px; font-weight: bold; color: #ffffff; background-color: #3c78b5; }
-.navItem a { color: #ffffff; text-decoration: none; }
-.navItem a:hover { color: #ffffff; text-decoration: none; }
-.navItem a:visited { color: #ffffff; text-decoration: none; }
-
-div.padded { padding: 4px; }
-div.thickPadded { padding: 10px; }
-h3.macrolibrariestitle {
-    margin: 0px 0px 0px 0px;
-}
-
-div.centered { text-align: center; margin: 10px; }
-div.centered table {margin: 0px auto; text-align: left; }
-
-.tableview table {
-    margin: 0;
-}
-
-.tableview th {
-    text-align: left;
-    color: #003366;
-    font-size: 12px;
-    padding: 5px 0px 0px 5px;
-    border-bottom: 2px solid #3c78b5;
-}
-.tableview td {
-    text-align: left;
-    border-color: #ccc;
-    border-width: 0px 0px 1px 0px;
-    border-style: solid;
-    margin: 0;
-    padding: 4px 10px 4px 5px;
-}
-
-.grid {
-    margin: 2px 0px 5px 0px;
-    border-collapse: collapse;
-}
-.grid th  {
-    border: 1px solid #ccc;
-    padding: 2px 4px 2px 4px;
-    background: #f0f0f0;
-    text-align: center;
-}
-.grid td  {
-    border: 1px solid #ccc;
-    padding: 3px 4px 3px 4px;
-}
-.gridHover {
-	background-color: #f9f9f9;
-}
-
-td.infocell {
-    background-color: #f0f0f0;
-}
-.label {
-	font-weight: bold;
-	color: #003366;
-}
-.error {
-	background-color: #fcc;
-}
-
-.errorBox {
-	background-color: #fcc;
-    border: 1px solid #c00;
-    padding: 5px;
-    margin: 5px;
-}
-
-.errorMessage {
-	color: #c00;
-}
-
-.success {
-	background-color: #dfd;
-}
-
-.successBox {
-	background-color: #dfd;
-    border: 1px solid #090;
-    padding: 5px;
-    margin-top:5px;
-    margin-bottom:5px;
-}
-
-blockquote {
-	padding-left: 10px;
-	padding-right: 10px;
-	margin-left: 5px;
-	margin-right: 0px;
-	border-left: 1px solid #3c78b5;
-}
-
-table.confluenceTable
-{
-    margin: 5px;
-    border-collapse: collapse;
-}
-
-/* Added as a temporary fix for CONF-4223. The table elements appear to be inheriting the border: none attribute from the sectionMacro class */
-table.confluenceTable td.confluenceTd
-{
-    border-width: 1px;
-    border-style: solid;
-    border-color: #ccc;
-    padding: 3px 4px 3px 4px;
-}
-
-/* Added as a temporary fix for CONF-4223. The table elements appear to be inheriting the border: none attribute from the sectionMacro class */
-table.confluenceTable th.confluenceTh
-{
-    border-width: 1px;
-    border-style: solid;
-    border-color: #ccc;
-    padding: 3px 4px 3px 4px;
-    background-color: #f0f0f0;
-    text-align: center;
-}
-
-td.confluenceTd
-{
-    border-width: 1px;
-    border-style: solid;
-    border-color: #ccc;
-    padding: 3px 4px 3px 4px;
-}
-
-th.confluenceTh
-{
-    border-width: 1px;
-    border-style: solid;
-    border-color: #ccc;
-    padding: 3px 4px 3px 4px;
-    background-color: #f0f0f0;
-    text-align: center;
-}
-
-DIV.small {
-	font-size: 9px;
-}
-
-H1.pagename {
-	margin-top: 0px;
-}
-
-IMG.inline  {}
-
-.loginform {
-    margin: 5px;
-    border: 1px solid #ccc;
-}
-
-/* The text how the "This is a preview" comment should be shown. */
-.previewnote { text-align: center;
-                font-size: 11px;
-                    color: red; }
-
-/* How the preview content should be shown */
-.previewcontent { background: #E0E0E0; }
-
-/* How the system messages should be shown (DisplayMessage.jsp) */
-.messagecontent { background: #E0E0E0; }
-
-/* How the "This page has been modified..." -comment should be shown. */
-.conflictnote { }
-
-.createlink {
-    color: maroon;
-}
-a.createlink {
-    color: maroon;
-}
-.templateparameter {
-    font-size: 9px;
-    color: darkblue;
-}
-
-.diffadded {
-    background: #ddffdd;
-    padding: 1px 1px 1px 4px;
-	border-left: 4px solid darkgreen;
-}
-.diffdeleted {
-    color: #999;
-    background: #ffdddd;
-    padding: 1px 1px 1px 4px;
-	border-left: 4px solid darkred;
-}
-.diffnochange {
-    padding: 1px 1px 1px 4px;
-	border-left: 4px solid lightgrey;
-}
-.differror {
-    background: brown;
-}
-.diff {
-    font-family: lucida console, courier new, fixed-width;
-	font-size: 12px;
-	line-height: 14px;
-}
-.diffaddedchars {
-    background-color:#99ff99;
-    font-weight:bolder;
-}
-.diffremovedchars {
-    background-color:#ff9999;
-    text-decoration: line-through;
-    font-weight:bolder;
-}
-
-.greybackground {
-    background: #f0f0f0
-}
-
-.greybox {
- 	border: 1px solid #ddd;
-	padding: 3px;
-	margin: 1px 1px 10px 1px;
-}
-
-.borderedGreyBox {
-    border: 1px solid #cccccc;
-    background-color: #f0f0f0;
-    padding: 10px;
-}
-
-.greyboxfilled {
- 	border: 1px solid #ddd;
-	padding: 5px;
-	margin: 10px 1px 10px 1px;
-	background: #f0f0f0;
-}
-
-.navBackgroundBox {
-    padding: 5px 5px 5px 5px;
-    font-size: 22px;
-	font-weight: bold;
-	font-family: Arial, sans-serif;
-	color: white;
-    background: #3c78b5;
-    text-decoration: none;
-}
-
-.previewBoxTop {
-	background-color: #f0f0f0;
-    border-width: 1px 1px 0px 1px;
-    border-style: solid;
-    border-color: #3c78b5;
-    padding: 5px;
-    margin: 5px 0px 0px 0px;
-    text-align: center;
-}
-.previewContent {
-    background-color: #fff;
- 	border-color: #3c78b5;
-	border-width: 0px 1px 0px 1px;
-	border-style: solid;
-	padding: 10px;
-	margin: 0px;
-}
-.previewBoxBottom {
-	background-color: #f0f0f0;
-    border-width: 0px 1px 1px 1px;
-    border-style: solid;
-    border-color: #3c78b5;
-    padding: 5px;
-    margin: 0px 0px 5px 0px;
-    text-align: center;
-}
-
-.functionbox {
-    background-color: #f0f0f0;
- 	border: 1px solid #3c78b5;
-	padding: 3px;
-	margin: 1px 1px 10px 1px;
-}
-
-.functionbox-greyborder {
-    background-color: #f0f0f0;
- 	border: 1px solid #ddd;
-	padding: 3px;
-	margin: 1px 1px 10px 1px;
-}
-
-.search-highlight {
-    background-color: #ffffcc;
-}
-
-/* normal (white) background */
-.rowNormal {
-    background-color: #ffffff;
- }
-
-/* alternate (pale yellow) background */
-.rowAlternate {
-    background-color: #f7f7f7;
-}
-
-/* used in the list attachments table */
-.rowAlternateNoBottomColor {
-    background-color: #f7f7f7;
-}
-
-.rowAlternateNoBottomNoColor {
-}
-
-.rowAlternateNoBottomColor td {
-    border-bottom: 0px;
-}
-
-.rowAlternateNoBottomNoColor td {
-    border-bottom: 0px;
-}
-
-/* row highlight (grey) background */
-.rowHighlight {
-    background-color: #f0f0f0;
-
-}
-
-TD.greenbar {FONT-SIZE: 2px; BACKGROUND: #00df00; BORDER: 1px solid #9c9c9c; PADDING: 0px; }
-TD.redbar {FONT-SIZE: 2px; BACKGROUND: #df0000; BORDER: 1px solid #9c9c9c; PADDING: 0px; }
-TD.darkredbar {FONT-SIZE: 2px; BACKGROUND: #af0000; BORDER: 1px solid #9c9c9c; PADDING: 0px; }
-
-TR.testpassed {FONT-SIZE: 2px; BACKGROUND: #ddffdd; PADDING: 0px; }
-TR.testfailed {FONT-SIZE: 2px; BACKGROUND: #ffdddd; PADDING: 0px; }
-
-.toolbar  {
-    margin: 0px;
-    border-collapse: collapse;
-}
-
-.toolbar td  {
-    border: 1px solid #ccc;
-    padding: 2px 2px 2px 2px;
-    color: #ccc;
-}
-
-td.noformatting {
-    border-width: 0px;
-    border-style: none;
-    text-align: center;
-	padding: 0px;
-}
-
-.commentblock {
-    margin: 12px 0 12px 0;
-}
-
-/*
- * Divs displaying the license information, if necessary.
- */
-.license-eval, .license-none, .license-nonprofit {
-    border-top: 1px solid #bbbbbb;
-    text-align: center;
-    font-size: 10px;
-    font-family: Verdana, Arial, Helvetica, sans-serif;
-}
-
-.license-eval, .license-none {
-    background-color: #ffcccc;
-}
-
-.license-eval b, .license-none b {
-    color: #990000
-}
-
-.license-nonprofit {
-    background-color: #ffffff;
-}
-
-/*
- * The shadow at the bottom of the page between the main content and the
- * "powered by" section.
- */
-.bottomshadow {
-    height: 12px;
-    background-image: url("$req.contextPath/images/border/border_bottom.gif");
-    background-repeat: repeat-x;
-}
-
-/*
- * Styling of the operations box
- */
-.navmenu .operations li, .navmenu .operations ul {
-    list-style: none;
-    margin-left: 0;
-    padding-left: 0;
-}
-
-.navmenu .operations ul {
-    margin-bottom: 9px;
-}
-
-.navmenu .label {
-    font-weight: inherit;
-}
-
-/*
- * Styling of ops as a toolbar
- */
-.toolbar div {
-    display: none;
-}
-
-.toolbar .label {
-    display: none;
-}
-
-.toolbar .operations {
-    display: block;
-}
-
-.toolbar .operations ul {
-    display: inline;
-    list-style: none;
-    margin-left: 10px;
-    padding-left: 0;
-}
-
-.toolbar .operations li {
-    list-style: none;
-    display: inline;
-}
-
-/* list page navigational tabs */
-#foldertab {
-padding: 3px 0px 3px 8px;
-margin-left: 0;
-border-bottom: 1px solid #3c78b5;
-font: bold 11px Verdana, sans-serif;
-}
-
-#foldertab li {
-list-style: none;
-margin: 0;
-display: inline;
-}
-
-#foldertab li a {
-padding: 3px 0.5em;
-margin-left: 3px;
-border: 1px solid #3c78b5;
-border-bottom: none;
-background: #3c78b5;
-text-decoration: none;
-}
-
-#foldertab li a:link { color: #ffffff; }
-#foldertab li a:visited { color: #ffffff; }
-
-#foldertab li a:hover {
-color: #ffffff;
-background: #003366;
-border-color: #003366;
-}
-
-#foldertab li a.current {
-background: white;
-border-bottom: 1px solid white;
-color: black;
-}
-
-#foldertab li a.current:link { color: black; }
-#foldertab li a.current:visited { color: black; }
-#foldertab li a.current:hover {
-background: white;
-border-bottom: 1px solid white;
-color: black;
-}
-
-/* alphabet list */
-ul#squaretab {
-margin-left: 0;
-padding-left: 0;
-white-space: nowrap;
-font: bold 8px Verdana, sans-serif;
-}
-
-#squaretab li {
-display: inline;
-list-style-type: none;
-}
-
-#squaretab a {
-padding: 2px 6px;
-border: 1px solid #3c78b5;
-}
-
-#squaretab a:link, #squaretab a:visited {
-color: #fff;
-background-color: #3c78b5;
-text-decoration: none;
-}
-
-#squaretab a:hover {
-color: #ffffff;
-background-color: #003366;
-border-color: #003366;
-text-decoration: none;
-}
-
-#squaretab li a#current {
-background: white;
-color: black;
-}
-
-.blogcalendar * {
-    font-family:verdana, arial, sans-serif;
-    font-size:x-small;
-    font-weight:normal;
-    line-height:140%;
-    padding:2px;
-}
-
-
-table.blogcalendar {
-    border: 1px solid #3c78b5;
-}
-
-.blogcalendar th.calendarhead, a.calendarhead {
-    font-size:x-small;
-    font-weight:bold;
-    padding:2px;
-    text-transform:uppercase;
-    background-color: #3c78b5;
-    color: #ffffff;
-    letter-spacing: .3em;
-    text-transform: uppercase;
-}
-
-.calendarhead:visited {color: white;}
-.calendarhead:active {color: white;}
-.calendarhead:hover {color: white;}
-
-.blogcalendar th {
-    font-size:x-small;
-    font-weight:bold;
-    padding:2px;
-    background-color:#f0f0f0;
-}
-
-.blogcalendar td {
-    font-size:x-small;
-    font-weight:normal;
-}
-
-.searchGroup { padding: 0 0 10px 0; background: #f0f0f0; }
-.searchGroupHeading { font-size: 10px; font-weight: bold; color: #ffffff; background-color: #3c78b5; padding: 2px 4px 1px 4px; }
-.searchItem { padding: 1px 4px 1px 4px; }
-.searchItemSelected { padding: 1px 4px 1px 4px; font-weight: bold; background: #ddd; }
-
-/* permissions page styles */
-.permissionHeading {
-    border-bottom: #bbb; border-width: 0 0 1px 0; border-style: solid; font-size: 16px; text-align: left;
-}
-.permissionTab {
-    border-width: 0 0 0 1px; border-style: solid; background: #3c78b5; color: #ffffff; font-size: 10px;
-}
-.permissionSuperTab {
-    border-width: 0 0 0 1px; border-style: solid; background: #003366; color: #ffffff;
-}
-.permissionCell {
-    border-left: #bbb; border-width: 0 0 0 1px; border-style: solid;
-}
-
-/* warning panel */
-.warningPanel { background: #FFFFCE; border:#F0C000 1px solid; padding: 8px; margin: 10px; }
-/* alert panel */
-.alertPanel { background: #FFCCCC; border:#C00 1px solid; padding: 8px; margin: 10px; }
-
-/* side menu highlighting (e.g. space content screen) */
-.optionPadded { padding: 2px; }
-.optionSelected { background-color: #ffffcc; padding: 2px; border: 1px solid #ddd; margin: -1px; }
-.optionSelected a { font-weight: bold; text-decoration: none; color: black; }
-
-/* information macros */
-.noteMacro { border-style: solid; border-width: 1px; border-color: #F0C000; background-color: #FFFFCE; text-align:left; margin-top: 5px; margin-bottom: 5px}
-.warningMacro { border-style: solid; border-width: 1px; border-color: #c00; background-color: #fcc; text-align:left; margin-top: 5px; margin-bottom: 5px}
-.infoMacro { border-style: solid; border-width: 1px; border-color: #3c78b5; background-color: #D8E4F1; text-align:left; margin-top: 5px; margin-bottom: 5px}
-.tipMacro { border-style: solid; border-width: 1px; border-color: #090; background-color: #dfd; text-align:left; margin-top: 5px; margin-bottom: 5px}
-.informationMacroPadding { padding: 5px 0 0 5px; }
-
-table.infoMacro td, table.warningMacro td, table.tipMacro td, table.noteMacro td, table.sectionMacro td {
-    border: none;
-}
-
-table.sectionMacroWithBorder td.columnMacro { border-style: dashed; border-width: 1px; border-color: #cccccc;}
-
-/* styles for links in the top bar */
-.topBarDiv a:link {color: #ffffff;}
-.topBarDiv a:visited {color: #ffffff;}
-.topBarDiv a:active {color: #ffffff;}
-.topBarDiv a:hover {color: #ffffff;}
-.topBarDiv {color: #ffffff;}
-
-.topBar {
-    background-color: #003366;
-}
-
-
-/* styles for extended operations */
-.greyLinks a:link {color: #666666; text-decoration:underline;}
-.greyLinks a:visited {color: #666666; text-decoration:underline;}
-.greyLinks a:active {color: #666666; text-decoration:underline;}
-.greyLinks a:hover {color: #666666; text-decoration:underline;}
-.greyLinks {color: #666666; display:block; padding: 10px}
-
-.logoSpaceLink {color: #999999; text-decoration: none}
-.logoSpaceLink a:link {color: #999999; text-decoration: none}
-.logoSpaceLink a:visited {color: #999999; text-decoration: none}
-.logoSpaceLink a:active {color: #999999; text-decoration: none}
-.logoSpaceLink a:hover {color: #003366; text-decoration: none}
-
-/* basic panel (basicpanel.vmd) style */
-.basicPanelContainer {border: 1px solid #3c78b5; margin-top: 2px; margin-bottom: 8px; width: 100%}
-.basicPanelTitle {padding: 5px; margin: 0px; background-color: #f0f0f0; color: black; font-weight: bold;}
-.basicPanelBody {padding: 5px; margin: 0px}
-
-.separatorLinks a:link {color: white}
-.separatorLinks a:visited {color: white}
-.separatorLinks a:active {color: white}
-
-.greynavbar {background-color: #f0f0f0; border-top: 1px solid #3c78b5; margin-top: 2px}
-
-div.headerField {
-    float: left;
-    width: auto;
-    height: 100%;
-}
-
-.headerFloat {
-    margin-left: auto;
-    width: 50%;
-}
-
-.headerFloatLeft {
-    float: left;
-    margin-right: 20px;
-    margin-bottom: 10px;
-}
-
-#headerRow {
-    padding: 10px;
-}
-
-div.license-personal {
-   background-color: #003366;
-   color: #ffffff;
-}
-
-div.license-personal a {
-   color: #ffffff;
-}
-
-.greyFormBox {
-    border: 1px solid #cccccc;
-    padding: 5px;
-}
-
-/* IE automatically adds a margin before and after form tags. Use this style to remove that */
-.marginlessForm {
-    margin: 0px;
-}
-
-.openPageHighlight {
-    background-color: #ffffcc;
-    padding: 2px;
-    border: 1px solid #ddd;
-}
-
-.editPageInsertLinks, .editPageInsertLinks a
-{
-    color: #666666;
-    font-weight: bold;
-    font-size: 10px;
-}
-
-/* Style for label heatmap. */
-.top10 a {
-    font-weight: bold;
-    font-size: 2em;
-    color: #003366;
-}
-.top25 a {
-    font-weight: bold;
-    font-size: 1.6em;
-    color: #003366;
-}
-.top50 a {
-    font-size: 1.4em;
-    color: #003366;
-}
-.top100 a {
-    font-size: 1.2em;
-    color: #003366;
-}
-
-.heatmap {
-    list-style:none;
-    width: 95%;
-    margin: 0px auto;
-}
-
-.heatmap a {
-    text-decoration:none;
-}
-
-.heatmap a:hover {
-    text-decoration:underline;
-}
-
-.heatmap li {
-    display: inline;
-}
-
-.minitab {
-padding: 3px 0px 3px 8px;
-margin-left: 0;
-margin-top: 1px;
-margin-bottom: 0px;
-border-bottom: 1px solid #3c78b5;
-font: bold 9px Verdana, sans-serif;
-text-decoration: none;
-float:none;
-}
-.selectedminitab {
-padding: 3px 0.5em;
-margin-left: 3px;
-margin-top: 1px;
-border: 1px solid #3c78b5;
-background: white;
-border-bottom: 1px solid white;
-color: #000000;
-text-decoration: none;
-}
-.unselectedminitab {
-padding: 3px 0.5em;
-margin-left: 3px;
-margin-top: 1px;
-border: 1px solid #3c78b5;
-border-bottom: none;
-background: #3c78b5;
-color: #ffffff;
-text-decoration: none;
-}
-
-a.unselectedminitab:hover {
-color: #ffffff;
-background: #003366;
-border-color: #003366;
-}
-
-a.unselectedminitab:link { color: white; }
-a.unselectedminitab:visited { color: white; }
-
-a.selectedminitab:link { color: black; }
-a.selectedminitab:visited { color: black; }
-
-.linkerror { background-color: #fcc;}
-
-a.labelOperationLink:link {text-decoration: underline}
-a.labelOperationLink:active {text-decoration: underline}
-a.labelOperationLink:visited {text-decoration: underline}
-a.labelOperationLink:hover {text-decoration: underline}
-
-a.newLabel:link {background-color: #ddffdd}
-a.newLabel:active {background-color: #ddffdd}
-a.newLabel:visited {background-color: #ddffdd}
-a.newLabel:hover {background-color: #ddffdd}
-
-ul.square {list-style-type: square}
-
-.inline-control-link {
-    background: #ffc;
-    font-size: 9px;
-    color: #666;
-    padding: 2px;
-    text-transform: uppercase;
-    text-decoration: none;
-}
-
-
-.inline-control-link a:link {text-decoration: none}
-.inline-control-link a:active {text-decoration: none}
-.inline-control-link a:visited {text-decoration: none}
-.inline-control-link a:hover {text-decoration: none}
-
-.inline-control-link {
-    background: #ffc;
-    font-size: 9px;
-    color: #666;
-    padding: 2px;
-    text-transform: uppercase;
-    text-decoration: none;
-    cursor: pointer;
-}
-
-div.auto_complete {
-    width: 350px;
-    background: #fff;
-}
-div.auto_complete ul {
-    border: 1px solid #888;
-    margin: 0;
-    padding: 0;
-    width: 100%;
-    list-style-type: none;
-}
-div.auto_complete ul li {
-    margin: 0;
-    padding: 3px;
-}
-div.auto_complete ul li.selected {
-    background-color: #ffb;
-}
-div.auto_complete ul strong.highlight {
-    color: #800;
-    margin: 0;
-    padding: 0;
-}
-
-/******* Edit Page Styles *******/
-.toogleFormDiv{
-    border:1px solid #A7A6AA;
-    background-color:white;
-    padding:5px;
-    margin-top: 5px;
-}
-
-.toogleInfoDiv{
-    border:1px solid #A7A6AA;
-    background-color:white;
-    display:none;
-    padding:5px;
-    margin-top: 10px;
-}
-
-.inputSection{
-    margin-bottom:20px;
-}
-
-#editBox{
-   border:1px solid lightgray;
-   background-color:#F0F0F0;
-}
-
-/******* Left Navigation Theme Styles ********/
-.leftnav li a {
-    text-decoration:none;
-    color:white;
-    margin:0px;
-    display:block;
-    padding:2px;
-    padding-left:5px;
-    background-color: #3c78b5;
-    border-top:1px solid #3c78b5;
-}
-
-.leftnav li a:active {color:white;}
-.leftnav li a:visited {color:white;}
-.leftnav li a:hover {background-color: #003366; color:white;}

http://git-wip-us.apache.org/repos/asf/incubator-rya/blob/ce4a10ff/extras/rya.manual/src/main/webapp/eval.conf
----------------------------------------------------------------------
diff --git a/extras/rya.manual/src/main/webapp/eval.conf b/extras/rya.manual/src/main/webapp/eval.conf
deleted file mode 100644
index f502b94..0000000
--- a/extras/rya.manual/src/main/webapp/eval.conf
+++ /dev/null
@@ -1,56 +0,0 @@
-h1. Prospects Table
-
-The Prospects Table provides statistics on the number of subject/predicate/object data found in the triple store. It is currently a
-Map Reduce job that will run against the Rya store and save all the statistics in the prosepcts table.
-
-h2. Build
-
-[Build the mmrts.git repo|./build-source.html]
-
-h2. Run
-
-Deploy the extras/rya.prospector/target/rya.prospector-<version>-shade.jar file to the hadoop cluster.
-
-The prospector also requires a configuration file that defines where Accumulo is, which Rya table (has to be the SPO table) to read from, and
-which table to output to. (Note: Make sure you follow the same schema as the Rya tables (prospects table name: tableprefix_prospects)
-
-A sample configuration file might look like the following:
-{code}
-<?xml version="1.0"?>
-<?xml-stylesheet type="text/xsl" href="configuration.xsl"?>
-<configuration>
-    <property>
-        <name>prospector.intable</name>
-        <value>triplestore_spo</value>
-    </property>
-    <property>
-        <name>prospector.outtable</name>
-        <value>triplestore_prospects</value>
-    </property>
-    <property>
-        <name>prospector.auths</name>
-        <value>U,FOUO</value>
-    </property>
-    <property>
-        <name>instance</name>
-        <value>accumulo</value>
-    </property>
-    <property>
-        <name>zookeepers</name>
-        <value>localhost:2181</value>
-    </property>
-    <property>
-        <name>username</name>
-        <value>root</value>
-    </property>
-    <property>
-        <name>password</name>
-        <value>secret</value>
-    </property>
-</configuration>
-{code}
-
-Run the command, filling in the correct information.
-{code}
-hadoop jar rya.prospector-3.0.4-SNAPSHOT-shade.jar mvm.rya.prospector.mr.Prospector /tmp/prospectorConf.xml
-{code}
\ No newline at end of file

http://git-wip-us.apache.org/repos/asf/incubator-rya/blob/ce4a10ff/extras/rya.manual/src/main/webapp/images/bg.png
----------------------------------------------------------------------
diff --git a/extras/rya.manual/src/main/webapp/images/bg.png b/extras/rya.manual/src/main/webapp/images/bg.png
deleted file mode 100644
index 878a84f..0000000
Binary files a/extras/rya.manual/src/main/webapp/images/bg.png and /dev/null differ

http://git-wip-us.apache.org/repos/asf/incubator-rya/blob/ce4a10ff/extras/rya.manual/src/main/webapp/images/bg02-blue-left.png
----------------------------------------------------------------------
diff --git a/extras/rya.manual/src/main/webapp/images/bg02-blue-left.png b/extras/rya.manual/src/main/webapp/images/bg02-blue-left.png
deleted file mode 100644
index 057efdf..0000000
Binary files a/extras/rya.manual/src/main/webapp/images/bg02-blue-left.png and /dev/null differ

http://git-wip-us.apache.org/repos/asf/incubator-rya/blob/ce4a10ff/extras/rya.manual/src/main/webapp/images/bg02-blue-right.png
----------------------------------------------------------------------
diff --git a/extras/rya.manual/src/main/webapp/images/bg02-blue-right.png b/extras/rya.manual/src/main/webapp/images/bg02-blue-right.png
deleted file mode 100644
index edf9a9d..0000000
Binary files a/extras/rya.manual/src/main/webapp/images/bg02-blue-right.png and /dev/null differ

http://git-wip-us.apache.org/repos/asf/incubator-rya/blob/ce4a10ff/extras/rya.manual/src/main/webapp/images/bg02-white-left-nogr.png
----------------------------------------------------------------------
diff --git a/extras/rya.manual/src/main/webapp/images/bg02-white-left-nogr.png b/extras/rya.manual/src/main/webapp/images/bg02-white-left-nogr.png
deleted file mode 100644
index 1c5186c..0000000
Binary files a/extras/rya.manual/src/main/webapp/images/bg02-white-left-nogr.png and /dev/null differ

http://git-wip-us.apache.org/repos/asf/incubator-rya/blob/ce4a10ff/extras/rya.manual/src/main/webapp/images/bg02-white-right-nogr.png
----------------------------------------------------------------------
diff --git a/extras/rya.manual/src/main/webapp/images/bg02-white-right-nogr.png b/extras/rya.manual/src/main/webapp/images/bg02-white-right-nogr.png
deleted file mode 100644
index 9734def..0000000
Binary files a/extras/rya.manual/src/main/webapp/images/bg02-white-right-nogr.png and /dev/null differ

http://git-wip-us.apache.org/repos/asf/incubator-rya/blob/ce4a10ff/extras/rya.manual/src/main/webapp/images/big-bullet.png
----------------------------------------------------------------------
diff --git a/extras/rya.manual/src/main/webapp/images/big-bullet.png b/extras/rya.manual/src/main/webapp/images/big-bullet.png
deleted file mode 100644
index f036db5..0000000
Binary files a/extras/rya.manual/src/main/webapp/images/big-bullet.png and /dev/null differ

http://git-wip-us.apache.org/repos/asf/incubator-rya/blob/ce4a10ff/extras/rya.manual/src/main/webapp/images/left-box-bottom.png
----------------------------------------------------------------------
diff --git a/extras/rya.manual/src/main/webapp/images/left-box-bottom.png b/extras/rya.manual/src/main/webapp/images/left-box-bottom.png
deleted file mode 100644
index 0495248..0000000
Binary files a/extras/rya.manual/src/main/webapp/images/left-box-bottom.png and /dev/null differ

http://git-wip-us.apache.org/repos/asf/incubator-rya/blob/ce4a10ff/extras/rya.manual/src/main/webapp/images/left-box-right.png
----------------------------------------------------------------------
diff --git a/extras/rya.manual/src/main/webapp/images/left-box-right.png b/extras/rya.manual/src/main/webapp/images/left-box-right.png
deleted file mode 100644
index 2698074..0000000
Binary files a/extras/rya.manual/src/main/webapp/images/left-box-right.png and /dev/null differ

http://git-wip-us.apache.org/repos/asf/incubator-rya/blob/ce4a10ff/extras/rya.manual/src/main/webapp/images/left-box-top.png
----------------------------------------------------------------------
diff --git a/extras/rya.manual/src/main/webapp/images/left-box-top.png b/extras/rya.manual/src/main/webapp/images/left-box-top.png
deleted file mode 100644
index e1fc26e..0000000
Binary files a/extras/rya.manual/src/main/webapp/images/left-box-top.png and /dev/null differ

http://git-wip-us.apache.org/repos/asf/incubator-rya/blob/ce4a10ff/extras/rya.manual/src/main/webapp/images/logo.png
----------------------------------------------------------------------
diff --git a/extras/rya.manual/src/main/webapp/images/logo.png b/extras/rya.manual/src/main/webapp/images/logo.png
deleted file mode 100644
index 42a1fb5..0000000
Binary files a/extras/rya.manual/src/main/webapp/images/logo.png and /dev/null differ

http://git-wip-us.apache.org/repos/asf/incubator-rya/blob/ce4a10ff/extras/rya.manual/src/main/webapp/images/logo2.png
----------------------------------------------------------------------
diff --git a/extras/rya.manual/src/main/webapp/images/logo2.png b/extras/rya.manual/src/main/webapp/images/logo2.png
deleted file mode 100644
index 9d29d0f..0000000
Binary files a/extras/rya.manual/src/main/webapp/images/logo2.png and /dev/null differ

http://git-wip-us.apache.org/repos/asf/incubator-rya/blob/ce4a10ff/extras/rya.manual/src/main/webapp/index.conf
----------------------------------------------------------------------
diff --git a/extras/rya.manual/src/main/webapp/index.conf b/extras/rya.manual/src/main/webapp/index.conf
deleted file mode 100644
index 2bcf160..0000000
--- a/extras/rya.manual/src/main/webapp/index.conf
+++ /dev/null
@@ -1,4 +0,0 @@
-h1. Rya
-
-This project contains documentation about the Rya, a scalable RDF triple store on top of Accumulo.
-

http://git-wip-us.apache.org/repos/asf/incubator-rya/blob/ce4a10ff/extras/rya.manual/src/main/webapp/infer.conf
----------------------------------------------------------------------
diff --git a/extras/rya.manual/src/main/webapp/infer.conf b/extras/rya.manual/src/main/webapp/infer.conf
deleted file mode 100644
index e9c16d2..0000000
--- a/extras/rya.manual/src/main/webapp/infer.conf
+++ /dev/null
@@ -1,13 +0,0 @@
-h1. Inferencing
-
-The current inferencing set supported includes:
-* rdfs:subClassOf
-* rdfs:subPropertyOf
-* owl:equivalentProperty
-* owl:inverseOf
-* owl:SymmetricProperty
-* owl:TransitiveProperty (* This is implemented, but probably not fully. Still in testing)
-
-Nothing special has to be done outside of making sure that the RdfCloudTripleStore object has the InferencingEngine object set on it and properly configured. This is usually done by default. See the [Query Data Section|./querydata.html] for a simple example.
-
-Also, the inferencing engine is set to pull down the latest model every 5 minutes currently (which is configurable). So if you load a new model, a previous RepositoryConnection may not pick up these changes into the Inferencing Engine yet. Getting the InferencingEngine object from the RdfCloudTripleStore and running the {{refreshGraph}} method can refresh the inferred graph immediately.
\ No newline at end of file

http://git-wip-us.apache.org/repos/asf/incubator-rya/blob/ce4a10ff/extras/rya.manual/src/main/webapp/loadPrecomputedJoin.conf
----------------------------------------------------------------------
diff --git a/extras/rya.manual/src/main/webapp/loadPrecomputedJoin.conf b/extras/rya.manual/src/main/webapp/loadPrecomputedJoin.conf
deleted file mode 100644
index 868b5a1..0000000
--- a/extras/rya.manual/src/main/webapp/loadPrecomputedJoin.conf
+++ /dev/null
@@ -1,27 +0,0 @@
-h1. Load Pre-computed Join
-
-A tool has been created to load a pre-computed join.  This tool will generate an index to support a pre-computed join on a user provided SPARQL query, and then register that query within Rya.
-
-
-h2. Registering a pre-computed join
-
-Generating a pre-computed join is done using Pig to execute a series of Map Reduce jobs.  The index (pre-computed join) is associated with a user defined SPARQL query.  
-  
-To execute the indexing tool, compile and run {code}mvm.rya.accumulo.pig.IndexWritingTool{code} 
-with the following seven input arguments: {code} [hdfsSaveLocation] [sparqlFile] [cbinstance] [cbzk] [cbuser] [cbpassword] [rdfTablePrefix]{code}
-
-
-Options:
-* hdfsSaveLocation: a working directory on hdfs for storing interim results
-* sparqlFile: the query to generate a precomputed join for
-* cbinstance: the accumulo instance name
-* cbzk: the accumulo zookeeper name
-* cbuser: the accumulo username
-* cbpassword:  the accumulo password for the supplied user
-* rdfTablePrefix : The tables (spo, po, osp) are prefixed with this qualifier. The tables become: (rdf.tablePrefix)spo,(rdf.tablePrefix)po,(rdf.tablePrefix)osp
-
-
-h1. Using a Pre-computed Join
-
-An example of using a pre-computed join can be referenced in 
-{code} mvm.rya.indexing.external.ExternalSailExample {code}

http://git-wip-us.apache.org/repos/asf/incubator-rya/blob/ce4a10ff/extras/rya.manual/src/main/webapp/loaddata.conf
----------------------------------------------------------------------
diff --git a/extras/rya.manual/src/main/webapp/loaddata.conf b/extras/rya.manual/src/main/webapp/loaddata.conf
deleted file mode 100644
index 9eb7c47..0000000
--- a/extras/rya.manual/src/main/webapp/loaddata.conf
+++ /dev/null
@@ -1,117 +0,0 @@
-h1. Load Data
-
-There are a few mechanisms to load data
-
-h2. Web REST endpoint
-
-The War sets up a Web REST endpoint at {{http://server/web.rya/loadrdf}} that allows POST data to get loaded into the Rdf Store. This short tutorial will use Java code to post data.
-
-First, you will need data to load and will need to figure out what format that data is in.
-
-For this sample, we will use the following N-Triples:
-{code}
-<http://mynamespace/ProductType1> <http://www.w3.org/1999/02/22-rdf-syntax-ns#type> <http://mynamespace/ProductType> .
-<http://mynamespace/ProductType1> <http://www.w3.org/2000/01/rdf-schema#label> "Thing" .
-<http://mynamespace/ProductType1> <http://purl.org/dc/elements/1.1/publisher> <http://mynamespace/Publisher1> .
-{code}
-
-Save this file somewhere $RDF_DATA
-
-Second, use the following Java code to load data to the REST endpoint:
-{code}
-import java.io.BufferedReader;
-import java.io.InputStream;
-import java.io.InputStreamReader;
-import java.io.OutputStream;
-import java.net.URL;
-import java.net.URLConnection;
-
-public class LoadDataServletRun {
-
-    public static void main(String[] args) {
-        try {
-            final InputStream resourceAsStream = Thread.currentThread().getContextClassLoader()
-                    .getResourceAsStream("$RDF_DATA");
-            URL url = new URL("http://server/web.rya/loadrdf" +
-                    "?format=N-Triples" +
-                    "");
-            URLConnection urlConnection = url.openConnection();
-            urlConnection.setRequestProperty("Content-Type", "text/plain");
-            urlConnection.setDoOutput(true);
-
-            final OutputStream os = urlConnection.getOutputStream();
-
-            int read;
-            while((read = resourceAsStream.read()) >= 0) {
-                os.write(read);
-            }
-            resourceAsStream.close();
-            os.flush();
-
-            BufferedReader rd = new BufferedReader(new InputStreamReader(
-                    urlConnection.getInputStream()));
-            String line;
-            while ((line = rd.readLine()) != null) {
-                System.out.println(line);
-            }
-            rd.close();
-            os.close();
-        } catch (Exception e) {
-            e.printStackTrace();
-        }
-    }
-}
-{code}
-
-Compile and run this code above, changing the references for $RDF_DATA and the url that your Rdf War is running at.
-
-The default "format" is RDF/XML, but these formats are supported : RDFXML, NTRIPLES, TURTLE, N3, TRIX, TRIG.
-
-h2. Bulk Loading data
-
-Bulk loading data is done through Map Reduce jobs
-
-h3. Bulk Load RDF data
-
-  This Map Reduce job will read a full file into memory and parse it into statements. The statements are saved into the store. Here is an example for storing in Accumulo:
-{code}
-hadoop jar target/accumulo.rya-3.0.4-SNAPSHOT-shaded.jar mvm.rya.accumulo.mr.fileinput.BulkNtripsInputTool -Dac.zk=localhost:2181 -Dac.instance=accumulo -Dac.username=root -Dac.pwd=secret -Drdf.tablePrefix=triplestore_ -Dio.sort.mb=64 /tmp/temp.ntrips
-{code}
-
-Options:
-* rdf.tablePrefix : The tables (spo, po, osp) are prefixed with this qualifier. The tables become: (rdf.tablePrefix)spo,(rdf.tablePrefix)po,(rdf.tablePrefix)osp
-* ac.* : Accumulo connection parameters
-* rdf.format : See RDFFormat from openrdf, samples include (Trig, N-Triples, RDF/XML)
-* io.sort.mb : Higher the value, the faster the job goes. Just remember that you will need this much ram at least per mapper
-
-The argument is the directory/file to load. This file needs to be loaded into HDFS before running.
-
-h2. Direct OpenRDF API
-
-Here is some sample code to load data directly through the OpenRDF API. (Loading N-Triples data)
-You will need at least cloudbase.rya-<version>, rya.api, rya.sail.impl on the classpath and transitive dependencies. I find that Maven is the easiest way to get a project dependency tree set up.
-
-{code}
-            final RdfCloudTripleStore store = new RdfCloudTripleStore();
-            AccumuloRdfConfiguration conf = new AccumuloRdfConfiguration();
-            AccumuloRyaDAO dao = new AccumuloRdfDAO();
-            Connector connector = new ZooKeeperInstance("instance", "zkserver:port").getConnector("user", "password");
-            dao.setConnector(connector);
-            conf.setTablePrefix("rya_");
-            dao.setConf(conf);
-            store.setRdfDao(dao);
-
-            Repository myRepository = new RyaSailRepository(store);
-            myRepository.initialize();
-            RepositoryConnection conn = myRepository.getConnection();
-
-            //load data from file
-            final File file = new File("ntriples.ntrips");
-            conn.add(new FileInputStream(file), file.getName(),
-                    RDFFormat.NTRIPLES, new Resource[]{});
-
-            conn.commit();
-
-            conn.close();
-            myRepository.shutDown();
-{code}
\ No newline at end of file

http://git-wip-us.apache.org/repos/asf/incubator-rya/blob/ce4a10ff/extras/rya.manual/src/main/webapp/overview.conf
----------------------------------------------------------------------
diff --git a/extras/rya.manual/src/main/webapp/overview.conf b/extras/rya.manual/src/main/webapp/overview.conf
deleted file mode 100644
index f351612..0000000
--- a/extras/rya.manual/src/main/webapp/overview.conf
+++ /dev/null
@@ -1,4 +0,0 @@
-h1. Overview
-
-The MVM RDF Cloud Triple Store is a scalable RDF Store that is built on top of a Columnar Index Store (such as Accumulo, Cloudbase, Hbase). It is implemented as an extension to OpenRdf to provide easy query mechanisms (SPARQL, SERQL, etc) and Rdf data storage (RDF/XML, NTriples, etc).
-

http://git-wip-us.apache.org/repos/asf/incubator-rya/blob/ce4a10ff/extras/rya.manual/src/main/webapp/querydata.conf
----------------------------------------------------------------------
diff --git a/extras/rya.manual/src/main/webapp/querydata.conf b/extras/rya.manual/src/main/webapp/querydata.conf
deleted file mode 100644
index 1f93944..0000000
--- a/extras/rya.manual/src/main/webapp/querydata.conf
+++ /dev/null
@@ -1,115 +0,0 @@
-h1. Query Data
-
-There are a few mechanisms to query data
-
-h2. Web JSP endpoint
-
-Open a url to {{http://server/web.rya/sparqlQuery.jsp}}. This simple form can run Sparql.
-
-h2. Web REST endpoint
-
-The War sets up a Web REST endpoint at {{http://server/web.rya/queryrdf}} that allows GET requests with queries.
-
-For this sample, we will assume you already loaded data from the [loaddata.html] tutorial
-
-Save this file somewhere $RDF_DATA
-
-Second, use the following Java code to load data to the REST endpoint:
-{code}
-import java.io.BufferedReader;
-import java.io.InputStreamReader;
-import java.net.URL;
-import java.net.URLConnection;
-import java.net.URLEncoder;
-
-public class QueryDataServletRun {
-
-    public static void main(String[] args) {
-        try {
-            String query = "select * where {\n" +
-                                "<http://mynamespace/ProductType1> ?p ?o.\n" +
-                                "}";
-
-            String queryenc = URLEncoder.encode(query, "UTF-8");
-
-            URL url = new URL("http://server/rdfTripleStore/queryrdf?query=" + queryenc);
-            URLConnection urlConnection = url.openConnection();
-            urlConnection.setDoOutput(true);
-
-            BufferedReader rd = new BufferedReader(new InputStreamReader(
-                    urlConnection.getInputStream()));
-            String line;
-            while ((line = rd.readLine()) != null) {
-                System.out.println(line);
-            }
-            rd.close();
-        } catch (Exception e) {
-            e.printStackTrace();
-        }
-    }
-}
-{code}
-
-Compile and run this code above, changing the url that your Rdf War is running at.
-
-h2. Direct Code
-
-Here is a code snippet for directly running against Accumulo with the code. You will need at least accumulo.rya.jar, rya.api, rya.sail.impl on the classpath and transitive dependencies. I find that Maven is the easiest way to get a project dependency tree set up.
-
-{code}
-            Connector connector = new ZooKeeperInstance("cbinstance", "zkserver:port").getConnector("cbuser", "cbpassword");
-
-            final RdfCloudTripleStore store = new RdfCloudTripleStore();
-            AccumuloRyaDAO crdfdao = new AccumuloRyaDAO();
-            crdfdao.setConnector(connector);
-
-            AccumuloRdfConfiguration conf = new AccumuloRdfConfiguration();
-            conf.setTablePrefix("rts_");
-            conf.setDisplayQueryPlan(true);
-            crdfdao.setConf(conf);
-            store.setRdfDao(crdfdao);
-
-            ProspectorServiceEvalStatsDAO evalDao = new ProspectorServiceEvalStatsDAO(connector, conf);
-            evalDao.init();
-            store.setRdfEvalStatsDAO(evalDao);
-
-            InferenceEngine inferenceEngine = new InferenceEngine();
-            inferenceEngine.setRdfDao(crdfdao);
-            inferenceEngine.setConf(conf);
-            store.setInferenceEngine(inferenceEngine);
-
-            Repository myRepository = new RyaSailRepository(store);
-            myRepository.initialize();
-
-            String query = "select * where {\n" +
-                                "<http://mynamespace/ProductType1> ?p ?o.\n" +
-                                "}";
-            RepositoryConnection conn = myRepository.getConnection();
-            System.out.println(query);
-            TupleQuery tupleQuery = conn.prepareTupleQuery(
-                    QueryLanguage.SPARQL, query);
-            ValueFactory vf = ValueFactoryImpl.getInstance();
-
-            TupleQueryResultHandler writer = new SPARQLResultsXMLWriter(System.out);
-            tupleQuery.evaluate(new TupleQueryResultHandler() {
-
-                int count = 0;
-
-                @Override
-                public void startQueryResult(List<String> strings) throws TupleQueryResultHandlerException {
-                }
-
-                @Override
-                public void endQueryResult() throws TupleQueryResultHandlerException {
-                }
-
-                @Override
-                public void handleSolution(BindingSet bindingSet) throws TupleQueryResultHandlerException {
-                    System.out.println(bindingSet);
-                }
-            });
-
-            conn.close();
-            myRepository.shutDown();
-{code}
-

http://git-wip-us.apache.org/repos/asf/incubator-rya/blob/ce4a10ff/extras/rya.manual/src/main/webapp/quickstart.conf
----------------------------------------------------------------------
diff --git a/extras/rya.manual/src/main/webapp/quickstart.conf b/extras/rya.manual/src/main/webapp/quickstart.conf
deleted file mode 100644
index ce0d20b..0000000
--- a/extras/rya.manual/src/main/webapp/quickstart.conf
+++ /dev/null
@@ -1,42 +0,0 @@
-h1. Quick Start
-
-This tutorial will outline the steps needed to get quickly started with the Rya store using the web based endpoint.
-
-h2. Prerequisites
-
-* Columnar Store (either Accumulo or Cloudbase) The tutorial will go forward using Accumulo
-* Rya code (Git: git@github.com:texeltek/rya.git)
-* Maven 2.2 +
-
-h2. Building from Source
-
-Using Git, pull down the latest code from the url above.
-
-Run the command to build the code {{mvn clean install}}
-
-If all goes well, the build should be successful and a war should be produced in {{web/web.rya/target/web.rya.war}}
-
-h2. Deployment
-
-(Using tomcat)
-
-Unwar the above war into the webapps directory.
-
-To point the web.rya war to the appropriate Accumulo instance, make a properties file {{environment.properties}} and put it in the classpath. Here is an example:
-{code}
-instance.name=accumulo  #Accumulo instance name
-instance.zk=localhost:2181  #Accumulo Zookeepers
-instance.username=root  #Accumulo username
-instance.password=secret  #Accumulo pwd
-rya.tableprefix=triplestore_  #Rya Table Prefix
-rya.displayqueryplan=true  #To display the query plan
-{code}
-
-Start the Tomcat server. {{./bin/startup.sh}}
-
-h2. Usage
-
-First, we need to load data. See the [Load Data Section|./loaddata]
-
-Second, we need to query that data. See the [Query Data Section|./querydata]
-

http://git-wip-us.apache.org/repos/asf/incubator-rya/blob/ce4a10ff/extras/rya.manual/src/main/webapp/sm-addauth.conf
----------------------------------------------------------------------
diff --git a/extras/rya.manual/src/main/webapp/sm-addauth.conf b/extras/rya.manual/src/main/webapp/sm-addauth.conf
deleted file mode 100644
index 371d8da..0000000
--- a/extras/rya.manual/src/main/webapp/sm-addauth.conf
+++ /dev/null
@@ -1,98 +0,0 @@
-h1. Add Authentication
-
-This tutorial will give a few examples on how to load and query data with authentication.
-
-This is only available for Cloudbase and Accumulo because they provide the security filters necessary to do row level authentication and visibility.
-
-h2. Load Data with Visibilities
-
-During the Load process, there are a few ways to set the Column Visibility you want set on each of the corresponding rdf rows.
-
-h3. Global Visibility
-
-You can set the Column Visibility globally on the RdfCloudTripleStore, and it will use that particular value for every row saved.
-
-To do this, once you create and set up the RdfCloudTripleStore, just set the property on the store configuration:
-
-{code}
-//setup
-final RdfCloudTripleStore store = new RdfCloudTripleStore();
-AccumuloRyaDAO crdfdao = new AccumuloRyaDAO();
-crdfdao.setConnector(connector);
-
-AccumuloRdfConfiguration conf = new AccumuloRdfConfiguration();
-conf.setTablePrefix("rts_");
-conf.setDisplayQueryPlan(true);
-
-//set global column Visibility
-conf.setCv("AUTH1|AUTH2");
-
-crdfdao.setConf(conf);
-store.setRdfDao(crdfdao);
-{code}
-
-The format is simply the same as the Column Visibility format.
-
-h3. Per triple or document based Visibility
-
-TODO: Not available as of yet
-
-h2. Query Data with Authentication
-
-Attaching an Authentication to the query process is very simple. It requires just adding the property {{RdfCloudTripleStoreConfiguration.CONF_QUERY_AUTH}} to the query {{BindingSet}}
-Example:
-
-{code}
-//setup
-Connector connector = new ZooKeeperInstance("cbinstance", "zkserver:port").getConnector("cbuser", "cbpassword");
-final RdfCloudTripleStore store = new RdfCloudTripleStore();
-AccumuloRyaDAO crdfdao = new AccumuloRyaDAO();
-crdfdao.setConnector(connector);
-
-AccumuloRdfConfiguration conf = new AccumuloRdfConfiguration();
-conf.setTablePrefix("rts_");
-conf.setDisplayQueryPlan(true);
-crdfdao.setConf(conf);
-//set global column Visibility
-conf.setCv("1|2");
-store.setRdfDao(crdfdao);
-
-InferenceEngine inferenceEngine = new InferenceEngine();
-inferenceEngine.setRdfDao(crdfdao);
-inferenceEngine.setConf(conf);
-store.setInferenceEngine(inferenceEngine);
-
-Repository myRepository = new RyaSailRepository(store);
-myRepository.initialize();
-RepositoryConnection conn = myRepository.getConnection();
-
-//define and add statement
-String litdupsNS = "urn:test:litdups#";
-URI cpu = vf.createURI(litdupsNS, "cpu");
-URI loadPerc = vf.createURI(litdupsNS, "loadPerc");
-URI uri1 = vf.createURI(litdupsNS, "uri1");
-conn.add(cpu, loadPerc, uri1);
-conn.commit();
-
-//query with auth
-String query = "select * where {" +
-                "<" + cpu.toString() + "> ?p ?o1." +
-                "}";
-TupleQuery tupleQuery = conn.prepareTupleQuery(QueryLanguage.SPARQL, query);
-tupleQuery.setBinding(RdfCloudTripleStoreConfiguration.CONF_QUERY_AUTH, vf.createLiteral("2"));
-TupleQueryResult result = tupleQuery.evaluate();
-while(result.hasNext()) {
-    System.out.println(result.next());
-}
-result.close();
-
-//close
-conn.close();
-myRepository.shutDown();
-{code}
-
-Or you can set a global auth using the configuration:
-
-{code}
-conf.setAuth("2")
-{code}
\ No newline at end of file

http://git-wip-us.apache.org/repos/asf/incubator-rya/blob/ce4a10ff/extras/rya.manual/src/main/webapp/sm-firststeps.conf
----------------------------------------------------------------------
diff --git a/extras/rya.manual/src/main/webapp/sm-firststeps.conf b/extras/rya.manual/src/main/webapp/sm-firststeps.conf
deleted file mode 100644
index 04a56b9..0000000
--- a/extras/rya.manual/src/main/webapp/sm-firststeps.conf
+++ /dev/null
@@ -1,55 +0,0 @@
-h1. Typical First Steps
-
-In this tutorial, I will give you a quick overview of some of the first steps I perform to get data loaded and read for query.
-
-h2. Prerequisites
-
- We are assuming Accumulo 1.4+ usage here.
-
- * Rya Source Code {{web.rya.war}})
- * Accumulo on top of Hadoop 0.20+
- * RDF Data (in N-Triples format, this format is the easiest to bulk load)
-
-h2. Building Source
-
-Skip this section if you already have the Map Reduce artifact and the WAR
-
-See the [Build From Source Section|./build-source.html] to get the appropriate artifacts built
-
-h2. Load Data
-
-I find that the best way to load the data is through the Bulk Load Map Reduce job.
-
-# Save the RDF Data above onto HDFS. From now on we will refer to this location as <RDF_HDFS_LOCATION>
-# Move the cloudbase.rya-<version>-job.jar onto the hadoop cluster
-# Bulk load the data. Here is a sample command line:
-{code}
-hadoop jar ../cloudbase.rya-2.0.0-SNAPSHOT-job.jar BulkNtripsInputTool -Drdf.tablePrefix=lubm_ -Dcb.username=cbuser -Dcb.pwd=cbpwd -Dcb.instance=cbinstance -Dcb.zk=zookeeperLocation -Drdf.format=N-Triples <RDF_HDFS_LOCATION>
-{code}
-
-Once the data is loaded, it is actually a good practice to compact your tables. You can do this by opening the cloudbase shell {{cbshell}} and running the {{compact}} command on the generated tables. Remember the generated tables will be prefixed by the {{rdf.tablePrefix}} property you assigned above. The default tablePrefix is {{rts}}.
-Here is a sample cloudbase shell command:
-{code}
-compact -p lubm_(.*)
-{code}
-
-See the [Load Data Section|./loaddata.html] for more options on loading rdf data
-
-h2. Run the Statistics Optimizer
-
-For the best query performance, it is recommended to run the Statistics Optimizer to create the Evaluation Statistics table. This job will read through your data and gather statistics on the distribution of the dataset. This table is then queried before query execution to reorder queries based on the data distribution.
-
-See the [Evaluation Statistics Table Section|eval.html] on how to do this.
-
-h2. Query data
-
-I find the easiest way to query is just to use the WAR. Load the WAR into your favorite web application container and go to the sparqlQuery.jsp page. Example:
-{code}
-http://localhost:8080/web.rya/sparqlQuery.jsp
-{code}
-
-This page provides a very simple text box for running queries against the store and getting data back. (SPARQL queries)
-
-Remember to update the connection information in the WAR: {{WEB-INF/spring/spring-cloudbase.xml}}
-
-See the [Query data section|./querydata.html] for more information.
\ No newline at end of file

http://git-wip-us.apache.org/repos/asf/incubator-rya/blob/ce4a10ff/extras/rya.manual/src/main/webapp/sm-infer.conf
----------------------------------------------------------------------
diff --git a/extras/rya.manual/src/main/webapp/sm-infer.conf b/extras/rya.manual/src/main/webapp/sm-infer.conf
deleted file mode 100644
index ab25281..0000000
--- a/extras/rya.manual/src/main/webapp/sm-infer.conf
+++ /dev/null
@@ -1,313 +0,0 @@
-h1. Inferencing
-
-The Rdf store provides simple inferencing currently. The supported list of inferred relationships include:
-
-* rdfs:subClassOf
-* rdfs:subPropertyOf
-* owl:EquivalentProperty
-* owl:inverseOf
-* owl:SymmetricProperty
-* owl:TransitiveProperty (This is currently in beta and will not work for every case)
-
-h2. Setup
-
-The Inferencing Engine is a scheduled job that runs by default every 5 minutes, this is configurable, to query the relationships in the store and develop the inferred graphs necessary to answer inferencing questions.
-
-This also means that if you load a model into the store, it could take up to 5 minutes for the inferred relationships to be available.
-
-As usual you will need to set up your {{RdfCloudTripleStore}} with the correct DAO, notice we add an {{InferencingEngine}} as well to the store. If this is not added, then no inferencing will be done on the queries:
-
-{code}
-//setup
-Connector connector = new ZooKeeperInstance("cbinstance", "zkserver:port").getConnector("cbuser", "cbpassword");
-final RdfCloudTripleStore store = new RdfCloudTripleStore();
-AccumuloRyaDAO crdfdao = new AccumuloRyaDAO();
-crdfdao.setConnector(connector);
-
-AccumuloRdfConfiguration conf = new AccumuloRdfConfiguration();
-conf.setTablePrefix("rts_");
-conf.setDisplayQueryPlan(true);
-crdfdao.setConf(conf);
-store.setRdfDao(crdfdao);
-
-ProspectorServiceEvalStatsDAO evalDao = new ProspectorServiceEvalStatsDAO(connector, conf);
-evalDao.init();
-store.setRdfEvalStatsDAO(evalDao);
-
-InferenceEngine inferenceEngine = new InferenceEngine();
-inferenceEngine.setRdfDao(crdfdao);
-inferenceEngine.setConf(conf);
-store.setInferenceEngine(inferenceEngine);
-
-Repository myRepository = new RyaSailRepository(store);
-myRepository.initialize();
-RepositoryConnection conn = myRepository.getConnection();
-
-//query code goes here
-
-//close
-conn.close();
-myRepository.shutDown();
-{code}
-
-h2. Samples
-
-We will go through some quick samples on loading inferred relationships, seeing and diagnosing the query plan, and checking the data
-
-h3. Rdfs:SubClassOf
-
-First the code, which will load the following subclassof relationship: {{UndergraduateStudent subclassof Student subclassof Person}}. Then we will load into the tables three triples defining {{UgradA rdf:type UndergraduateStudent, StudentB rdf:type Student, PersonC rdf:type Person}}
-
-{code}
-conn.add(new StatementImpl(vf.createURI(litdupsNS, "UndergraduateStudent"), RDFS.SUBCLASSOF, vf.createURI(litdupsNS, "Student")));
-conn.add(new StatementImpl(vf.createURI(litdupsNS, "Student"), RDFS.SUBCLASSOF, vf.createURI(litdupsNS, "Person")));
-conn.add(new StatementImpl(vf.createURI(litdupsNS, "UgradA"), RDF.TYPE, vf.createURI(litdupsNS, "UndergraduateStudent")));
-conn.add(new StatementImpl(vf.createURI(litdupsNS, "StudentB"), RDF.TYPE, vf.createURI(litdupsNS, "Student")));
-conn.add(new StatementImpl(vf.createURI(litdupsNS, "PersonC"), RDF.TYPE, vf.createURI(litdupsNS, "Person")));
-conn.commit();
-{code}
-
-Remember that once the model is committed, it may take up to 5 minutes for the inferred relationships to be ready. Though you can override this property in the {{InferencingEngine}}.
-
-We shall run the following query:
-
-{code}
-PREFIX rdfs: <http://www.w3.org/2000/01/rdf-schema#>
-PREFIX rdf: <http://www.w3.org/1999/02/22-rdf-syntax-ns#>
-PREFIX lit: <urn:test:litdups#>
-select * where {?s rdf:type lit:Person.}
-{code}
-
-And should get back the following results:
-{code}
-[s=urn:test:litdups#StudentB]
-[s=urn:test:litdups#PersonC]
-[s=urn:test:litdups#UgradA]
-{code}
-
-h4. How it works
-
-Let us look at the query plan:
-{code}
-QueryRoot
-   Projection
-      ProjectionElemList
-         ProjectionElem "s"
-      Join
-         FixedStatementPattern
-            Var (name=79f261ee-e930-4af1-bc09-e637cc0affef)
-            Var (name=c-79f261ee-e930-4af1-bc09-e637cc0affef, value=http://www.w3.org/2000/01/rdf-schema#subClassOf)
-            Var (name=-const-2, value=urn:test:litdups#Person, anonymous)
-         DoNotExpandSP
-            Var (name=s)
-            Var (name=-const-1, value=http://www.w3.org/1999/02/22-rdf-syntax-ns#type, anonymous)
-            Var (name=79f261ee-e930-4af1-bc09-e637cc0affef)
-{code}
-
-Basically, we first find out (through the InferencingEngine) what triples have subclassof with Person. The InferencingEngine will do the graph analysis to find the both Student and UndergraduateStudent are Person classes.
-Then this information is joined with the statement pattern {{(?s rdf:type ?inf)}} where {{?inf}} is the results from the InferencingEngine.
-
-h3. Rdfs:SubPropertyOf
-
-SubPropertyOf defines that a property can be an instance of another property. For example, a {{gradDegreeFrom subPropertyOf degreeFrom}}.
-
-Also, EquivalentProperty can be thought of as specialized SubPropertyOf relationship where if {{propA equivalentProperty propB}} then that means that {{propA subPropertyOf propB AND propB subPropertyOf propA}}
-
-Sample Code:
-{code}
-conn.add(new StatementImpl(vf.createURI(litdupsNS, "undergradDegreeFrom"), RDFS.SUBPROPERTYOF, vf.createURI(litdupsNS, "degreeFrom")));
-conn.add(new StatementImpl(vf.createURI(litdupsNS, "gradDegreeFrom"), RDFS.SUBPROPERTYOF, vf.createURI(litdupsNS, "degreeFrom")));
-conn.add(new StatementImpl(vf.createURI(litdupsNS, "degreeFrom"), RDFS.SUBPROPERTYOF, vf.createURI(litdupsNS, "memberOf")));
-conn.add(new StatementImpl(vf.createURI(litdupsNS, "memberOf"), RDFS.SUBPROPERTYOF, vf.createURI(litdupsNS, "associatedWith")));
-conn.add(new StatementImpl(vf.createURI(litdupsNS, "UgradA"), vf.createURI(litdupsNS, "undergradDegreeFrom"), vf.createURI(litdupsNS, "Harvard")));
-conn.add(new StatementImpl(vf.createURI(litdupsNS, "GradB"), vf.createURI(litdupsNS, "gradDegreeFrom"), vf.createURI(litdupsNS, "Yale")));
-conn.add(new StatementImpl(vf.createURI(litdupsNS, "ProfessorC"), vf.createURI(litdupsNS, "memberOf"), vf.createURI(litdupsNS, "Harvard")));
-conn.commit();
-{code}
-
-With query:
-{code}
-PREFIX rdfs: <http://www.w3.org/2000/01/rdf-schema#>
-PREFIX rdf: <http://www.w3.org/1999/02/22-rdf-syntax-ns#>
-PREFIX lit: <urn:test:litdups#>
-select * where {?s lit:memberOf lit:Harvard.}
-{code}
-
-Will return results:
-{code}
-[s=urn:test:litdups#UgradA]
-[s=urn:test:litdups#ProfessorC]
-{code}
-
-Since UgradA has undergraduateDegreeFrom Harvard and ProfessorC is memberOf Harvard.
-
-h4. How it works
-
-This is very similar to the subClassOf relationship above. Basically the InferencingEngine provides what properties are subPropertyOf relationships with memberOf, and the second part of the Join checks to see if those properties are predicates with object "Harvard".
-
-Query Plan:
-{code}
-QueryRoot
-   Projection
-      ProjectionElemList
-         ProjectionElem "s"
-      Join
-         FixedStatementPattern
-            Var (name=0bad69f3-4769-4293-8318-e828b23dc52a)
-            Var (name=c-0bad69f3-4769-4293-8318-e828b23dc52a, value=http://www.w3.org/2000/01/rdf-schema#subPropertyOf)
-            Var (name=-const-1, value=urn:test:litdups#memberOf, anonymous)
-         DoNotExpandSP
-            Var (name=s)
-            Var (name=0bad69f3-4769-4293-8318-e828b23dc52a)
-            Var (name=-const-2, value=urn:test:litdups#Harvard, anonymous)
-{code}
-
-h3. InverseOf
-
-InverseOf defines a property that is an inverse relation of another property. For example, a student who has a {{degreeFrom}} a University also means that the University {{hasAlumnus}} student.
-
-Code:
-{code}
-conn.add(new StatementImpl(vf.createURI(litdupsNS, "degreeFrom"), OWL.INVERSEOF, vf.createURI(litdupsNS, "hasAlumnus")));
-conn.add(new StatementImpl(vf.createURI(litdupsNS, "UgradA"), vf.createURI(litdupsNS, "degreeFrom"), vf.createURI(litdupsNS, "Harvard")));
-conn.add(new StatementImpl(vf.createURI(litdupsNS, "GradB"), vf.createURI(litdupsNS, "degreeFrom"), vf.createURI(litdupsNS, "Harvard")));
-conn.add(new StatementImpl(vf.createURI(litdupsNS, "Harvard"), vf.createURI(litdupsNS, "hasAlumnus"), vf.createURI(litdupsNS, "AlumC")));
-conn.commit();
-{code}
-
-Query:
-{code}
-PREFIX rdfs: <http://www.w3.org/2000/01/rdf-schema#>
-PREFIX rdf: <http://www.w3.org/1999/02/22-rdf-syntax-ns#>
-PREFIX lit: <urn:test:litdups#>
-select * where {lit:Harvard lit:hasAlumnus ?s.}
-{code}
-
-Result:
-{code}
-[s=urn:test:litdups#AlumC]
-[s=urn:test:litdups#GradB]
-[s=urn:test:litdups#UgradA]
-{code}
-
-h4. How it works
-
-The query planner will expand the statement pattern {{Harvard hasAlumnus ?s}} to a Union between {{Harvard hasAlumnus ?s. and ?s degreeFrom Harvard}}
-
-As a caveat, it is important to note that in general Union queries do not have the best performance, so having a property that has an inverseOf and subPropertyOf, could cause a query plan that might take long depending on how the query planner orders the joins.
-
-Query Plan
-{code}
-QueryRoot
-   Projection
-      ProjectionElemList
-         ProjectionElem "s"
-      InferUnion
-         StatementPattern
-            Var (name=-const-1, value=urn:test:litdups#Harvard, anonymous)
-            Var (name=-const-2, value=urn:test:litdups#hasAlumnus, anonymous)
-            Var (name=s)
-         StatementPattern
-            Var (name=s)
-            Var (name=-const-2, value=urn:test:litdups#degreeFrom)
-            Var (name=-const-1, value=urn:test:litdups#Harvard, anonymous)
-{code}
-
-h3. SymmetricProperty
-
-SymmetricProperty defines a relationship where, for example, if Bob is a friendOf Jeff, then Jeff is a friendOf Bob. (Hopefully)
-
-Code:
-{code}
-conn.add(new StatementImpl(vf.createURI(litdupsNS, "friendOf"), RDF.TYPE, OWL.SYMMETRICPROPERTY));
-conn.add(new StatementImpl(vf.createURI(litdupsNS, "Bob"), vf.createURI(litdupsNS, "friendOf"), vf.createURI(litdupsNS, "Jeff")));
-conn.add(new StatementImpl(vf.createURI(litdupsNS, "James"), vf.createURI(litdupsNS, "friendOf"), vf.createURI(litdupsNS, "Jeff")));
-conn.commit();
-{code}
-
-Query:
-{code}
-PREFIX rdfs: <http://www.w3.org/2000/01/rdf-schema#>
-PREFIX rdf: <http://www.w3.org/1999/02/22-rdf-syntax-ns#>
-PREFIX lit: <urn:test:litdups#>
-select * where {?s lit:friendOf lit:Bob.}
-{code}
-
-Results:
-{code}
-[s=urn:test:litdups#Jeff]
-{code}
-
-h4. How it works
-
-The query planner will recognize that {{friendOf}} is a SymmetricProperty and devise a Union to find the specified relationship and inverse.
-
-Query Plan:
-{code}
-QueryRoot
-   Projection
-      ProjectionElemList
-         ProjectionElem "s"
-      InferUnion
-         StatementPattern
-            Var (name=s)
-            Var (name=-const-1, value=urn:test:litdups#friendOf, anonymous)
-            Var (name=-const-2, value=urn:test:litdups#Bob, anonymous)
-         StatementPattern
-            Var (name=-const-2, value=urn:test:litdups#Bob, anonymous)
-            Var (name=-const-1, value=urn:test:litdups#friendOf, anonymous)
-            Var (name=s)
-{code}
-
-h3. TransitiveProperty
-
-TransitiveProperty provides a transitive relationship between resources. For example, if Queens is subRegionOf NYC and NYC is subRegionOf NY, then Queens is transitively a subRegionOf NY.
-
-Code:
-{code}
-conn.add(new StatementImpl(vf.createURI(litdupsNS, "subRegionOf"), RDF.TYPE, OWL.TRANSITIVEPROPERTY));
-conn.add(new StatementImpl(vf.createURI(litdupsNS, "Queens"), vf.createURI(litdupsNS, "subRegionOf"), vf.createURI(litdupsNS, "NYC")));
-conn.add(new StatementImpl(vf.createURI(litdupsNS, "NYC"), vf.createURI(litdupsNS, "subRegionOf"), vf.createURI(litdupsNS, "NY")));
-conn.add(new StatementImpl(vf.createURI(litdupsNS, "NY"), vf.createURI(litdupsNS, "subRegionOf"), vf.createURI(litdupsNS, "US")));
-conn.add(new StatementImpl(vf.createURI(litdupsNS, "US"), vf.createURI(litdupsNS, "subRegionOf"), vf.createURI(litdupsNS, "NorthAmerica")));
-conn.add(new StatementImpl(vf.createURI(litdupsNS, "NorthAmerica"), vf.createURI(litdupsNS, "subRegionOf"), vf.createURI(litdupsNS, "World")));
-conn.commit();
-{code}
-
-Query:
-{code}
-PREFIX rdfs: <http://www.w3.org/2000/01/rdf-schema#>
-PREFIX rdf: <http://www.w3.org/1999/02/22-rdf-syntax-ns#>
-PREFIX lit: <urn:test:litdups#>
-select * where {?s lit:subRegionOf lit:NorthAmerica.}
-{code}
-
-Results:
-{code}
-[s=urn:test:litdups#Queens]
-[s=urn:test:litdups#NYC]
-[s=urn:test:litdups#NY]
-[s=urn:test:litdups#US]
-{code}
-
-h4. How it works
-
-The TransitiveProperty relationship works by running recursive queries till all the results are returned.
-
-It is important to note that certain TransitiveProperty relationships will not work:
-* Open ended property: ?s subRegionOf ?o (At least one of the properties must be filled or will be filled as the query gets answered)
-* Closed property: Queens subRegionOf NY (At least one of the properties must be empty)
-
-We are working on fixing these issues.
-
-Query Plan:
-{code}
-QueryRoot
-   Projection
-      ProjectionElemList
-         ProjectionElem "s"
-      TransitivePropertySP
-         Var (name=s)
-         Var (name=-const-1, value=urn:test:litdups#subRegionOf, anonymous)
-         Var (name=-const-2, value=urn:test:litdups#NorthAmerica, anonymous)
-{code}
\ No newline at end of file

http://git-wip-us.apache.org/repos/asf/incubator-rya/blob/ce4a10ff/extras/rya.manual/src/main/webapp/sm-namedgraph.conf
----------------------------------------------------------------------
diff --git a/extras/rya.manual/src/main/webapp/sm-namedgraph.conf b/extras/rya.manual/src/main/webapp/sm-namedgraph.conf
deleted file mode 100644
index ce2a0da..0000000
--- a/extras/rya.manual/src/main/webapp/sm-namedgraph.conf
+++ /dev/null
@@ -1,129 +0,0 @@
-h1. Named Graphs
-
-Named graphs are supported simply in the Rdf Store in a few ways. OpenRdf supports sending {{contexts}} as each triple is saved.
-
-h2. Simple Named Graph Load and Query
-
-Here is a very simple example of using the API to Insert data in named graphs and querying with Sparql
-
-First we will define a Trig document to load
-Trig document
-{code}
-@prefix rdf: <http://www.w3.org/1999/02/22-rdf-syntax-ns#> .
-@prefix xsd: <http://www.w3.org/2001/XMLSchema#> .
-@prefix swp: <http://www.w3.org/2004/03/trix/swp-1/> .
-@prefix dc: <http://purl.org/dc/elements/1.1/> .
-@prefix ex: <http://www.example.org/vocabulary#> .
-@prefix : <http://www.example.org/exampleDocument#> .
-:G1 { :Monica ex:name "Monica Murphy" .
-      :Monica ex:homepage <http://www.monicamurphy.org> .
-      :Monica ex:email <ma...@monicamurphy.org> .
-      :Monica ex:hasSkill ex:Management }
-
-:G2 { :Monica rdf:type ex:Person .
-      :Monica ex:hasSkill ex:Programming }
-
-:G4 { :Phobe ex:name "Phobe Buffet" }
-
-:G3 { :G1 swp:assertedBy _:w1 .
-      _:w1 swp:authority :Chris .
-      _:w1 dc:date "2003-10-02"^^xsd:date .
-      :G2 swp:quotedBy _:w2 .
-      :G4 swp:assertedBy _:w2 .
-      _:w2 dc:date "2003-09-03"^^xsd:date .
-      _:w2 swp:authority :Tom .
-      :Chris rdf:type ex:Person .
-      :Chris ex:email <ma...@bizer.de>.
-      :Tom rdf:type ex:Person .
-      :Tom ex:email <ma...@bizer.de>}
-{code}
-
-We will assume that this file is saved on your classpath somewhere at {{<TRIG_FILE>}}
-
-Load data through API:
-{code}
-InputStream stream = Thread.currentThread().getContextClassLoader().getResourceAsStream("namedgraphs.trig");
-RepositoryConnection conn = repository.getConnection();
-conn.add(stream, "", RDFFormat.TRIG);
-conn.commit();
-{code}
-
-Now that the data is loaded we can easily query it. For example, we will query to find what {{hasSkill}} is defined in graph G2, and relate that to someone defined in G1.
-
-Query:
-{code}
-PREFIX  ex:  <http://www.example.org/exampleDocument#>
-PREFIX  voc:  <http://www.example.org/vocabulary#>
-PREFIX  foaf:  <http://xmlns.com/foaf/0.1/>
-PREFIX  rdfs:  <http://www.w3.org/2000/01/rdf-schema#>
-
-SELECT *
-WHERE
-{
-  GRAPH ex:G1
-  {
-    ?m voc:name ?name ;
-       voc:homepage ?hp .
-  } .
- GRAPH ex:G2
-  {
-    ?m voc:hasSkill ?skill .
-  } .
-}
-{code}
-
-Results:
-{code}
-[hp=http://www.monicamurphy.org;m=http://www.example.org/exampleDocument#Monica;skill=http://www.example.org/vocabulary#Programming;name="Monica Murphy"]
-{code}
-
-Here is the Query Plan as well:
-{code}
-QueryRoot
-   Projection
-      ProjectionElemList
-         ProjectionElem "m"
-         ProjectionElem "name"
-         ProjectionElem "hp"
-         ProjectionElem "skill"
-      Join
-         Join
-            StatementPattern FROM NAMED CONTEXT
-               Var (name=m)
-               Var (name=-const-2, value=http://www.example.org/vocabulary#name, anonymous)
-               Var (name=name)
-               Var (name=-const-1, value=http://www.example.org/exampleDocument#G1, anonymous)
-            StatementPattern FROM NAMED CONTEXT
-               Var (name=m)
-               Var (name=-const-3, value=http://www.example.org/vocabulary#homepage, anonymous)
-               Var (name=hp)
-               Var (name=-const-1, value=http://www.example.org/exampleDocument#G1, anonymous)
-         StatementPattern FROM NAMED CONTEXT
-            Var (name=m)
-            Var (name=-const-5, value=http://www.example.org/vocabulary#hasSkill, anonymous)
-            Var (name=skill)
-            Var (name=-const-4, value=http://www.example.org/exampleDocument#G2, anonymous)
-{code}
-
-h2. Inserting named graph data through Sparql
-
-The new Sparql update standard provides another way to insert data, even into named graphs.
-
-First the insert update:
-{code}
-PREFIX dc: <http://purl.org/dc/elements/1.1/>
-PREFIX ex: <http://example/addresses#>
-INSERT DATA
-{
-    GRAPH ex:G1 {
-        <http://example/book3> dc:title    "A new book" ;
-                               dc:creator  "A.N.Other" .
-    }
-}
-{code}
-
-To perform this update, it requires different code than querying the data directly:
-{code}
-Update update = conn.prepareUpdate(QueryLanguage.SPARQL, insert);
-update.execute();
-{code}
\ No newline at end of file

http://git-wip-us.apache.org/repos/asf/incubator-rya/blob/ce4a10ff/extras/rya.manual/src/main/webapp/sm-simpleaqr.conf
----------------------------------------------------------------------
diff --git a/extras/rya.manual/src/main/webapp/sm-simpleaqr.conf b/extras/rya.manual/src/main/webapp/sm-simpleaqr.conf
deleted file mode 100644
index 7616197..0000000
--- a/extras/rya.manual/src/main/webapp/sm-simpleaqr.conf
+++ /dev/null
@@ -1,54 +0,0 @@
-h1. Simple Add Query and Remove of Statements
-
-This quick tutorial will give a small example on how to add, query, and remove statements from Cloudbase
-
-h2. Code
-
-{code}
-//setup
-Connector connector = new ZooKeeperInstance("cbinstance", "zkserver:port").getConnector("cbuser", "cbpassword");
-final RdfCloudTripleStore store = new RdfCloudTripleStore();
-AccumuloRyaDAO crdfdao = new AccumuloRyaDAO();
-crdfdao.setConnector(connector);
-
-AccumuloRdfConfiguration conf = new AccumuloRdfConfiguration();
-conf.setTablePrefix("rts_");
-conf.setDisplayQueryPlan(true);
-crdfdao.setConf(conf);
-store.setRdfDao(crdfdao);
-
-ProspectorServiceEvalStatsDAO evalDao = new ProspectorServiceEvalStatsDAO(connector, conf);
-evalDao.init();
-store.setRdfEvalStatsDAO(evalDao);
-
-InferenceEngine inferenceEngine = new InferenceEngine();
-inferenceEngine.setRdfDao(crdfdao);
-inferenceEngine.setConf(conf);
-store.setInferenceEngine(inferenceEngine);
-
-Repository myRepository = new RyaSailRepository(store);
-myRepository.initialize();
-RepositoryConnection conn = myRepository.getConnection();
-
-//define and add statement
-String litdupsNS = "urn:test:litdups#";
-URI cpu = vf.createURI(litdupsNS, "cpu");
-URI loadPerc = vf.createURI(litdupsNS, "loadPerc");
-URI uri1 = vf.createURI(litdupsNS, "uri1");
-conn.add(cpu, loadPerc, uri1);
-conn.commit();
-
-//query for all statements that have subject=cpu and pred=loadPerc (wildcard object)
-RepositoryResult<Statement> result = conn.getStatements(cpu, loadPerc, null, true)
-while(result.hasNext()) {
-    System.out.println(result.next());
-}
-result.close();
-
-//remove statement
-conn.remove(cpu, loadPerc, uri1);
-
-//close
-conn.close();
-myRepository.shutDown();
-{code}
\ No newline at end of file

http://git-wip-us.apache.org/repos/asf/incubator-rya/blob/ce4a10ff/extras/rya.manual/src/main/webapp/sm-sparqlquery.conf
----------------------------------------------------------------------
diff --git a/extras/rya.manual/src/main/webapp/sm-sparqlquery.conf b/extras/rya.manual/src/main/webapp/sm-sparqlquery.conf
deleted file mode 100644
index 985e21a..0000000
--- a/extras/rya.manual/src/main/webapp/sm-sparqlquery.conf
+++ /dev/null
@@ -1,58 +0,0 @@
-h1. Simple Add Query and Remove of Statements
-
-This quick tutorial will give a small example on how to query data with Sparql
-
-h2. Code
-
-{code}
-//setup
-Connector connector = new ZooKeeperInstance("cbinstance", "zkserver:port").getConnector("cbuser", "cbpassword");
-final RdfCloudTripleStore store = new RdfCloudTripleStore();
-AccumuloRyaDAO crdfdao = new AccumuloRyaDAO();
-crdfdao.setConnector(connector);
-
-AccumuloRdfConfiguration conf = new AccumuloRdfConfiguration();
-conf.setTablePrefix("rts_");
-conf.setDisplayQueryPlan(true);
-crdfdao.setConf(conf);
-store.setRdfDao(crdfdao);
-
-ProspectorServiceEvalStatsDAO evalDao = new ProspectorServiceEvalStatsDAO(connector, conf);
-evalDao.init();
-store.setRdfEvalStatsDAO(evalDao);
-
-InferenceEngine inferenceEngine = new InferenceEngine();
-inferenceEngine.setRdfDao(crdfdao);
-inferenceEngine.setConf(conf);
-store.setInferenceEngine(inferenceEngine);
-
-Repository myRepository = new RyaSailRepository(store);
-myRepository.initialize();
-RepositoryConnection conn = myRepository.getConnection();
-
-//define and add statements
-String litdupsNS = "urn:test:litdups#";
-URI cpu = vf.createURI(litdupsNS, "cpu");
-URI loadPerc = vf.createURI(litdupsNS, "loadPerc");
-URI uri1 = vf.createURI(litdupsNS, "uri1");
-URI pred2 = vf.createURI(litdupsNS, "pred2");
-URI uri2 = vf.createURI(litdupsNS, "uri2");
-conn.add(cpu, loadPerc, uri1);
-conn.commit();
-
-//query using sparql
-String query = "select * where {" +
-                "?x <" + loadPerc.stringValue() + "> ?o1." +
-                "?x <" + pred2.stringValue() + "> ?o2." +
-                "}";
-TupleQuery tupleQuery = conn.prepareTupleQuery(QueryLanguage.SPARQL, query);
-TupleQueryResult result = tupleQuery.evaluate();
-while(result.hasNext()) {
-      System.out.println(result.next());
-}
-result.close();
-
-//close
-conn.close();
-myRepository.shutDown();
-{code}
\ No newline at end of file

http://git-wip-us.apache.org/repos/asf/incubator-rya/blob/ce4a10ff/extras/rya.manual/src/main/webapp/sm-updatedata.conf
----------------------------------------------------------------------
diff --git a/extras/rya.manual/src/main/webapp/sm-updatedata.conf b/extras/rya.manual/src/main/webapp/sm-updatedata.conf
deleted file mode 100644
index b3e3a5b..0000000
--- a/extras/rya.manual/src/main/webapp/sm-updatedata.conf
+++ /dev/null
@@ -1,56 +0,0 @@
-h1. Sparql Update
-
-OpenRDF supports the Sparql Update functionality. Here are a few samples:
-
-Remember, you have to use {{RepositoryConnection.prepareUpdate(..)}} to perform these queries
-
-Insert:
-{code}
-PREFIX dc: <http://purl.org/dc/elements/1.1/>
-INSERT DATA
-{ <http://example/book3> dc:title    "A new book" ;
-                         dc:creator  "A.N.Other" .
-}
-{code}
-
-Delete:
-{code}
-PREFIX dc: <http://purl.org/dc/elements/1.1/>
-DELETE DATA
-{ <http://example/book3> dc:title    "A new book" ;
-                         dc:creator  "A.N.Other" .
-}
-{code}
-
-Update:
-{code}
-PREFIX dc: <http://purl.org/dc/elements/1.1/>
-DELETE { ?book dc:title ?title }
-INSERT { ?book dc:title "A newer book".         ?book dc:add "Additional Info" }
-WHERE
-  { ?book dc:creator "A.N.Other" .
-  }
-{code}
-
-Insert Named Graph:
-{code}
-PREFIX dc: <http://purl.org/dc/elements/1.1/>
-PREFIX ex: <http://example/addresses#>
-INSERT DATA
-{ GRAPH ex:G1 {
-<http://example/book3> dc:title    "A new book" ;
-                         dc:creator  "A.N.Other" .
-}
-}
-{code}
-
-Update Named Graph:
-{code}
-PREFIX dc: <http://purl.org/dc/elements/1.1/>
-WITH <http://example/addresses#G1>
-DELETE { ?book dc:title ?title }
-INSERT { ?book dc:title "A newer book".         ?book dc:add "Additional Info" }
-WHERE
-  { ?book dc:creator "A.N.Other" .
-  }
-{code}



[3/3] incubator-rya git commit: RYA-3 Changing Manual to Markdown; linking README

Posted by mi...@apache.org.
RYA-3 Changing Manual to Markdown; linking README


Project: http://git-wip-us.apache.org/repos/asf/incubator-rya/repo
Commit: http://git-wip-us.apache.org/repos/asf/incubator-rya/commit/ce4a10ff
Tree: http://git-wip-us.apache.org/repos/asf/incubator-rya/tree/ce4a10ff
Diff: http://git-wip-us.apache.org/repos/asf/incubator-rya/diff/ce4a10ff

Branch: refs/heads/develop
Commit: ce4a10ff5eafa2b85cb6829e5876956d5dd2278a
Parents: 2a6f730
Author: Aaron Mihalik <mi...@alum.mit.edu>
Authored: Sun Nov 22 12:27:56 2015 -0500
Committer: Aaron Mihalik <mi...@alum.mit.edu>
Committed: Sun Nov 22 16:17:24 2015 -0500

----------------------------------------------------------------------
 README.md                                       |  230 +--
 extras/rya.manual/pom.xml                       |  274 +--
 extras/rya.manual/src/main/java/temp/TempJ.java |   29 -
 extras/rya.manual/src/main/webapp/.livereload   |   19 -
 .../WEB-INF/scalate/layouts/default.scaml       |   97 --
 .../webapp/WEB-INF/scalate/layouts/print.ssp    |  778 ---------
 .../rya.manual/src/main/webapp/WEB-INF/web.xml  |   50 -
 extras/rya.manual/src/main/webapp/_banner.ssp   |    8 -
 .../rya.manual/src/main/webapp/_navigation.conf |   24 -
 .../rya.manual/src/main/webapp/_quicklinks.ssp  |    5 -
 extras/rya.manual/src/main/webapp/alx.conf      |   59 -
 .../src/main/webapp/build-source.conf           |   19 -
 .../src/main/webapp/css/pygmentize.css          |  134 --
 .../src/main/webapp/css/scalate/console.css     |   71 -
 extras/rya.manual/src/main/webapp/css/style.css | 1588 ------------------
 extras/rya.manual/src/main/webapp/eval.conf     |   56 -
 extras/rya.manual/src/main/webapp/images/bg.png |  Bin 244 -> 0 bytes
 .../src/main/webapp/images/bg02-blue-left.png   |  Bin 1068 -> 0 bytes
 .../src/main/webapp/images/bg02-blue-right.png  |  Bin 1114 -> 0 bytes
 .../main/webapp/images/bg02-white-left-nogr.png |  Bin 951 -> 0 bytes
 .../webapp/images/bg02-white-right-nogr.png     |  Bin 958 -> 0 bytes
 .../src/main/webapp/images/big-bullet.png       |  Bin 285 -> 0 bytes
 .../src/main/webapp/images/left-box-bottom.png  |  Bin 588 -> 0 bytes
 .../src/main/webapp/images/left-box-right.png   |  Bin 213 -> 0 bytes
 .../src/main/webapp/images/left-box-top.png     |  Bin 466 -> 0 bytes
 .../rya.manual/src/main/webapp/images/logo.png  |  Bin 18119 -> 0 bytes
 .../rya.manual/src/main/webapp/images/logo2.png |  Bin 9271 -> 0 bytes
 extras/rya.manual/src/main/webapp/index.conf    |    4 -
 extras/rya.manual/src/main/webapp/infer.conf    |   13 -
 .../src/main/webapp/loadPrecomputedJoin.conf    |   27 -
 extras/rya.manual/src/main/webapp/loaddata.conf |  117 --
 extras/rya.manual/src/main/webapp/overview.conf |    4 -
 .../rya.manual/src/main/webapp/querydata.conf   |  115 --
 .../rya.manual/src/main/webapp/quickstart.conf  |   42 -
 .../rya.manual/src/main/webapp/sm-addauth.conf  |   98 --
 .../src/main/webapp/sm-firststeps.conf          |   55 -
 extras/rya.manual/src/main/webapp/sm-infer.conf |  313 ----
 .../src/main/webapp/sm-namedgraph.conf          |  129 --
 .../src/main/webapp/sm-simpleaqr.conf           |   54 -
 .../src/main/webapp/sm-sparqlquery.conf         |   58 -
 .../src/main/webapp/sm-updatedata.conf          |   56 -
 extras/rya.manual/src/site/markdown/_index.md   |   23 +
 extras/rya.manual/src/site/markdown/alx.md      |   61 +
 .../src/site/markdown/build-source.md           |   15 +
 extras/rya.manual/src/site/markdown/eval.md     |   58 +
 extras/rya.manual/src/site/markdown/index.md    |   24 +
 extras/rya.manual/src/site/markdown/infer.md    |   14 +
 .../src/site/markdown/loadPrecomputedJoin.md    |   28 +
 extras/rya.manual/src/site/markdown/loaddata.md |  121 ++
 extras/rya.manual/src/site/markdown/overview.md |    5 +
 .../rya.manual/src/site/markdown/querydata.md   |  116 ++
 .../rya.manual/src/site/markdown/quickstart.md  |   41 +
 .../rya.manual/src/site/markdown/sm-addauth.md  |   98 ++
 .../src/site/markdown/sm-firststeps.md          |   59 +
 extras/rya.manual/src/site/markdown/sm-infer.md |  332 ++++
 .../src/site/markdown/sm-namedgraph.md          |  136 ++
 .../src/site/markdown/sm-simpleaqr.md           |   54 +
 .../src/site/markdown/sm-sparqlquery.md         |   58 +
 .../src/site/markdown/sm-updatedata.md          |   62 +
 .../src/site/resources/js/fixmarkdownlinks.js   |    6 +
 extras/rya.manual/src/site/site.xml             |   45 +
 61 files changed, 1486 insertions(+), 4396 deletions(-)
----------------------------------------------------------------------


http://git-wip-us.apache.org/repos/asf/incubator-rya/blob/ce4a10ff/README.md
----------------------------------------------------------------------
diff --git a/README.md b/README.md
index 04628cc..bc548a0 100644
--- a/README.md
+++ b/README.md
@@ -2,10 +2,14 @@
 
 ## Overview
 
-RYA is a scalable RDF Store that is built on top of a Columnar Index Store (such as Accumulo, Hbase(not currently supported)). It is implemented as an extension to OpenRdf to provide easy query mechanisms (SPARQL, SERQL, etc) and Rdf data storage (RDF/XML, NTriples, etc).
+RYA is a scalable RDF Store that is built on top of a Columnar Index Store (such as Accumulo). It is implemented as an extension to OpenRdf to provide easy query mechanisms (SPARQL, SERQL, etc) and Rdf data storage (RDF/XML, NTriples, etc).
 
 RYA stands for RDF y(and) Accumulo.
 
+## Rya Manual
+
+A copy of the Rya Manual is located [here](extras/rya.manual/src/site/markdown/index.md).  The material in the manual and below may be out of sync.
+
 ## Upgrade Path
 
 Since the data encodings changed in the 3.2.2 release, you will need to run the Upgrade322Tool MapReduce job to perform the upgrade.
@@ -13,7 +17,10 @@ Since the data encodings changed in the 3.2.2 release, you will need to run the
 1. Build the project with -Pmr to build the mapreduce artifacts
 2. Make sure to clone the rya tables before doing the upgrade
 3. Run
-    <pre> hadoop jar accumulo.rya-mr.jar mvm.rya.accumulo.mr.upgrade.Upgrade322Tool -Dac.instance={} -Dac.username={} -Dac.pwd={} </pre>
+    
+```
+hadoop jar accumulo.rya-mr.jar mvm.rya.accumulo.mr.upgrade.Upgrade322Tool -Dac.instance={} -Dac.username={} -Dac.pwd={}
+```
 
 ## Quick Start
 
@@ -22,41 +29,42 @@ This tutorial will outline the steps needed to get quickly started with the Rya
 ### Prerequisites
 
 * Columnar Store (either Accumulo) The tutorial will go forward using Accumulo
-* Rya code (Git: git@github.com:texeltek/rya.git)
-* Maven 2.2 +
+* Rya code (Git: git://git.apache.org/incubator-rya.git)
+* Maven 3.0 +
 
 ### Building from Source
 
 Using Git, pull down the latest code from the url above.
 
-Run the command to build the code {{mvn clean install}}
-
-If all goes well, the build should be successful and a war should be produced in {{web/web.rya/target/web.rya.war}}
+Run the command to build the code `mvn clean install`
 
-Note: To perform a build of the geomesa/lucene indexing, run the build with the profile 'indexing' (-P indexing)
+If all goes well, the build should be successful and a war should be produced in `web/web.rya/target/web.rya.war`
 
-Note: If you are building on windows, you will need hadoop-common 2.6.0's {{winutils.exe}} and {{hadoop.dll}}.  You can download it from [here](https://github.com/amihalik/hadoop-common-2.6.0-bin/archive/master.zip).  This build requires the [Visual C++ Redistributable for Visual Studio 2015 (x64)](https://www.microsoft.com/en-us/download/details.aspx?id=48145).   Also you will need to set your path and Hadoop home using the commands below:
+Note: To perform a build of the geomesa/lucene indexing, run the build with the profile 'indexing' `mvn clean install -P indexing`
 
-    set HADOOP_HOME=c:\hadoop-common-2.6.0-bin
-    set PATH=%PATH%;c:\hadoop-common-2.6.0-bin\bin
+Note: If you are building on windows, you will need hadoop-common 2.6.0's `winutils.exe` and `hadoop.dll`.  You can download it from [here](https://github.com/amihalik/hadoop-common-2.6.0-bin/archive/master.zip).  This build requires the [Visual C++ Redistributable for Visual Studio 2015 (x64)](https://www.microsoft.com/en-us/download/details.aspx?id=48145).   Also you will need to set your path and Hadoop home using the commands below:
 
-### Deployment
+```
+set HADOOP_HOME=c:\hadoop-common-2.6.0-bin
+set PATH=%PATH%;c:\hadoop-common-2.6.0-bin\bin
+```
 
-(Using tomcat)
+### Deployment Using Tomcat
 
 Unwar the above war into the webapps directory.
 
-To point the web.rya war to the appropriate Accumulo instance, make a properties file {{environment.properties}} and put it in the classpath. Here is an example:
+To point the web.rya war to the appropriate Accumulo instance, make a properties file `environment.properties` and put it in the classpath. Here is an example:
 
-    instance.name=accumulo  #Accumulo instance name
-    instance.zk=localhost:2181  #Accumulo Zookeepers
-    instance.username=root  #Accumulo username
-    instance.password=secret  #Accumulo pwd
-    rya.tableprefix=triplestore_  #Rya Table Prefix
-    rya.displayqueryplan=true  #To display the query plan
+```
+instance.name=accumulo  #Accumulo instance name
+instance.zk=localhost:2181  #Accumulo Zookeepers
+instance.username=root  #Accumulo username
+instance.password=secret  #Accumulo pwd
+rya.tableprefix=triplestore_  #Rya Table Prefix
+rya.displayqueryplan=true  #To display the query plan
+```
 
-
-Start the Tomcat server. {{./bin/startup.sh}}
+Start the Tomcat server. `./bin/startup.sh`
 
 ## Usage
 
@@ -64,29 +72,31 @@ Start the Tomcat server. {{./bin/startup.sh}}
 
 #### Web REST endpoint
 
-The War sets up a Web REST endpoint at {{http://server/web.rya/loadrdf}} that allows POST data to get loaded into the Rdf Store. This short tutorial will use Java code to post data.
+The War sets up a Web REST endpoint at `http://server/web.rya/loadrdf` that allows POST data to get loaded into the Rdf Store. This short tutorial will use Java code to post data.
 
 First, you will need data to load and will need to figure out what format that data is in.
 
 For this sample, we will use the following N-Triples:
 
-    <http://mynamespace/ProductType1> <http://www.w3.org/1999/02/22-rdf-syntax-ns#type> <http://mynamespace/ProductType> .
-    <http://mynamespace/ProductType1> <http://www.w3.org/2000/01/rdf-schema#label> "Thing" .
-    <http://mynamespace/ProductType1> <http://purl.org/dc/elements/1.1/publisher> <http://mynamespace/Publisher1> .
-
+```
+<http://mynamespace/ProductType1> <http://www.w3.org/1999/02/22-rdf-syntax-ns#type> <http://mynamespace/ProductType> .
+<http://mynamespace/ProductType1> <http://www.w3.org/2000/01/rdf-schema#label> "Thing" .
+<http://mynamespace/ProductType1> <http://purl.org/dc/elements/1.1/publisher> <http://mynamespace/Publisher1> .
+```
 
 Save this file somewhere $RDF_DATA
 
 Second, use the following Java code to load data to the REST endpoint:
 
-    import java.io.BufferedReader;
-    import java.io.InputStream;
-    import java.io.InputStreamReader;
-    import java.io.OutputStream;
-    import java.net.URL;
-    import java.net.URLConnection;
+``` JAVA
+import java.io.BufferedReader;
+import java.io.InputStream;
+import java.io.InputStreamReader;
+import java.io.OutputStream;
+import java.net.URL;
+import java.net.URLConnection;
 
-    public class LoadDataServletRun {
+public class LoadDataServletRun {
 
     public static void main(String[] args) {
         try {
@@ -120,8 +130,8 @@ Second, use the following Java code to load data to the REST endpoint:
             e.printStackTrace();
         }
     }
-    }
-
+}
+```
 
 Compile and run this code above, changing the references for $RDF_DATA and the url that your Rdf War is running at.
 
@@ -134,11 +144,13 @@ Bulk loading data is done through Map Reduce jobs
 ##### Bulk Load RDF data
 
   This Map Reduce job will read a full file into memory and parse it into statements. The statements are saved into the store. Here is an example for storing in Accumulo:
-{code}
+
+```
 hadoop jar target/accumulo.rya-3.0.4-SNAPSHOT-shaded.jar mvm.rya.accumulo.mr.fileinput.BulkNtripsInputTool -Dac.zk=localhost:2181 -Dac.instance=accumulo -Dac.username=root -Dac.pwd=secret -Drdf.tablePrefix=triplestore_ -Dio.sort.mb=64 /tmp/temp.ntrips
-{code}
+```
 
 Options:
+
 * rdf.tablePrefix : The tables (spo, po, osp) are prefixed with this qualifier. The tables become: (rdf.tablePrefix)spo,(rdf.tablePrefix)po,(rdf.tablePrefix)osp
 * ac.* : Accumulo connection parameters
 * rdf.format : See RDFFormat from openrdf, samples include (Trig, N-Triples, RDF/XML)
@@ -151,41 +163,41 @@ The argument is the directory/file to load. This file needs to be loaded into HD
 Here is some sample code to load data directly through the OpenRDF API. (Loading N-Triples data)
 You will need at least accumulo.rya-<version>, rya.api, rya.sail.impl on the classpath and transitive dependencies. I find that Maven is the easiest way to get a project dependency tree set up.
 
+``` JAVA
+final RdfCloudTripleStore store = new RdfCloudTripleStore();
+AccumuloRdfConfiguration conf = new AccumuloRdfConfiguration();
+AccumuloRyaDAO dao = new AccumuloRdfDAO();
+Connector connector = new ZooKeeperInstance("instance", "zoo1,zoo2,zoo3").getConnector("user", "password");
+dao.setConnector(connector);
+conf.setTablePrefix("rya_");
+dao.setConf(conf);
+store.setRdfDao(dao);
 
-            final RdfCloudTripleStore store = new RdfCloudTripleStore();
-            AccumuloRdfConfiguration conf = new AccumuloRdfConfiguration();
-            AccumuloRyaDAO dao = new AccumuloRdfDAO();
-            Connector connector = new ZooKeeperInstance("instance", "zkserver:port").getConnector("user", "password");
-            dao.setConnector(connector);
-            conf.setTablePrefix("rya_");
-            dao.setConf(conf);
-            store.setRdfDao(dao);
-
-            Repository myRepository = new RyaSailRepository(store);
-            myRepository.initialize();
-            RepositoryConnection conn = myRepository.getConnection();
+Repository myRepository = new RyaSailRepository(store);
+myRepository.initialize();
+RepositoryConnection conn = myRepository.getConnection();
 
-            //load data from file
-            final File file = new File("ntriples.ntrips");
-            conn.add(new FileInputStream(file), file.getName(),
-                    RDFFormat.NTRIPLES, new Resource[]{});
+//load data from file
+final File file = new File("ntriples.ntrips");
+conn.add(new FileInputStream(file), file.getName(),
+        RDFFormat.NTRIPLES, new Resource[]{});
 
-            conn.commit();
-
-            conn.close();
-            myRepository.shutDown();
+conn.commit();
 
+conn.close();
+myRepository.shutDown();
+```
 
 
 ### Query Data
 
 #### Web JSP endpoint
 
-Open a url to {{http://server/web.rya/sparqlQuery.jsp}}. This simple form can run Sparql.
+Open a url to `http://server/web.rya/sparqlQuery.jsp`. This simple form can run Sparql.
 
 ### Web REST endpoint
 
-The War sets up a Web REST endpoint at {{http://server/web.rya/queryrdf}} that allows GET requests with queries.
+The War sets up a Web REST endpoint at `http://server/web.rya/queryrdf` that allows GET requests with queries.
 
 For this sample, we will assume you already loaded data from the [loaddata.html] tutorial
 
@@ -193,13 +205,14 @@ Save this file somewhere $RDF_DATA
 
 Second, use the following Java code to load data to the REST endpoint:
 
-    import java.io.BufferedReader;
-    import java.io.InputStreamReader;
-    import java.net.URL;
-    import java.net.URLConnection;
-    import java.net.URLEncoder;
+``` JAVA
+import java.io.BufferedReader;
+import java.io.InputStreamReader;
+import java.net.URL;
+import java.net.URLConnection;
+import java.net.URLEncoder;
 
-    public class QueryDataServletRun {
+public class QueryDataServletRun {
 
     public static void main(String[] args) {
         try {
@@ -224,8 +237,8 @@ Second, use the following Java code to load data to the REST endpoint:
             e.printStackTrace();
         }
     }
-    }
-
+}
+```
 
 Compile and run this code above, changing the url that your Rdf War is running at.
 
@@ -233,54 +246,55 @@ Compile and run this code above, changing the url that your Rdf War is running a
 
 Here is a code snippet for directly running against Accumulo with the code. You will need at least accumulo.rya.jar, rya.api, rya.sail.impl on the classpath and transitive dependencies. I find that Maven is the easiest way to get a project dependency tree set up.
 
+``` JAVA
+Connector connector = new ZooKeeperInstance("instance", "zoo1,zoo2,zoo3").getConnector("user", "password");
 
-            Connector connector = new ZooKeeperInstance("cbinstance", "zkserver:port").getConnector("cbuser", "cbpassword");
+final RdfCloudTripleStore store = new RdfCloudTripleStore();
+AccumuloRyaDAO crdfdao = new AccumuloRyaDAO();
+crdfdao.setConnector(connector);
 
-            final RdfCloudTripleStore store = new RdfCloudTripleStore();
-            AccumuloRyaDAO crdfdao = new AccumuloRyaDAO();
-            crdfdao.setConnector(connector);
+AccumuloRdfConfiguration conf = new AccumuloRdfConfiguration();
+conf.setTablePrefix("rts_");
+conf.setDisplayQueryPlan(true);
+crdfdao.setConf(conf);
+store.setRdfDao(crdfdao);
 
-            AccumuloRdfConfiguration conf = new AccumuloRdfConfiguration();
-            conf.setTablePrefix("rts_");
-            conf.setDisplayQueryPlan(true);
-            crdfdao.setConf(conf);
-            store.setRdfDao(crdfdao);
+InferenceEngine inferenceEngine = new InferenceEngine();
+inferenceEngine.setRdfDao(crdfdao);
+inferenceEngine.setConf(conf);
+store.setInferenceEngine(inferenceEngine);
 
-            InferenceEngine inferenceEngine = new InferenceEngine();
-            inferenceEngine.setRdfDao(crdfdao);
-            inferenceEngine.setConf(conf);
-            store.setInferenceEngine(inferenceEngine);
+Repository myRepository = new RyaSailRepository(store);
+myRepository.initialize();
 
-            Repository myRepository = new RyaSailRepository(store);
-            myRepository.initialize();
-
-            String query = "select * where {\n" +
-                                "<http://mynamespace/ProductType1> ?p ?o.\n" +
-                                "}";
-            RepositoryConnection conn = myRepository.getConnection();
-            System.out.println(query);
-            TupleQuery tupleQuery = conn.prepareTupleQuery(
-                    QueryLanguage.SPARQL, query);
-            ValueFactory vf = ValueFactoryImpl.getInstance();
+String query = "select * where {\n" +
+                    "<http://mynamespace/ProductType1> ?p ?o.\n" +
+                    "}";
+RepositoryConnection conn = myRepository.getConnection();
+System.out.println(query);
+TupleQuery tupleQuery = conn.prepareTupleQuery(
+        QueryLanguage.SPARQL, query);
+ValueFactory vf = ValueFactoryImpl.getInstance();
 
-            TupleQueryResultHandler writer = new SPARQLResultsXMLWriter(System.out);
-            tupleQuery.evaluate(new TupleQueryResultHandler() {
+TupleQueryResultHandler writer = new SPARQLResultsXMLWriter(System.out);
+tupleQuery.evaluate(new TupleQueryResultHandler() {
 
-                int count = 0;
+    int count = 0;
 
-                @Override
-                public void startQueryResult(List<String> strings) throws TupleQueryResultHandlerException {
-                }
+    @Override
+    public void startQueryResult(List<String> strings) throws TupleQueryResultHandlerException {
+    }
 
-                @Override
-                public void endQueryResult() throws TupleQueryResultHandlerException {
-                }
+    @Override
+    public void endQueryResult() throws TupleQueryResultHandlerException {
+    }
 
-                @Override
-                public void handleSolution(BindingSet bindingSet) throws TupleQueryResultHandlerException {
-                    System.out.println(bindingSet);
-                }
-            });
+    @Override
+    public void handleSolution(BindingSet bindingSet) throws TupleQueryResultHandlerException {
+        System.out.println(bindingSet);
+    }
+});
 
-            conn.close();
-            myRepository.shutDown();
+conn.close();
+myRepository.shutDown();
+```

http://git-wip-us.apache.org/repos/asf/incubator-rya/blob/ce4a10ff/extras/rya.manual/pom.xml
----------------------------------------------------------------------
diff --git a/extras/rya.manual/pom.xml b/extras/rya.manual/pom.xml
index 2e27b9f..f79359a 100644
--- a/extras/rya.manual/pom.xml
+++ b/extras/rya.manual/pom.xml
@@ -11,284 +11,26 @@
 
     <artifactId>rya.manual</artifactId>
     <name>RYA Manual</name>
-    <packaging>war</packaging>
-
-    <properties>
-        <must-succeed>false</must-succeed>
-        <manual.dir>${project.build.directory}/${project.artifactId}</manual.dir>
-        <manual>${manual.dir}/${project.artifactId}-${project.version}</manual>
-        <netbeans.hint.deploy.server>Tomcat60</netbeans.hint.deploy.server>
-        <jetty-port>8080</jetty-port>
-        <jetty-war-dir>${project.build.directory}/webapp/</jetty-war-dir>
-        <scalate.version>1.7.1</scalate.version>
-        <scala.version>2.11.1</scala.version>
-        <wikitext.version>1.3</wikitext.version>
-        <scalate.editor>${env.SCALATE_EDITOR}</scalate.editor>
-        <scalate.mode>production</scalate.mode>
-        <scalate.workdir>${project.build.directory}/scalateWorkDir</scalate.workdir>
-        <maven.wagon.version>1.0-beta-6</maven.wagon.version>
-        <remoteTomcatManager>http://10.40.190.248:8080/manager</remoteTomcatManager>
-    </properties>
-
-    <dependencies>
-        <dependency>
-            <groupId>org.scalatra.scalate</groupId>
-            <artifactId>scalate-wikitext_2.11</artifactId>
-            <version>${scalate.version}</version>
-        </dependency>
-        <dependency>
-            <groupId>ch.qos.logback</groupId>
-            <artifactId>logback-classic</artifactId>
-            <version>0.9.24</version>
-        </dependency>
-        <dependency>
-            <groupId>org.slf4j</groupId>
-            <artifactId>slf4j-api</artifactId>
-            <version>1.6.0</version>
-        </dependency>
-        <dependency>
-            <groupId>org.scala-lang</groupId>
-            <artifactId>scala-library</artifactId>
-            <version>${scala.version}</version>
-        </dependency>
-        <dependency>
-            <groupId>org.scala-lang</groupId>
-            <artifactId>scala-compiler</artifactId>
-            <version>${scala.version}</version>
-        </dependency>
-    </dependencies>
+    <packaging>jar</packaging>
 
     <build>
         <plugins>
             <plugin>
                 <groupId>org.apache.maven.plugins</groupId>
-                <artifactId>maven-resources-plugin</artifactId>
-                <executions>
-                    <execution>
-                        <id>filter</id>
-                        <phase>generate-resources</phase>
-                        <goals>
-                            <goal>copy-resources</goal>
-                        </goals>
-                        <configuration>
-                            <outputDirectory>${project.build.directory}/webapp</outputDirectory>
-                            <resources>
-                                <resource>
-                                    <directory>src/main/webapp</directory>
-                                    <filtering>true</filtering>
-                                </resource>
-                            </resources>
-                        </configuration>
-                    </execution>
-                </executions>
-            </plugin>
-            <plugin>
-                <groupId>org.codehaus.mojo</groupId>
-                <artifactId>build-helper-maven-plugin</artifactId>
-                <version>1.7</version>
-                <executions>
-                    <execution>
-                        <id>add-source</id>
-                        <phase>generate-resources</phase>
-                        <goals>
-                            <goal>add-source</goal>
-                        </goals>
-                        <configuration>
-                            <sources>
-                                <source>${project.build.directory}/webapp</source>
-                            </sources>
-                        </configuration>
-                    </execution>
-                    <!--execution>
-                        <id>attach-artifacts</id>
-                        <phase>install</phase>
-                        <goals>
-                            <goal>attach-artifact</goal>
-                        </goals>
-                        <configuration>
-                            <artifacts>
-                                <artifact>
-                                    <file>${manual}.html</file>
-                                    <type>html</type>
-                                </artifact>
-                            </artifacts>
-                        </configuration>
-                    </execution-->
-                </executions>
-            </plugin>
-            <plugin>
-                <groupId>org.scalatra.scalate</groupId>
-                <artifactId>maven-scalate-plugin_2.11</artifactId>
-                <version>${scalate.version}</version>
-                <executions>
-                    <execution>
-                        <id>generate-htmls</id>
-                        <phase>compile</phase>
-                        <goals>
-                            <goal>sitegen</goal>
-                        </goals>
-                        <configuration>
-                            <webappDirectory>${project.build.directory}/webapp</webappDirectory>
-                        </configuration>
-                    </execution>
-                </executions>
-                <configuration>
-                    <target>${basedir}/src/main/webapp</target>
-                </configuration>
+                <artifactId>maven-site-plugin</artifactId>
                 <dependencies>
                     <dependency>
-                        <groupId>org.apache.maven.wagon</groupId>
-                        <artifactId>wagon-ssh</artifactId>
-                        <version>${maven.wagon.version}</version>
-                    </dependency>
-                    <dependency>
-                        <groupId>org.fusesource.wikitext</groupId>
-                        <artifactId>confluence-core</artifactId>
-                        <version>${wikitext.version}</version>
+                        <groupId>org.apache.maven.doxia</groupId>
+                        <artifactId>doxia-module-markdown</artifactId>
+                        <version>1.6</version>
                     </dependency>
                 </dependencies>
-            </plugin>
-            <plugin>
-                <groupId>org.apache.maven.plugins</groupId>
-                <artifactId>maven-war-plugin</artifactId>
                 <configuration>
-                    <webResources>
-                        <resource>
-                            <directory>${project.build.directory}/webapp/</directory>
-                        </resource>
-                    </webResources>
-                </configuration>
-            </plugin>
-            <plugin>
-                <groupId>org.codehaus.mojo</groupId>
-                <artifactId>tomcat-maven-plugin</artifactId>
-                <version>1.1</version>
-                <configuration>
-                    <url>${remoteTomcatManager}</url>
-                </configuration>
-                <!--executions>
-                    <execution>
-                        <id>redeploy-war</id>
-                        <phase>deploy</phase>
-                        <goals>
-                            <goal>redeploy</goal>
-                        </goals>
-                    </execution>
-                </executions-->
-            </plugin>
-            <plugin>
-                <groupId>org.mortbay.jetty</groupId>
-                <artifactId>jetty-maven-plugin</artifactId>
-                <version>7.1.5.v20100705</version>
-                <configuration>
-                    <!--webAppConfig>
-                        <contextPath>${project.artifactId}</contextPath>
-                    </webAppConfig-->
-                    <!-- When editing the conf files, you can comment this line to run
-                        mvn jetty:run
-                       to have a live web site
-                    -->
-                    <webAppSourceDirectory>${jetty-war-dir}</webAppSourceDirectory>
-
-                    <systemProperties>
-                        <systemProperty>
-                            <name>scalate.editor</name>
-                            <value>${scalate.editor}</value>
-                        </systemProperty>
-                        <systemProperty>
-                            <name>scalate.workdir</name>
-                            <value>${scalate.workdir}</value>
-                        </systemProperty>
-                        <systemProperty>
-                            <name>scalate.mode</name>
-                            <value>${scalate.mode}</value>
-                        </systemProperty>
-                    </systemProperties>
-                </configuration>
+                    <inputEncoding>UTF-8</inputEncoding>
+                    <outputEncoding>UTF-8</outputEncoding>
+                </configuration> 
             </plugin>
         </plugins>
     </build>
-
-    <reporting>
-        <plugins>
-            <plugin>
-                <groupId>org.scalatra.scalate</groupId>
-                <artifactId>maven-scalate-plugin</artifactId>
-                <version>${scalate.version}</version>
-            </plugin>
-        </plugins>
-    </reporting>
-
-    <profiles>
-        <profile>
-            <id>live</id>
-            <properties>
-                <jetty-war-dir>${basedir}/src/main/webapp/</jetty-war-dir>
-                <scalate.mode>development</scalate.mode>
-            </properties>
-        </profile>
-        <profile>
-            <id>pdf-manual</id>
-            <build>
-                <plugins>
-                    <plugin>
-                        <groupId>org.apache.maven.plugins</groupId>
-                        <artifactId>maven-antrun-plugin</artifactId>
-                        <executions>
-                            <execution>
-                                <id>attach-manual</id>
-                                <phase>install</phase>
-                                <goals>
-                                    <goal>run</goal>
-                                </goals>
-                                <configuration>
-                                    <tasks>
-                                        <mkdir dir="${manual.dir}" />
-                                        <move file="${project.build.directory}/sitegen/manual.html" tofile="${manual}.html" />
-                                        <echo message="Generating PDF using Prince XML (http://www.princexml.com/)" />
-                                        <exec executable="prince">
-                                            <arg value="${manual}.html" />
-                                            <arg value="${manual}.pdf" />
-                                            <arg value="--log" />
-                                            <arg value="${project.build.directory}/prince.log" />
-                                        </exec>
-                                    </tasks>
-                                </configuration>
-                            </execution>
-                        </executions>
-                    </plugin>
-                    <plugin>
-                        <groupId>org.codehaus.mojo</groupId>
-                        <artifactId>build-helper-maven-plugin</artifactId>
-                        <executions>
-                            <execution>
-                                <id>attach-artifacts</id>
-                                <phase>install</phase>
-                                <goals>
-                                    <goal>attach-artifact</goal>
-                                </goals>
-                                <configuration>
-                                    <artifacts>
-                                        <artifact>
-                                            <file>${manual}.pdf</file>
-                                            <type>pdf</type>
-                                        </artifact>
-                                    </artifacts>
-                                </configuration>
-                            </execution>
-                        </executions>
-                    </plugin>
-
-                </plugins>
-            </build>
-        </profile>
-        <profile>
-            <id>release</id>
-            <properties>
-                <must-succeed>true</must-succeed>
-            </properties>
-        </profile>
-    </profiles>
-
 </project>
 

http://git-wip-us.apache.org/repos/asf/incubator-rya/blob/ce4a10ff/extras/rya.manual/src/main/java/temp/TempJ.java
----------------------------------------------------------------------
diff --git a/extras/rya.manual/src/main/java/temp/TempJ.java b/extras/rya.manual/src/main/java/temp/TempJ.java
deleted file mode 100644
index be473e4..0000000
--- a/extras/rya.manual/src/main/java/temp/TempJ.java
+++ /dev/null
@@ -1,29 +0,0 @@
-package temp;
-
-/*
- * #%L
- * RYA Manual
- * %%
- * Copyright (C) 2014 Rya
- * %%
- * Licensed under the Apache License, Version 2.0 (the "License");
- * you may not use this file except in compliance with the License.
- * You may obtain a copy of the License at
- * 
- *      http://www.apache.org/licenses/LICENSE-2.0
- * 
- * Unless required by applicable law or agreed to in writing, software
- * distributed under the License is distributed on an "AS IS" BASIS,
- * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
- * See the License for the specific language governing permissions and
- * limitations under the License.
- * #L%
- */
-
-/**
-* Marker class so that target/classes directory gets generated
-* and the bundle plugin does not fail
-*/
-public class TempJ {
-	
-}
\ No newline at end of file

http://git-wip-us.apache.org/repos/asf/incubator-rya/blob/ce4a10ff/extras/rya.manual/src/main/webapp/.livereload
----------------------------------------------------------------------
diff --git a/extras/rya.manual/src/main/webapp/.livereload b/extras/rya.manual/src/main/webapp/.livereload
deleted file mode 100644
index b185045..0000000
--- a/extras/rya.manual/src/main/webapp/.livereload
+++ /dev/null
@@ -1,19 +0,0 @@
-# Lines starting with pound sign (#) are ignored.
-
-# additional extensions to monitor
-#config.exts << 'haml'
-
-# exclude files with NAMES matching this mask
-#config.exclusions << '~*'
-# exclude files with PATHS matching this mask (if the mask contains a slash)
-#config.exclusions << '/excluded_dir/*'
-# exclude files with PATHS matching this REGEXP
-#config.exclusions << /somedir.*(ab){2,4}.(css|js)$/
-
-# reload the whole page when .js changes
-#config.apply_js_live = false
-# reload the whole page when .css changes
-#config.apply_css_live = false
-
-# wait 100ms for more changes before reloading a page
-#config.grace_period = 0.1

http://git-wip-us.apache.org/repos/asf/incubator-rya/blob/ce4a10ff/extras/rya.manual/src/main/webapp/WEB-INF/scalate/layouts/default.scaml
----------------------------------------------------------------------
diff --git a/extras/rya.manual/src/main/webapp/WEB-INF/scalate/layouts/default.scaml b/extras/rya.manual/src/main/webapp/WEB-INF/scalate/layouts/default.scaml
deleted file mode 100644
index 8b27470..0000000
--- a/extras/rya.manual/src/main/webapp/WEB-INF/scalate/layouts/default.scaml
+++ /dev/null
@@ -1,97 +0,0 @@
-
--@ var body: String
--@ var title : String = "Rya Manual"
-- response.setContentType("text/html")
-
--# Only include the console if it's available and the engine is in dev mode.
-- val include_console = engine.isDevelopmentMode && engine.resourceLoader.exists("/org/fusesource/scalate/console/console_head.scaml")
-
-!!! Basic
-%html(lang="en")
-  %head
-    %meta(http-equiv="Content-Type" content="text/html; charset=utf-8")
-    %meta(name="description" content="description goes here")
-    %meta(name="keywords" content="keywords,goes,here")
-    %meta(name="author" content="The Team")
-
-    - if (include_console)
-      = include("/org/fusesource/scalate/console/console_head.scaml")
-
-    %link(href={uri("/css/style.css")} rel="stylesheet" type="text/css")
-    %link(href={uri("/css/pygmentize.css")} rel="stylesheet" type="text/css")
-
-    - if (include_console)
-      %link(href={uri("/css/scalate/console.css")} rel="stylesheet" type="text/css")
-
-    %title
-      = title
-  
-  %body
-    %table{:width => "100%", :cellpadding => "0", :cellspacing => "0"}
-      %tr{:width => "100%"}
-        %td#cell-0-0{:colspan => "2"}
-          &nbsp;
-        %td#cell-0-1
-          &nbsp;
-        %td#cell-0-2{:colspan => "2"}
-          &nbsp;
-      %tr{:width => "100%"}
-        %td#cell-1-0
-          &nbsp;
-        %td#cell-1-1
-          &nbsp;
-        %td#cell-1-2
-          %div{:style => "padding: 5px;"}
-            #banner
-              = include("/_banner.ssp")
-            #top-menu
-              %table{:border => "0", :cellpadding => "1", :cellspacing => "0", :width => "100%"}
-                %tr
-                  %td
-                    %div{:align => "left"}
-                  %td
-                    %div{:align => "right"}
-                      = include("/_quicklinks.ssp")
-        %td#cell-1-3
-          &nbsp;
-        %td#cell-1-4
-          &nbsp;
-      %tr{:width => "100%"}
-        %td#cell-2-0{:colspan => "2"}
-          &nbsp;
-        %td#cell-2-1
-          %table
-            %tr{:height => "100%", :valign => "top"}
-              %td{:height => "100%"}
-                #wrapper-menu-page-right
-                  #wrapper-menu-page-top
-                    #wrapper-menu-page-bottom
-                      #menu-page
-                        = include("_navigation.conf")
-              %td{:height =>"100%", :width => "100%"}
-                .wiki-content
-                  !~~ body
-        %td#cell-2-2{:colspan => "2"}
-          &nbsp;
-      %tr{:width => "100%"}
-        %td#cell-3-0
-          &nbsp;
-        %td#cell-3-1
-          &nbsp;
-        %td#cell-3-2
-          #footer
-            #site-footer
-              %br
-        %td#cell-3-3
-          &nbsp;
-        %td#cell-3-4
-          &nbsp;
-      %tr{:width => "100%"}
-        %td#cell-4-0{:colspan => "2"}
-          &nbsp;
-        %td#cell-4-1
-          &nbsp;
-        %td#cell-4-2{:colspan => "2"}
-          &nbsp;
-    - if (include_console)
-      = include("/org/fusesource/scalate/console/console.scaml")

http://git-wip-us.apache.org/repos/asf/incubator-rya/blob/ce4a10ff/extras/rya.manual/src/main/webapp/WEB-INF/scalate/layouts/print.ssp
----------------------------------------------------------------------
diff --git a/extras/rya.manual/src/main/webapp/WEB-INF/scalate/layouts/print.ssp b/extras/rya.manual/src/main/webapp/WEB-INF/scalate/layouts/print.ssp
deleted file mode 100644
index 1676547..0000000
--- a/extras/rya.manual/src/main/webapp/WEB-INF/scalate/layouts/print.ssp
+++ /dev/null
@@ -1,778 +0,0 @@
-
-<%@ var body: String %>
-<%@ var title: String = "Rya Manual" %>
-<% escapeMarkup = false %>
-<!DOCTYPE html>
-<html>
-<head>
-  <style type="text/css">
-
-html {
-  margin: 0; padding: 0;
-  font: 10pt/1.26 "Gill Sans", sans-serif;
-}
-
-h1, h2, h3, h4, h5, h6 {
-  font-family: "Gill Sans", sans-serif;
-  margin: 2em 0 0.5em 0;
-  page-break-after: avoid;
-}
-
-h1 {
-  font-size: 2.0em;
-  font-weight: 900;
-
-  margin: 0;
-  margin-left:-2cm;
-  margin-top:-1cm;
-  margin-bottom:1.5cm;
-  top: 0cm;
-  left: 0cm;
-  padding: 2cm;
-  padding-top: 0cm;
-  padding-bottom: 1cm;
-  background: #888;
-  border-bottom-right-radius: 2cm;
-  page-break-before: always;
-  page-break-inside: avoid;
-}
-
-@media screen, handheld {
-h1 {
-  margin-top:1cm;
-  background-image: url("chapter-rounded-bottom.png");
-  background-repeat: no-repeat;
-  background-position: bottom right;
-}
-div.page-reset > h1 {
-  margin-top:0cm;
-}
-}
-
-
-h2 {
-  font-size: 1.2em;
-  text-transform: uppercase;
-  font-weight: bold;
-}
-
-h3 {
-  font-size: 1em;
-  font-weight: bold;
-}
-
-q::before {
-  content: "\201C";
-}
-
-q::after {
-  content: "\201D";
-}
-
-p { margin: 0 }
-p + p { text-indent: 1.3em ; margin-top: 0.2em; }
-p.sidenote + p, p.caption, p.art { text-indent: 0 }
-
-p.author {
-  margin-top: 2em;
-  text-indent: 0;
-  text-align: right;
-}
-
-a { text-decoration: none; color: black }
-
-/* cross-references */
-
-a.pageref::after { content: " on page " target-counter(attr(href), page); }
-a.chapref::before { content: " Chapter " target-counter(attr(href), chapter) ", "; }
-a.figref { content: " Figure " target-counter(attr(href), figure); }
-a.tableref { content: " Table " target-counter(attr(href), figure); }
-a.listingref { content: " Listing " target-counter(attr(href), listing); }
-
-/* sidenotes */
-
-.sidenote {
-  float: left;
-  clear: left;
-  margin: 0 0 1em -41%;
-  width: 37%;
-  font-size: 0.9em;
-  font-style: normal;
-  text-indent: 0;
-  text-align: right;
-  page-break-inside: avoid;
-}
-
-/* sidebars */
-
-div.sidebar {
-  float: top-next;
-  margin: 1.2em 0 1.2em 0;
-  border: thin solid;
-  background: #CCC;
-  padding: 0.5em 1em;
-  page-break-inside: avoid;
-  column-count: 2;
-  column-gap: 1.5em;
-}
-
-div.sidebar h2 {
-  margin-top: 0;
-}
-
-/* frontpage */
-
-.title p{
-	font-size:22pt;
-  	font-family: "Gill Sans", sans-serif;
-  	text-align: center;
-}
-
-.copyright-section {
-	text-align: center;
-	font-size: 9pt;
-	page-break-after: always;
-	margin-top: 50pt;
-	margin-bottom: 20pt;
-}
-
-.toc-title {
-	font-size:18pt;
-  	font-family: "Gill Sans", sans-serif;
-  	text-align: left;
-  	margin-left:20pt;
-  	margin-bottom: 40pt;
-
-}
-
-/* figures, tables, and listings */
-
-div.confluenceTableSmall th.confluenceTh {
-  font-size: 11px;
-}
-
-div.confluenceTableSmall td.confluenceTd {
-  font-size: 7px;
-}
-
-div.figure {
-  margin: 1em 0;
-  counter-increment: figure;
-}
-
-div.figure .caption, div.table .caption {
-  font-size: 0.9em;
-  font-style: italic;
-}
-
-div.figure .caption::before {
-  content: "Figure " counter(figure) ". ";
-  font-weight: bold;
-  font-style: normal;
-}
-
-div.table .caption::before {
-  content: "Table " counter(table) ". ";
-  font-weight: bold;
-  font-style: normal;
-}
-
-div.table {
-  margin: 1em 0;
-  counter-increment: table;
-}
-
-div.table th {
-  text-align: left;
-}
-
-table th, table td {
-  text-align: left;
-  padding-right: 1em;
-  page-break-inside: avoid;
-}
-
-table th {
-  border-top: thin solid;
-  border-bottom: thin solid;
-  padding-top: 0.2em;
-  padding-bottom: 0.2em;
-}
-table td {
-  border-top: none;
-  border-bottom: thin dotted;
-  padding-top: 0.2em;
-  padding-bottom: 0.2em;
-}
-
-div.Scrollbar {
-	display: none;
-}
-
-
-/* Weird div.codeHeader a b::before would be a better selection
-   but prince does not properly match it.. Firefox does. */
-div.codeHeader::before {
-  content: "Listing " counter(listing) ". ";
-  font-weight: bold;
-  font-style: normal;
-}
-div.codeHeader a b {
-  font-style: italic;
-  font-weight: normal;
-}
-div.codeHeader {
-  font-size: 0.9em;
-  counter-increment: listing;
-}
-div.code {
-	border: 1px dashed #c0c0c0;
-    font-size: 12px;
-	font-family: Courier;
-    margin: 5px;
-	line-height: 13px;
-	padding: 3px;
-	background-color: #f8f8f8;
-
-}
-
-
-@page {
-  margin: 2cm;
-  size: 7in 9.25in;
-
-  @footnotes {
-    border-top: thin solid black;
-    padding-top: 0.3em;
-    margin-top: 0.6em;
-  }
-}
-
-
-/* define default page and names pages: cover, blank, frontmatter */
-div.page-reset {
-    counter-reset: page 1;
-}
-
-@page :left {
-  @top-left-corner {
-    font-weight: 900; font: 9pt "Gill Sans", serif;
-    content: counter(page);
-    text-align: left;
-    margin-left: 1cm;
-    visibility: hidden;
-  }
-  @top-left {
-    font-weight: 900;
-    font: 9pt "Gill Sans", serif; white-space: pre; text-transform: uppercase; letter-spacing: 0.1em;
-    content: string(header, first);
-    visibility: hidden;
-  }
-  @bottom-left-corner {
-    font-weight: 900; font: 9pt "Gill Sans", serif;
-    content: counter(page);
-    text-align: left;
-    margin-left: 1cm;
-  }
-  @bottom-left {
-    font-weight: 900;
-    font: 9pt "Gill Sans", serif; white-space: pre; text-transform: uppercase; letter-spacing: 0.1em;
-    content: string(header, first);
-  }
-}
-
-@page :right {
-  @top-right-corner {
-    font-weight: 900; font: 9pt "Gill Sans", serif;
-    content: counter(page);
-    text-align: left;
-    margin-left: 1cm;
-    visibility: hidden;
-  }
-  @top-right {
-    font-weight: 900;
-    font: 9pt "Gill Sans", serif; white-space: pre; text-transform: uppercase; letter-spacing: 0.1em;
-    content: string(header, first)
-    visibility: hidden;
-  }
-  @bottom-right-corner {
-    font-weight: 900; font: 9pt "Gill Sans", serif;
-    content: counter(page);
-    text-align: right;
-    margin-right: 1cm;
-  }
-  @bottom-right {
-    font-weight: 900; font: 9pt "Gill Sans", serif;
-    white-space: pre; text-transform: uppercase; letter-spacing: 0.1em;
-    content: string(header, first)
-  }
-}
-
-/*
-  In theory we should be able to use the :first selector so taht
-  we can put the page numbering on the bottom of the first page of the chapter
-  but have the rest of the pages number at the top.  But this does not seem
-  to work.  See http://www.princexml.com/doc/6.0/page-selectors/
-
-  So for now just always number at the bottom :(
-*/
-/*
-div.chapter { page: bottom-number; }
-@page bottom-number :first {
-  @top-left {
-    visibility: hidden;
-  }
-  @bottom-left {
-    visibility: visible;
-  }
-  @top-right {
-    visibility: hidden;
-  }
-  @bottom-right {
-    visibility: visible;
-  }
-}
-*/
-
-@page cover { margin: 0; }
-
-@page frontmatter :left {
-  @bottom-left-corner {
-    content: counter(page, lower-roman);
-  }
-  @bottom-left-corner {
-    content: counter(page, lower-roman);
-  }
-}
-
-@page frontmatter :right {
-  @bottom-right-corner {
-    content: counter(page, lower-roman);
-  }
-  @bottom-right-corner {
-    content: counter(page, lower-roman);
-  }
-}
-
-@page blank :left {
-  @top-left { visibility: hidden; }
-  @bottom-left { visibility: hidden; }
-  @top-left-corner { visibility: hidden; }
-  @bottom-left-corner { visibility: hidden; }
-}
-
-@page blank :right {
-  @top-right { visibility: hidden; }
-  @bottom-right { visibility: hidden; }
-  @top-right-corner { visibility: hidden; }
-  @bottom-right-corner { visibility: hidden; }
-}
-
-/* footnotes */
-.footnote {
-  display: none;                   /* default rule */
-
-  display: prince-footnote;        /* prince-specific rules */
-  position: footnote;
-  footnote-style-position: inside;
-
-  counter-increment: footnote;
-  margin-left: 1.4em;
-  font-size: 90%;
-  line-height: 1.4;
-}
-
-.footnote::footnote-call {
-  vertical-align: super;
-  font-size: 80%;
-}
-
-.footnote::footnote-marker {
-  vertical-align: super;
-  color: green;
-  padding-right: 0.4em;
-}
-
-/* Confluence contents to hide */
-#labels-section {
-	display: none;
-}
-#comments-section {
-	display: none;
-}
-#footer {
-	display: none;
-}
-.hidden {
-	display: none;
-}
-
-/*
-   A book consists of different types of sections. We propose to use
-   DIV elements with these class names:
-
-    frontcover
-    halftitlepage: contains the title of the book
-    titlepage: contains the title of the book, name of author(s) and publisher
-    imprint: left page with copyright, publisher, library printing information
-    dedication: right page with short dedication
-    foreword: written by someone other than the author(s)
-    toc: table of contents
-    preface: preface, including acknowledgements
-    chapter: each chapter is given its own DIV element
-    references: contains list of references
-    appendix: each appendix is given its own
-    bibliography
-    glossary
-    index
-    colophon: describes how the book was produced
-    backcover
-
-   A book will use several of the types listed above, but few books
-   will use all of them.
-*/
-
-/* which section uses which named page */
-
-div.halftitlepage, div.titlepage, div.imprint, div.dedication { page: blank }
-div.foreword, div.toc, div.preface { page: frontmatter }
-
-
-/* page breaks */
-div.frontcover, div.halftitlepage, div.titlepage { page-break-before: right }
-div.imprint { page-break-before: always; }
-div.chapter { page-break-before: always; }
-div.dedication, div.foreword, div.toc, div.preface, div.reference,
-div.appendix, div.bibliography, div.glossary, div.index, div.colophon {
-  page-break-before: always
-}
-div.backcover { page-break-before: left }
-
-/* titlepage, halftitlepage */
-
-div.titlepage h1, div.halftitlepage h1 { margin-bottom: 2em; }
-div.titlepage h2, div.halftitlepage h2 { font-size: 1.2em; margin-bottom: 3em; }
-div.titlepage h3, div.halftitlepage h3 { font-size: 1em; margin-bottom: 3em; }
-div.titlepage p, div.halftitlepage p {
-  font-size: 1.4em;
-  font-weight: bold;
-  margin: 0; padding: 0;
-}
-
-
-/* TOC */
-
-ul.toc, ul.toc ul {
-  list-style-type: none;
-  margin: 0; padding: 0;
-  margin-left: 3cm;
-}
-ul.toc ul {
-  margin-left: 1em;
-  font-weight: normal;
-}
-ul.toc > li {
-  font-weight: bold;
-  margin-bottom: 0.5em;
-}
-ul.toc a::after {
-  content: leader('.') target-counter(attr(href), page);
-  font-style: normal;
-}
-ul.toc > li.frontmatter a::after {
-  content: leader('.') target-counter(attr(href), page, lower-roman);
-  font-style: normal;
-}
-ul.toc > li.endmatter a::after {
-  content: leader('.') target-counter(attr(href), page);
-  font-style: normal;
-}
-ul.toc > li.chapter::before {
-  content: "Chapter " counter(toc-chapter, decimal);
-  display: block;
-  margin: 1em 0 0.1em -2.5cm;
-  font-weight: normal;
-  counter-increment: toc-chapter;
-  page-break-after: avoid;
-}
-
-/* chapter numbers */
-
-div.chapter { counter-increment: chapter; }
-div.chapter h1::before {
-  text-transform: uppercase;
-  letter-spacing: 0.15em;
-  content: "Chapter  " counter(chapter) " \A\B0 \B0 \B0 \B0\A";
-  white-space: pre;
-  font-size: 50%;
-}
-
-div.frontcover h1::before, div.titlepage h1::before, div.halftitlepage h1::before {
-  content: normal; /* that is, none */
-}
-
-h1 { string-set: header content();}
-div.chapter h1 { string-set: header "Chapter " counter(chapter) " - " content(); }
-
-/* index */
-
-ul.index {
-  list-style-type: none;
-  margin: 0; padding: 0;
-  column-count: 2;
-  column-gap: 1em;
-}
-
-ul.index a::after { content: ", " target-counter(attr(href), page); }
-
-
-span.element, span.attribute {
-  text-transform: uppercase;
-  font-weight: bold;
-  font-size: 80%;
-}
-span.property { font-weight: bold }
-code, span.css, span.value, span.declaration {
-  font: 90% "Lucida Console", "Lucida Sans Typewriter", monospace;
-}
-
-
-@media screen, handheld {
-  html {font: 14px "Gill Sans", sans-serif; }
-  h1 { margin-bottom: 0.5em }
-  div.frontcover, div.halftitlepage, div.titlepage, div.imprint,
-  div.dedication, div.foreword, div.toc, div.index { display: none }
-  body {
-      margin: 0cm;
-      margin-left: 2cm;
-      margin-right: 2cm;
-  }
-}
-
-/*
- * Enhancements to the take advantage of some of the style markup that
- * Confluence generates
- */
-a sup img { visibility: hidden; position: absolute;}
-
-img {
-  prince-image-resolution:150dpi;
-}
-
-table {
-  font: "Lucida Console", "Lucida Sans Typewriter", monospace;
-}
-
-table td {
-  font-size: 10pt;
-}
-
-pre {
-   white-space: pre-wrap;
-}
-
-.codeContent {
-  font-size: 80%;
-}
-.code {
-}
-.code-keyword {
-  color: #000091;
-  background-color: inherit;
-}
-
-.code-object {
-  color: #910091;
-  background-color: inherit;
-}
-
-.code-quote {
-  color: #009100;
-  background-color: inherit;
-}
-
-.code-comment {
-  color: #808080;
-  background-color: inherit;
-}
-
-
-.code-xml .code-keyword {
-  color: inherit;
-  font-weight: bold;
-}
-
-.code-tag {
-  color: #000091;
-  background-color: inherit;
-}
-
-.noteMacro { border-color: #F0C000; background-color: #FFFFCE;}
-.warningMacro { border-color: #CC0000; background-color: #FFCCCC }
-.infoMacro { border-color: #3c78b5; background-color: #D8E4F1; }
-.tipMacro { border-color: #090; background-color: #dfd;}
-.noteMacro, .warningMacro, .infoMacro, .tipMacro, .informationMacroPadding {
-  border: thin solid;
-  float: top-next;
-  margin: 1em 0 1.2em 0;
-  padding: 0.5em;
-  column-count: 2;
-  column-gap: 1.5em;
-  width: 100%;
-}
-table.infoMacro td, table.warningMacro td, table.tipMacro td, table.noteMacro td, table.sectionMacro td {
-    border: none;
-}
-table.infoMacro p, table.warningMacro p, table.tipMacro p, table.noteMacro p, table.sectionMacro p {
-    font-size:x-small;
-    margin-top: 1em;
-}
-  </style>
-  <style type="text/css">
-.syntax .hll { background-color: #ffffcc }
-.syntax  { background: #f0f0f0; }
-.syntax .c { color: #60a0b0; font-style: italic } /* Comment */
-.syntax .err { border: 1px solid #FF0000 } /* Error */
-.syntax .k { color: #007020; font-weight: bold } /* Keyword */
-.syntax .o { color: #666666 } /* Operator */
-.syntax .cm { color: #60a0b0; font-style: italic } /* Comment.Multiline */
-.syntax .cp { color: #007020 } /* Comment.Preproc */
-.syntax .c1 { color: #60a0b0; font-style: italic } /* Comment.Single */
-.syntax .cs { color: #60a0b0; background-color: #fff0f0 } /* Comment.Special */
-.syntax .gd { color: #A00000 } /* Generic.Deleted */
-.syntax .ge { font-style: italic } /* Generic.Emph */
-.syntax .gr { color: #FF0000 } /* Generic.Error */
-.syntax .gh { color: #000080; font-weight: bold } /* Generic.Heading */
-.syntax .gi { color: #00A000 } /* Generic.Inserted */
-.syntax .go { color: #808080 } /* Generic.Output */
-.syntax .gp { color: #c65d09; font-weight: bold } /* Generic.Prompt */
-.syntax .gs { font-weight: bold } /* Generic.Strong */
-.syntax .gu { color: #800080; font-weight: bold } /* Generic.Subheading */
-.syntax .gt { color: #0040D0 } /* Generic.Traceback */
-.syntax .kc { color: #007020; font-weight: bold } /* Keyword.Constant */
-.syntax .kd { color: #007020; font-weight: bold } /* Keyword.Declaration */
-.syntax .kn { color: #007020; font-weight: bold } /* Keyword.Namespace */
-.syntax .kp { color: #007020 } /* Keyword.Pseudo */
-.syntax .kr { color: #007020; font-weight: bold } /* Keyword.Reserved */
-.syntax .kt { color: #902000 } /* Keyword.Type */
-.syntax .m { color: #40a070 } /* Literal.Number */
-.syntax .s { color: #4070a0 } /* Literal.String */
-.syntax .na { color: #4070a0 } /* Name.Attribute */
-.syntax .nb { color: #007020 } /* Name.Builtin */
-.syntax .nc { color: #0e84b5; font-weight: bold } /* Name.Class */
-.syntax .no { color: #60add5 } /* Name.Constant */
-.syntax .nd { color: #555555; font-weight: bold } /* Name.Decorator */
-.syntax .ni { color: #d55537; font-weight: bold } /* Name.Entity */
-.syntax .ne { color: #007020 } /* Name.Exception */
-.syntax .nf { color: #06287e } /* Name.Function */
-.syntax .nl { color: #002070; font-weight: bold } /* Name.Label */
-.syntax .nn { color: #0e84b5; font-weight: bold } /* Name.Namespace */
-.syntax .nt { color: #062873; font-weight: bold } /* Name.Tag */
-.syntax .nv { color: #bb60d5 } /* Name.Variable */
-.syntax .ow { color: #007020; font-weight: bold } /* Operator.Word */
-.syntax .w { color: #bbbbbb } /* Text.Whitespace */
-.syntax .mf { color: #40a070 } /* Literal.Number.Float */
-.syntax .mh { color: #40a070 } /* Literal.Number.Hex */
-.syntax .mi { color: #40a070 } /* Literal.Number.Integer */
-.syntax .mo { color: #40a070 } /* Literal.Number.Oct */
-.syntax .sb { color: #4070a0 } /* Literal.String.Backtick */
-.syntax .sc { color: #4070a0 } /* Literal.String.Char */
-.syntax .sd { color: #4070a0; font-style: italic } /* Literal.String.Doc */
-.syntax .s2 { color: #4070a0 } /* Literal.String.Double */
-.syntax .se { color: #4070a0; font-weight: bold } /* Literal.String.Escape */
-.syntax .sh { color: #4070a0 } /* Literal.String.Heredoc */
-.syntax .si { color: #70a0d0; font-style: italic } /* Literal.String.Interpol */
-.syntax .sx { color: #c65d09 } /* Literal.String.Other */
-.syntax .sr { color: #235388 } /* Literal.String.Regex */
-.syntax .s1 { color: #4070a0 } /* Literal.String.Single */
-.syntax .ss { color: #517918 } /* Literal.String.Symbol */
-.syntax .bp { color: #007020 } /* Name.Builtin.Pseudo */
-.syntax .vc { color: #bb60d5 } /* Name.Variable.Class */
-.syntax .vg { color: #bb60d5 } /* Name.Variable.Global */
-.syntax .vi { color: #bb60d5 } /* Name.Variable.Instance */
-.syntax .il { color: #40a070 } /* Literal.Number.Integer.Long */
-
-
-/* don't highlight errors */
-.syntax .err {
-  border: none;
-}
-
-.syntax {
-  font-size: .9em;
-  font-family:Monaco,"Courier New","DejaVu Sans Mono","Bitstream Vera Sans Mono",monospace;
-  background-color: #F8F8FF;
-
-  overflow:auto;
-  -moz-background-clip:border;
-  -moz-background-inline-policy:continuous;
-  -moz-background-origin:padding;
-  margin: 1em 0 1em 0;
-  border:1px solid #DDDDDD;
-
-  border-top-left-radius: 8px; -webkit-border-top-left-radius: 8px; -moz-border-radius-topleft: 8px;
-  border-top-right-radius: 8px; -webkit-border-top-right-radius: 8px; -moz-border-radius-topright: 8px;
-  border-style: solid;  border-width: 1px; border-color: #dedede !important;
-  padding: 1em;
-}
-.syntax .linenodiv  {
-  background-color:#ECECEC;
-  border-right:1px solid #DDDDDD;
-  color:#AAAAAA;
-  padding: .5em;
-  text-align:right;
-}
-.syntax .highlight  {
-}
-.syntax pre {
-  margin:0;
-}
-
-pre.syntax {
-  padding: .5em;
-  background-color: #F8F8FF; overflow:auto;
-}
-
-.syntax code {
-  font-family:Monaco,"Courier New","DejaVu Sans Mono","Bitstream Vera Sans Mono",monospace;
-  font-size: 10pt;
-}
-
-div.compare { width: 700px; }
-div.compare div.compare-left { float:left; width:340px; padding:5px; margin-top: 15px; }
-div.compare div.compare-right { float:right; width:340px; padding:5px; margin-top: 15px; }
-div.compare div h3 {
-  margin-left: 15px;
-  padding: 5px 15px;
-  display: inline;
-  font-size: .8em;
-  color: #666;
-
-  border-top: 1px solid #ccc; -moz-border-top-colors: #ccc white white #e5e5e5;
-  border-left: 1px solid #ccc; -moz-border-left-colors: #ccc white white #e5e5e5;
-  border-right: 1px solid #ccc;-moz-border-right-colors: #ccc white white #e5e5e5;
-  border-top-left-radius: 8px; -webkit-border-top-left-radius: 8px;  -moz-border-radius-topleft: 8px;
-  border-top-right-radius: 8px; -webkit-border-top-right-radius: 8px; -moz-border-radius-topright: 8px;
-}
-div.compare div div {
-  margin: 5px 0px 0px 0px;
-}
-.clear {
-  clear:both;
-}
-.wide div.compare div.compare-left { float:none; width:700px; }
-.wide div.compare div.compare-right { float:none; width:700px; }
-
-  </style>
-
-  <title>${title}</title>
-</head>
-<body>
-  <div id="titlepage">
-    <div id="title">Distillery Mgmt</div>
-  </div>
-  <div id="main">
-    <%= body %>
-  </div>
-</body>
-</html>

http://git-wip-us.apache.org/repos/asf/incubator-rya/blob/ce4a10ff/extras/rya.manual/src/main/webapp/WEB-INF/web.xml
----------------------------------------------------------------------
diff --git a/extras/rya.manual/src/main/webapp/WEB-INF/web.xml b/extras/rya.manual/src/main/webapp/WEB-INF/web.xml
deleted file mode 100644
index 71398d6..0000000
--- a/extras/rya.manual/src/main/webapp/WEB-INF/web.xml
+++ /dev/null
@@ -1,50 +0,0 @@
-<?xml version="1.0" encoding="UTF-8"?>
-<!--
-
-    Copyright (C) 2009-2010 the original author or authors.
-    See the notice.md file distributed with this work for additional
-    information regarding copyright ownership.
-
-    Licensed under the Apache License, Version 2.0 (the "License");
-    you may not use this file except in compliance with the License.
-    You may obtain a copy of the License at
-
-        http://www.apache.org/licenses/LICENSE-2.0
-
-    Unless required by applicable law or agreed to in writing, software
-    distributed under the License is distributed on an "AS IS" BASIS,
-    WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
-    See the License for the specific language governing permissions and
-    limitations under the License.
-
--->
-
-<web-app xmlns:xsi="http://www.w3.org/2001/XMLSchema-instance"
-	xmlns="http://java.sun.com/xml/ns/javaee" xmlns:web="http://java.sun.com/xml/ns/javaee/web-app_2_5.xsd"
-	xsi:schemaLocation="http://java.sun.com/xml/ns/javaee http://java.sun.com/xml/ns/javaee/web-app_2_5.xsd"
-	version="2.5">
-
-  <display-name>MVM ALX Documentation</display-name>
-
-  <description>
-    MVM ALX Documentation
-  </description>
-
-  <filter>
-    <filter-name>TemplateEngineFilter</filter-name>
-    <filter-class>org.fusesource.scalate.servlet.TemplateEngineFilter</filter-class>
-  </filter>
-  <filter-mapping>
-    <filter-name>TemplateEngineFilter</filter-name>
-    <url-pattern>/*</url-pattern>
-  </filter-mapping>
-
-  <welcome-file-list>
-    <welcome-file>index.conf</welcome-file>
-  </welcome-file-list>
-
-  <error-page>
-    <error-code>500</error-code>
-    <location>/WEB-INF/scalate/errors/500.scaml</location>
-  </error-page>
-</web-app>

http://git-wip-us.apache.org/repos/asf/incubator-rya/blob/ce4a10ff/extras/rya.manual/src/main/webapp/_banner.ssp
----------------------------------------------------------------------
diff --git a/extras/rya.manual/src/main/webapp/_banner.ssp b/extras/rya.manual/src/main/webapp/_banner.ssp
deleted file mode 100644
index 7322e88..0000000
--- a/extras/rya.manual/src/main/webapp/_banner.ssp
+++ /dev/null
@@ -1,8 +0,0 @@
-      <table border="0" cellpadding="0" cellspacing="0" width="100%">
-        <tr>
-          <td align="left" class="topbardiv" nowrap="">
-        <img border="0" src="${uri("/images/logo2.png")}" width="400px" height="100px"></img>
-</a>
-          </td>
-        </tr>
-      </table>

http://git-wip-us.apache.org/repos/asf/incubator-rya/blob/ce4a10ff/extras/rya.manual/src/main/webapp/_navigation.conf
----------------------------------------------------------------------
diff --git a/extras/rya.manual/src/main/webapp/_navigation.conf b/extras/rya.manual/src/main/webapp/_navigation.conf
deleted file mode 100644
index 6ab6ce8..0000000
--- a/extras/rya.manual/src/main/webapp/_navigation.conf
+++ /dev/null
@@ -1,24 +0,0 @@
-h3. [Home|index]
-
-h3. Rya
-- [Overview|./overview.html]
-- [Quick Start|./quickstart.html]
-- [Load Data|./loaddata.html]
-- [Query Data|./querydata.html]
-- [Evaluation Table|./eval.html]
-- [Pre-computed Joins|./loadPrecomputedJoin.html]
-- [Inferencing|./infer.html]
-
-h3. Samples
-- [Typical First Steps|./sm-firststeps.html]
-- [Simple Add/Query/Remove Statements|./sm-simpleaqr.html]
-- [Sparql query|./sm-sparqlquery.html]
-- [Adding Authentication|./sm-addauth.html]
-- [Inferencing|./sm-infer.html]
-- [Named Graph|./sm-namedgraph.html]
-- [Update data|./sm-updatedata.html]
-- [Alx|./alx.html]
-
-h3. Development
-- [Building From Source|./build-source.html]
-- [LTS Maven Settings XML|./maven-settings.html]
\ No newline at end of file

http://git-wip-us.apache.org/repos/asf/incubator-rya/blob/ce4a10ff/extras/rya.manual/src/main/webapp/_quicklinks.ssp
----------------------------------------------------------------------
diff --git a/extras/rya.manual/src/main/webapp/_quicklinks.ssp b/extras/rya.manual/src/main/webapp/_quicklinks.ssp
deleted file mode 100644
index c9b8966..0000000
--- a/extras/rya.manual/src/main/webapp/_quicklinks.ssp
+++ /dev/null
@@ -1,5 +0,0 @@
-<p>
-<a href="${uri("/index.html")}" title="Home">Home</a>
-&#124;
-</p>
-

http://git-wip-us.apache.org/repos/asf/incubator-rya/blob/ce4a10ff/extras/rya.manual/src/main/webapp/alx.conf
----------------------------------------------------------------------
diff --git a/extras/rya.manual/src/main/webapp/alx.conf b/extras/rya.manual/src/main/webapp/alx.conf
deleted file mode 100644
index e9d7875..0000000
--- a/extras/rya.manual/src/main/webapp/alx.conf
+++ /dev/null
@@ -1,59 +0,0 @@
-h1. Alx Rya Integration
-
-Alx is a modular framework for developing applications. Rya has mechanisms to integrate directly into Alx to provide other modules access to queries.
-
-Currently, the Alx Rya extension only allows interacting with an Accumulo store.
-
-h2. Prerequisites
-
-- Alx 1.0.5+ (we will refer to it at the ALX_HOME directory from now on)
-- alx.rya features xml (can be found in maven at mvn:mvm.rya/alx.rya/<version>/xml/features)
-
-h2. Steps
-
-1. Start up Alx
-2. features:addurl alx.rya features xml
-3. features:install alx.rya
-4. (optional) features:install alx.rya.console
-
-That's it. To make sure, run {{ls <alx.rya bundle id>}} and make sure something like this pops up:
-{code}
-mvm.rya.alx.rya (99) provides:
-------------------------------
-Bundle-SymbolicName = mvm.rya.alx.rya
-Bundle-Version = 3.0.4.SNAPSHOT
-objectClass = org.osgi.service.cm.ManagedService
-service.id = 226
-service.pid = mvm.rya.alx
-----
-...
-{code}
-
-h2. Using
-
-The bundle registers a Sail Repository, so you can interact with it directly as in the other code examples. Here is a quick groovy example of the usage:
-{code}
-import org.springframework.osgi.extensions.annotation.*;
-import org.openrdf.repository.*;
-import org.openrdf.model.ValueFactory;
-import static mvm.rya.api.RdfCloudTripleStoreConstants.*;
-
-class TstRepo {
-
-	@ServiceReference
-	public void setRepo(Repository repo) {
-		println repo
-		RepositoryConnection conn = repo.getConnection();
-		ValueFactory vf = VALUE_FACTORY;
-        def statements = conn.getStatements(vf.createURI("http://www.Department0.University0.edu"), null, null, true);
-        while(statements.hasNext()) {
-            System.out.println(statements.next());
-        }
-        statements.close();
-        conn.close();
-	}
-
-}
-{code}
-
-The bundle also registers a RyaDAO, so you can interact with the RyaDAO interface directly
\ No newline at end of file

http://git-wip-us.apache.org/repos/asf/incubator-rya/blob/ce4a10ff/extras/rya.manual/src/main/webapp/build-source.conf
----------------------------------------------------------------------
diff --git a/extras/rya.manual/src/main/webapp/build-source.conf b/extras/rya.manual/src/main/webapp/build-source.conf
deleted file mode 100644
index 49f9a4a..0000000
--- a/extras/rya.manual/src/main/webapp/build-source.conf
+++ /dev/null
@@ -1,19 +0,0 @@
-h1. Building from Source
-
-h2. Prerequisites
-
-* Rya code
-* Maven 2.2 +
-
-h2. Building
-
-Using Git, pull down the latest code from the url above.
-
-Run the command to build the code {{mvn clean install}}
-
-If all goes well, here are the artifacts that you will be interested in:
-* Rya-WAR : web/web-rya/target/web.rya.war
-
-h3. Cloudbase
-
-To build cloudbase projects, run with the 'cloudbase' profile: {{mvn clean install -Pcloudbase}}

http://git-wip-us.apache.org/repos/asf/incubator-rya/blob/ce4a10ff/extras/rya.manual/src/main/webapp/css/pygmentize.css
----------------------------------------------------------------------
diff --git a/extras/rya.manual/src/main/webapp/css/pygmentize.css b/extras/rya.manual/src/main/webapp/css/pygmentize.css
deleted file mode 100644
index 80efdcd..0000000
--- a/extras/rya.manual/src/main/webapp/css/pygmentize.css
+++ /dev/null
@@ -1,134 +0,0 @@
-.syntax .hll { background-color: #ffffcc }
-.syntax  { background: #f0f0f0; }
-.syntax .c { color: #60a0b0; font-style: italic } /* Comment */
-.syntax .err { border: 1px solid #FF0000 } /* Error */
-.syntax .k { color: #007020; font-weight: bold } /* Keyword */
-.syntax .o { color: #666666 } /* Operator */
-.syntax .cm { color: #60a0b0; font-style: italic } /* Comment.Multiline */
-.syntax .cp { color: #007020 } /* Comment.Preproc */
-.syntax .c1 { color: #60a0b0; font-style: italic } /* Comment.Single */
-.syntax .cs { color: #60a0b0; background-color: #fff0f0 } /* Comment.Special */
-.syntax .gd { color: #A00000 } /* Generic.Deleted */
-.syntax .ge { font-style: italic } /* Generic.Emph */
-.syntax .gr { color: #FF0000 } /* Generic.Error */
-.syntax .gh { color: #000080; font-weight: bold } /* Generic.Heading */
-.syntax .gi { color: #00A000 } /* Generic.Inserted */
-.syntax .go { color: #808080 } /* Generic.Output */
-.syntax .gp { color: #c65d09; font-weight: bold } /* Generic.Prompt */
-.syntax .gs { font-weight: bold } /* Generic.Strong */
-.syntax .gu { color: #800080; font-weight: bold } /* Generic.Subheading */
-.syntax .gt { color: #0040D0 } /* Generic.Traceback */
-.syntax .kc { color: #007020; font-weight: bold } /* Keyword.Constant */
-.syntax .kd { color: #007020; font-weight: bold } /* Keyword.Declaration */
-.syntax .kn { color: #007020; font-weight: bold } /* Keyword.Namespace */
-.syntax .kp { color: #007020 } /* Keyword.Pseudo */
-.syntax .kr { color: #007020; font-weight: bold } /* Keyword.Reserved */
-.syntax .kt { color: #902000 } /* Keyword.Type */
-.syntax .m { color: #40a070 } /* Literal.Number */
-.syntax .s { color: #4070a0 } /* Literal.String */
-.syntax .na { color: #4070a0 } /* Name.Attribute */
-.syntax .nb { color: #007020 } /* Name.Builtin */
-.syntax .nc { color: #0e84b5; font-weight: bold } /* Name.Class */
-.syntax .no { color: #60add5 } /* Name.Constant */
-.syntax .nd { color: #555555; font-weight: bold } /* Name.Decorator */
-.syntax .ni { color: #d55537; font-weight: bold } /* Name.Entity */
-.syntax .ne { color: #007020 } /* Name.Exception */
-.syntax .nf { color: #06287e } /* Name.Function */
-.syntax .nl { color: #002070; font-weight: bold } /* Name.Label */
-.syntax .nn { color: #0e84b5; font-weight: bold } /* Name.Namespace */
-.syntax .nt { color: #062873; font-weight: bold } /* Name.Tag */
-.syntax .nv { color: #bb60d5 } /* Name.Variable */
-.syntax .ow { color: #007020; font-weight: bold } /* Operator.Word */
-.syntax .w { color: #bbbbbb } /* Text.Whitespace */
-.syntax .mf { color: #40a070 } /* Literal.Number.Float */
-.syntax .mh { color: #40a070 } /* Literal.Number.Hex */
-.syntax .mi { color: #40a070 } /* Literal.Number.Integer */
-.syntax .mo { color: #40a070 } /* Literal.Number.Oct */
-.syntax .sb { color: #4070a0 } /* Literal.String.Backtick */
-.syntax .sc { color: #4070a0 } /* Literal.String.Char */
-.syntax .sd { color: #4070a0; font-style: italic } /* Literal.String.Doc */
-.syntax .s2 { color: #4070a0 } /* Literal.String.Double */
-.syntax .se { color: #4070a0; font-weight: bold } /* Literal.String.Escape */
-.syntax .sh { color: #4070a0 } /* Literal.String.Heredoc */
-.syntax .si { color: #70a0d0; font-style: italic } /* Literal.String.Interpol */
-.syntax .sx { color: #c65d09 } /* Literal.String.Other */
-.syntax .sr { color: #235388 } /* Literal.String.Regex */
-.syntax .s1 { color: #4070a0 } /* Literal.String.Single */
-.syntax .ss { color: #517918 } /* Literal.String.Symbol */
-.syntax .bp { color: #007020 } /* Name.Builtin.Pseudo */
-.syntax .vc { color: #bb60d5 } /* Name.Variable.Class */
-.syntax .vg { color: #bb60d5 } /* Name.Variable.Global */
-.syntax .vi { color: #bb60d5 } /* Name.Variable.Instance */
-.syntax .il { color: #40a070 } /* Literal.Number.Integer.Long */
-
-
-/* don't highlight errors */
-.syntax .err {
-  border: none;
-}
-
-.syntax {
-  font-size: .9em;
-  font-family:Monaco,"Courier New","DejaVu Sans Mono","Bitstream Vera Sans Mono",monospace;
-  background-color: #F8F8FF; 
-  
-  overflow:auto;
-  -moz-background-clip:border;
-  -moz-background-inline-policy:continuous;
-  -moz-background-origin:padding;
-  margin: 1em 0 1em 0;
-  border:1px solid #DDDDDD;
-  
-  border-top-left-radius: 8px; -webkit-border-top-left-radius: 8px; -moz-border-radius-topleft: 8px;  
-  border-top-right-radius: 8px; -webkit-border-top-right-radius: 8px; -moz-border-radius-topright: 8px;
-  border-style: solid;  border-width: 1px; border-color: #dedede !important; 
-  padding: 1em;
-}
-.syntax .linenodiv  {
-  background-color:#ECECEC;
-  border-right:1px solid #DDDDDD;
-  color:#AAAAAA;
-  padding: .5em;
-  text-align:right;
-}
-.syntax .highlight  {
-}
-.syntax pre {
-  margin:0;
-}
-
-pre.syntax {
-  padding: .5em;
-  background-color: #F8F8FF; overflow:auto;
-}
-
-.syntax code {
-  font-family:Monaco,"Courier New","DejaVu Sans Mono","Bitstream Vera Sans Mono",monospace;
-  font-size: 10pt;
-}
-
-div.compare { width: 700px; }
-div.compare div.compare-left { float:left; width:340px; padding:5px; margin-top: 15px; }
-div.compare div.compare-right { float:right; width:340px; padding:5px; margin-top: 15px; }
-div.compare div h3 {
-  margin-left: 15px;
-  padding: 5px 15px;
-  display: inline;
-  font-size: .8em;
-  color: #666;
-
-  border-top: 1px solid #ccc; -moz-border-top-colors: #ccc white white #e5e5e5;
-  border-left: 1px solid #ccc; -moz-border-left-colors: #ccc white white #e5e5e5;
-  border-right: 1px solid #ccc;-moz-border-right-colors: #ccc white white #e5e5e5;
-  border-top-left-radius: 8px; -webkit-border-top-left-radius: 8px;  -moz-border-radius-topleft: 8px;
-  border-top-right-radius: 8px; -webkit-border-top-right-radius: 8px; -moz-border-radius-topright: 8px; 
-}
-div.compare div div {
-  margin: 5px 0px 0px 0px;
-}
-.clear {
-  clear:both;
-}
-.wide div.compare div.compare-left { float:none; width:700px; }
-.wide div.compare div.compare-right { float:none; width:700px; }
-

http://git-wip-us.apache.org/repos/asf/incubator-rya/blob/ce4a10ff/extras/rya.manual/src/main/webapp/css/scalate/console.css
----------------------------------------------------------------------
diff --git a/extras/rya.manual/src/main/webapp/css/scalate/console.css b/extras/rya.manual/src/main/webapp/css/scalate/console.css
deleted file mode 100644
index 4b2dfe7..0000000
--- a/extras/rya.manual/src/main/webapp/css/scalate/console.css
+++ /dev/null
@@ -1,71 +0,0 @@
-/**
- * Copyright (C) 2009-2010 the original author or authors.
- * See the notice.md file distributed with this work for additional
- * information regarding copyright ownership.
- *
- * Licensed under the Apache License, Version 2.0 (the "License");
- * you may not use this file except in compliance with the License.
- * You may obtain a copy of the License at
- *
- *     http://www.apache.org/licenses/LICENSE-2.0
- *
- * Unless required by applicable law or agreed to in writing, software
- * distributed under the License is distributed on an "AS IS" BASIS,
- * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
- * See the License for the specific language governing permissions and
- * limitations under the License.
- */
-
-#scalate-console {
-  position:fixed;
-  top:40px;
-  right:10px;
-  background: #FDFCAD;
-  padding:10px;
-  border: 1px solid #BDBD81;
-  font-family:Arial,Verdana,Helvetica,sans-serif;
-  font-size: 80%;
-  max-width: 400px;
-}
-
-#scalate-console .archetypes ul, #scalate-console .templates ul, #scalate-console .resources ul, #scalate-console .layouts ul   {
-  padding-left:20px;
-  overflow: auto;
-}
-
-#scalate-console .toggle {
-  text-align: right;
-}
-
-#scalate-console .toggle img {
-  vertical-align: middle;
-}
-
-#scalate-console div.attributes {
-  margin: 10px;
-  overflow: auto;
-  background: white;
-}
-
-#scalate-console div.attributes table th {
-  text-align: left;
-  color: #993333;
-}
-#scalate-console div.attributes table td {
-  padding:2px;
-  border-top: 1px solid #BDBD81;
-}
-
-#scalate-console div.body {
-  margin: 10px;
-  overflow: auto;
-  max-height: 300px;
-  background: white;
-}
-
-#scalate-console div.systemProperties {
-  margin: 10px;
-  overflow: auto;
-  max-height: 300px;
-  background: white;
-}