You are viewing a plain text version of this content. The canonical link for it is here.
Posted to commits@tinkerpop.apache.org by sp...@apache.org on 2020/02/20 11:20:44 UTC

[tinkerpop] 02/02: Merge branch '3.4-dev'

This is an automated email from the ASF dual-hosted git repository.

spmallette pushed a commit to branch master
in repository https://gitbox.apache.org/repos/asf/tinkerpop.git

commit ad66b5d5566b52421198f6a2a4d5ac0d96b10103
Merge: c3b07a8 ed03483
Author: Stephen Mallette <sp...@genoprime.com>
AuthorDate: Thu Feb 20 06:20:27 2020 -0500

    Merge branch '3.4-dev'

 CHANGELOG.asciidoc                                 |  2 +-
 docs/src/upgrade/release-3.4.x.asciidoc            | 12 ++++++
 .../strategy/optimization/LazyBarrierStrategy.java |  5 +++
 .../optimization/LazyBarrierStrategyTest.java      |  9 ++++-
 .../Gherkin/GherkinTestRunner.cs                   |  2 +-
 .../ModernGraphTypeInformation.cs                  |  1 +
 .../src/main/python/radish/feature_steps.py        |  2 +
 gremlin-test/features/filter/Drop.feature          | 43 +++++++++++++++++++++-
 .../process/traversal/step/filter/DropTest.java    | 37 +++++++++++++++++++
 9 files changed, 109 insertions(+), 4 deletions(-)

diff --cc gremlin-dotnet/test/Gremlin.Net.IntegrationTest/Gherkin/GherkinTestRunner.cs
index 0a3d5ed,59fd11e..2649e68
--- a/gremlin-dotnet/test/Gremlin.Net.IntegrationTest/Gherkin/GherkinTestRunner.cs
+++ b/gremlin-dotnet/test/Gremlin.Net.IntegrationTest/Gherkin/GherkinTestRunner.cs
@@@ -41,8 -41,7 +41,8 @@@ namespace Gremlin.Net.IntegrationTest.G
              new Dictionary<string, IgnoreReason>
              {
                  // Add here the name of scenarios to ignore and the reason, e.g.:
 +                { "g_V_group_byXageX", IgnoreReason.NullKeysInMapNotSupported }
-                 // { "g_V_peerPressure_withXpropertyName_clusterX_withXedges_outEXknowsXX_pageRankX1X_byXrankX_withXedges_outEXknowsX_withXtimes_2X_group_byXclusterX_byXrank_sumX_limitX100X", IgnoreReason.NoReason },
+                 //{ "g_V_properties_propertiesXstartTimeX_drop", IgnoreReason.NoReason },
              };
          
          private static class Keywords
diff --cc gremlin-python/src/main/python/radish/feature_steps.py
index 0b26bf4,0000000..5b9808f
mode 100644,000000..100644
--- a/gremlin-python/src/main/python/radish/feature_steps.py
+++ b/gremlin-python/src/main/python/radish/feature_steps.py
@@@ -1,293 -1,0 +1,295 @@@
 +#
 +# Licensed to the Apache Software Foundation (ASF) under one
 +# or more contributor license agreements.  See the NOTICE file
 +# distributed with this work for additional information
 +# regarding copyright ownership.  The ASF licenses this file
 +# to you under the Apache License, Version 2.0 (the
 +# "License"); you may not use this file except in compliance
 +# with the License.  You may obtain a copy of the License at
 +# 
 +# http://www.apache.org/licenses/LICENSE-2.0
 +# 
 +# Unless required by applicable law or agreed to in writing,
 +# software distributed under the License is distributed on an
 +# "AS IS" BASIS, WITHOUT WARRANTIES OR CONDITIONS OF ANY
 +# KIND, either express or implied.  See the License for the
 +# specific language governing permissions and limitations
 +# under the License.
 +#
 +
 +import json
 +import re
 +from gremlin_python.statics import long
 +from gremlin_python.structure.graph import Path
 +from gremlin_python.process.anonymous_traversal import traversal
 +from gremlin_python.process.graph_traversal import __
 +from gremlin_python.process.traversal import Barrier, Cardinality, P, TextP, Pop, Scope, Column, Order, Direction, T, Pick, Operator, IO, WithOptions
 +from radish import given, when, then
 +from hamcrest import *
 +
 +regex_all = re.compile(r"Pop\.all")
 +regex_and = re.compile(r"([(.,\s])and\(")
 +regex_as = re.compile(r"([(.,\s])as\(")
 +regex_from = re.compile(r"([(.,\s])from\(")
 +regex_global = re.compile(r"([(.,\s])global")
++regex_cardlist = re.compile(r"Cardinality\.list")
 +regex_in = re.compile(r"([(.,\s])in\(")
 +regex_is = re.compile(r"([(.,\s])is\(")
 +regex_not = re.compile(r"([(.,\s])not\(")
 +regex_or = re.compile(r"([(.,\s])or\(")
 +regex_with = re.compile(r"([(.,\s])with\(")
 +regex_true = re.compile(r"(true)")
 +regex_false = re.compile(r"(false)")
 +regex_null = re.compile(r"(null)")
 +
 +outV = __.outV
 +label = __.label
 +inV = __.inV
 +project = __.project
 +tail = __.tail
 +
 +ignores = []
 +
 +
 +@given("the {graph_name:w} graph")
 +def choose_graph(step, graph_name):
 +    step.context.graph_name = graph_name
 +    step.context.g = traversal().withRemote(step.context.remote_conn[graph_name])
 +
 +
 +@given("the graph initializer of")
 +def initialize_graph(step):
 +    t = _make_traversal(step.context.g, step.text, {})
 +
 +    # just be sure that the traversal returns something to prove that it worked to some degree. probably
 +    # is overkill to try to assert the complete success of this init operation. presumably the test
 +    # suite would fail elsewhere if this didn't work which would help identify a problem.
 +    result = t.toList()
 +    assert len(result) > 0
 +
 +    # add the first result - if a map - to the bindings. this is useful for cases when parameters for
 +    # the test traversal need to come from the original graph initializer (i.e. a new graph is created
 +    # and you need the id of a vertex from that graph)
 +
 +
 +@given("an unsupported test")
 +def unsupported_scenario(step):
 +    # this is a do nothing step as the test can't be supported for whatever reason
 +    return
 +
 +
 +@given("using the parameter {param_name:w} defined as {param:QuotedString}")
 +def add_parameter(step, param_name, param):
 +    if not hasattr(step.context, "traversal_params"):
 +        step.context.traversal_params = {}
 +
 +    step.context.traversal_params[param_name] = _convert(param.replace('\\"', '"'), step.context)
 +
 +
 +@given("the traversal of")
 +def translate_traversal(step):
 +    step.context.ignore = ignores.__contains__(step.text)
 +    step.context.traversal = _make_traversal(
 +        step.context.g, step.text,
 +        step.context.traversal_params if hasattr(step.context, "traversal_params") else {})
 +
 +
 +@when("iterated to list")
 +def iterate_the_traversal(step):
 +    if step.context.ignore:
 +        return
 +    
 +    step.context.result = list(map(lambda x: _convert_results(x), step.context.traversal.toList()))
 +
 +
 +@when("iterated next")
 +def next_the_traversal(step):
 +    if step.context.ignore:
 +        return
 +
 +    step.context.result = list(map(lambda x: _convert_results(x), step.context.traversal.next()))
 +
 +
 +@then("the result should be {characterized_as:w}")
 +def assert_result(step, characterized_as):
 +    if step.context.ignore:
 +        return
 +
 +    if characterized_as == "empty":        # no results
 +        assert_that(len(step.context.result), equal_to(0))
 +    elif characterized_as == "ordered":    # results asserted in the order of the data table
 +        _table_assertion(step.table, step.context.result, step.context, True)
 +    elif characterized_as == "unordered":  # results asserted in any order
 +        _table_assertion(step.table, step.context.result, step.context, False)
 +    elif characterized_as == "of":         # results may be of any of the specified items in the data table
 +        _any_assertion(step.table, step.context.result, step.context)
 +    else:
 +        raise ValueError("unknown data characterization of " + characterized_as)
 +
 +
 +@then("the graph should return {count:d} for count of {traversal_string:QuotedString}")
 +def assert_side_effects(step, count, traversal_string):
 +    if step.context.ignore:
 +        return
 +
 +    t = _make_traversal(step.context.g, traversal_string.replace('\\"', '"'),
 +                        step.context.traversal_params if hasattr(step.context, "traversal_params") else {})
 +    assert_that(t.count().next(), equal_to(count))
 +
 +
 +@then("the result should have a count of {count:d}")
 +def assert_count(step, count):
 +    assert_that(len(list(step.context.result)), equal_to(count))
 +
 +
 +@then("nothing should happen because")
 +def nothing_happening(step):
 +    return
 +
 +
 +def _convert(val, ctx):
 +    graph_name = ctx.graph_name
 +    if isinstance(val, dict):                                           # convert dictionary keys/values
 +        n = {}
 +        for key, value in val.items():
 +            n[_convert(key, ctx)] = _convert(value, ctx)
 +        return n
 +    elif isinstance(val, str) and re.match(r"^l\[.*\]$", val):           # parse list
 +        return [] if val == "l[]" else list(map((lambda x: _convert(x, ctx)), val[2:-1].split(",")))
 +    elif isinstance(val, str) and re.match(r"^s\[.*\]$", val):           # parse set
 +        return set() if val == "s[]" else set(map((lambda x: _convert(x, ctx)), val[2:-1].split(",")))
 +    elif isinstance(val, str) and re.match(r"^d\[.*\]\.[ilfdm]$", val):  # parse numeric
 +        return float(val[2:-3]) if val[2:-3].__contains__(".") else long(val[2:-3])
 +    elif isinstance(val, str) and re.match(r"^v\[.*\]\.id$", val):       # parse vertex id
 +        return __find_cached_element(ctx, graph_name, val[2:-4], "v").id
 +    elif isinstance(val, str) and re.match(r"^v\[.*\]\.sid$", val):      # parse vertex id as string
 +        return str(__find_cached_element(ctx, graph_name, val[2:-5], "v").id)
 +    elif isinstance(val, str) and re.match(r"^v\[.*\]$", val):           # parse vertex
 +        return __find_cached_element(ctx, graph_name, val[2:-1], "v")
 +    elif isinstance(val, str) and re.match(r"^e\[.*\]\.id$", val):       # parse edge id
 +        return __find_cached_element(ctx, graph_name, val[2:-4], "e").id
 +    elif isinstance(val, str) and re.match(r"^e\[.*\]\.sid$", val):      # parse edge id as string
 +        return str(__find_cached_element(ctx, graph_name, val[2:-5], "e").id)
 +    elif isinstance(val, str) and re.match(r"^e\[.*\]$", val):           # parse edge
 +        return __find_cached_element(ctx, graph_name, val[2:-1], "e")
 +    elif isinstance(val, str) and re.match(r"^m\[.*\]$", val):           # parse json as a map
 +        return _convert(json.loads(val[2:-1]), ctx)
 +    elif isinstance(val, str) and re.match(r"^p\[.*\]$", val):           # parse path
 +        path_objects = list(map((lambda x: _convert(x, ctx)), val[2:-1].split(",")))
 +        return Path([set([])], path_objects)
 +    elif isinstance(val, str) and re.match(r"^c\[.*\]$", val):           # parse lambda/closure
 +        return lambda: (val[2:-1], "gremlin-groovy")
 +    elif isinstance(val, str) and re.match(r"^t\[.*\]$", val):           # parse instance of T enum
 +        return T[val[2:-1]]
 +    elif isinstance(val, str) and re.match(r"^D\[.*\]$", val):           # parse instance of Direction enum
 +        return Direction[val[2:-1]]
 +    elif isinstance(val, str) and re.match(r"^null$", val):              # parse null to None
 +        return None
 +    else:
 +        return val
 +
 +
 +def __find_cached_element(ctx, graph_name, identifier, element_type):
 +    if graph_name == "empty":
 +        cache = __create_lookup_v(ctx.remote_conn["empty"]) if element_type == "v" else __create_lookup_e(ctx.remote_conn["empty"])
 +    else:
 +        cache = ctx.lookup_v[graph_name] if element_type == "v" else ctx.lookup_e[graph_name]
 +
 +    return cache[identifier]
 +
 +
 +def _convert_results(val):
 +    if isinstance(val, Path):
 +        # kill out labels as they aren't in the assertion logic
 +        return Path([set([])], val.objects)
 +    else:
 +        return val
 +
 +
 +def _any_assertion(data, result, ctx):
 +    converted = [_convert(line['result'], ctx) for line in data]
 +    for r in result:
 +        assert_that(r, is_in(converted))
 +
 +
 +def _table_assertion(data, result, ctx, ordered):
 +    # results from traversal should have the same number of entries as the feature data table
 +    assert_that(len(result), equal_to(len(data)), "result:" + str(result))
 +
 +    results_to_test = list(result)
 +
 +    # finds a match in the results for each line of data to assert and then removes that item
 +    # from the list - in the end there should be no items left over and each will have been asserted
 +    for ix, line in enumerate(data):
 +        val = _convert(line['result'], ctx)
 +
 +        # clear the labels since we don't define them in .feature files
 +        if isinstance(val, Path):
 +            val.labels = [set([])]
 +
 +        if ordered:
 +            assert_that(results_to_test[ix], equal_to(val))
 +        else:
 +            assert_that(val, is_in(results_to_test))
 +            results_to_test.remove(val)
 +
 +    if not ordered:
 +        assert_that(len(results_to_test), is_(0))
 +
 +
 +def _translate(traversal_):
 +    replaced = traversal_.replace("\n", "")
 +    replaced = regex_all.sub(r"Pop.all_", replaced)
++    replaced = regex_cardlist.sub(r"Cardinality.list_", replaced)
 +    replaced = regex_and.sub(r"\1and_(", replaced)
 +    replaced = regex_from.sub(r"\1from_(", replaced)
 +    replaced = regex_global.sub(r"\1global_", replaced)
 +    replaced = regex_as.sub(r"\1as_(", replaced)
 +    replaced = regex_is.sub(r"\1is_(", replaced)
 +    replaced = regex_not.sub(r"\1not_(", replaced)
 +    replaced = regex_or.sub(r"\1or_(", replaced)
 +    replaced = regex_in.sub(r"\1in_(", replaced)
 +    replaced = regex_with.sub(r"\1with_(", replaced)
 +    replaced = regex_true.sub(r"True", replaced)
 +    replaced = regex_false.sub(r"False", replaced)
 +    return regex_null.sub(r"None", replaced)
 +
 +
 +def _make_traversal(g, traversal_string, params):
 +    b = {"g": g,
 +         "__": __,
 +         "Barrier": Barrier,
 +         "Cardinality": Cardinality,
 +         "Column": Column,
 +         "Direction": Direction,
 +         "Order": Order,
 +         "P": P,
 +         "TextP": TextP,
 +         "IO": IO,
 +         "Pick": Pick,
 +         "Pop": Pop,
 +         "Scope": Scope,
 +         "Operator": Operator,
 +         "T": T,
 +         "WithOptions": WithOptions}
 +
 +    b.update(params)
 +
 +    return eval(_translate(traversal_string), b)
 +
 +
 +def __create_lookup_v(remote):
 +    g = traversal().withRemote(remote)
 +
 +    # hold a map of name/vertex for use in asserting results
 +    return g.V().group().by('name').by(tail()).next()
 +
 +
 +def __create_lookup_e(remote):
 +    g = traversal().withRemote(remote)
 +
 +    # hold a map of the "name"/edge for use in asserting results - "name" in this context is in the form of
 +    # outgoingV-label->incomingV
 +    return g.E().group(). \
 +        by(lambda: ("it.outVertex().value('name') + '-' + it.label() + '->' + it.inVertex().value('name')", "gremlin-groovy")). \
 +        by(tail()).next()