You are viewing a plain text version of this content. The canonical link for it is here.
Posted to commits@cassandra.apache.org by "Russ Hatch (JIRA)" <ji...@apache.org> on 2016/03/19 22:06:33 UTC

[jira] [Comment Edited] (CASSANDRA-11195) paging may returns incomplete results on small page size

    [ https://issues.apache.org/jira/browse/CASSANDRA-11195?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=15202968#comment-15202968 ] 

Russ Hatch edited comment on CASSANDRA-11195 at 3/19/16 9:06 PM:
-----------------------------------------------------------------

here's some tracing from the python driver when the incomplete data is returned:

the first ccm node is acting as coordinator (127.0.0.1):
{noformat}
127.0.0.1 ***** Parsing SELECT * FROM test
127.0.0.1 ***** Preparing statement
127.0.0.1 ***** Computing ranges to query
127.0.0.1 ***** Submitting range requests on 8 ranges with a concurrency of 1 (0.0 rows per range expected)
127.0.0.1 ***** Submitted 1 concurrent range requests
127.0.0.2 ***** RANGE_SLICE message received from /127.0.0.1
127.0.0.1 ***** Executing seq scan across 0 sstables for [3, max(9057792379311864889)]
127.0.0.2 ***** Executing seq scan across 0 sstables for (max(9057792379311864889), max(9063681715701432110)]
127.0.0.1 ***** Read 2 live and 0 tombstone cells
127.0.0.2 ***** Scanned 0 rows and matched 0
127.0.0.1 ***** Enqueuing request to /127.0.0.2
127.0.0.2 ***** Enqueuing response to /127.0.0.1
127.0.0.1 ***** Submitted 1 concurrent range requests
127.0.0.2 ***** Sending REQUEST_RESPONSE message to /127.0.0.1
127.0.0.1 ***** Sending RANGE_SLICE message to /127.0.0.2
127.0.0.1 ***** REQUEST_RESPONSE message received from /127.0.0.2
127.0.0.2 ***** RANGE_SLICE message received from /127.0.0.1
127.0.0.1 ***** Processing response from /127.0.0.2
127.0.0.2 ***** Executing seq scan across 0 sstables for (max(9066543152329569058), max(9149440404233844277)]
127.0.0.1 ***** Enqueuing request to /127.0.0.2
127.0.0.2 ***** Scanned 0 rows and matched 0
127.0.0.1 ***** Submitted 2 concurrent range requests
127.0.0.2 ***** Enqueuing response to /127.0.0.1
127.0.0.1 ***** Executing seq scan across 0 sstables for (max(9063681715701432110), max(9066543152329569058)]
127.0.0.2 ***** Sending REQUEST_RESPONSE message to /127.0.0.1
127.0.0.1 ***** Read 0 live and 0 tombstone cells
127.0.0.1 ***** Sending RANGE_SLICE message to /127.0.0.2
127.0.0.1 ***** REQUEST_RESPONSE message received from /127.0.0.2
127.0.0.1 ***** Processing response from /127.0.0.2
127.0.0.1 ***** Enqueuing request to /127.0.0.2
127.0.0.1 ***** Submitted 2 concurrent range requests
127.0.0.2 ***** RANGE_SLICE message received from /127.0.0.1
127.0.0.1 ***** Executing seq scan across 0 sstables for (max(9149440404233844277), max(9221933478161079123)]
127.0.0.2 ***** Executing seq scan across 0 sstables for [max(9221933478161079123), min(-9223372036854775808)]
127.0.0.1 ***** Read 0 live and 0 tombstone cells
127.0.0.2 ***** Scanned 0 rows and matched 0
127.0.0.1 ***** Sending RANGE_SLICE message to /127.0.0.2
127.0.0.2 ***** Enqueuing response to /127.0.0.1
127.0.0.1 ***** REQUEST_RESPONSE message received from /127.0.0.2
127.0.0.2 ***** Sending REQUEST_RESPONSE message to /127.0.0.1
127.0.0.1 ***** Processing response from /127.0.0.2
{noformat}


was (Author: rhatch):
here's some tracing from the python driver when the error occurs.

the first ccm node is acting as coordinator (127.0.0.1):
{noformat}
127.0.0.1 ***** Parsing SELECT * FROM test
127.0.0.1 ***** Preparing statement
127.0.0.1 ***** Computing ranges to query
127.0.0.1 ***** Submitting range requests on 8 ranges with a concurrency of 1 (0.0 rows per range expected)
127.0.0.1 ***** Submitted 1 concurrent range requests
127.0.0.2 ***** RANGE_SLICE message received from /127.0.0.1
127.0.0.1 ***** Executing seq scan across 0 sstables for [3, max(9057792379311864889)]
127.0.0.2 ***** Executing seq scan across 0 sstables for (max(9057792379311864889), max(9063681715701432110)]
127.0.0.1 ***** Read 2 live and 0 tombstone cells
127.0.0.2 ***** Scanned 0 rows and matched 0
127.0.0.1 ***** Enqueuing request to /127.0.0.2
127.0.0.2 ***** Enqueuing response to /127.0.0.1
127.0.0.1 ***** Submitted 1 concurrent range requests
127.0.0.2 ***** Sending REQUEST_RESPONSE message to /127.0.0.1
127.0.0.1 ***** Sending RANGE_SLICE message to /127.0.0.2
127.0.0.1 ***** REQUEST_RESPONSE message received from /127.0.0.2
127.0.0.2 ***** RANGE_SLICE message received from /127.0.0.1
127.0.0.1 ***** Processing response from /127.0.0.2
127.0.0.2 ***** Executing seq scan across 0 sstables for (max(9066543152329569058), max(9149440404233844277)]
127.0.0.1 ***** Enqueuing request to /127.0.0.2
127.0.0.2 ***** Scanned 0 rows and matched 0
127.0.0.1 ***** Submitted 2 concurrent range requests
127.0.0.2 ***** Enqueuing response to /127.0.0.1
127.0.0.1 ***** Executing seq scan across 0 sstables for (max(9063681715701432110), max(9066543152329569058)]
127.0.0.2 ***** Sending REQUEST_RESPONSE message to /127.0.0.1
127.0.0.1 ***** Read 0 live and 0 tombstone cells
127.0.0.1 ***** Sending RANGE_SLICE message to /127.0.0.2
127.0.0.1 ***** REQUEST_RESPONSE message received from /127.0.0.2
127.0.0.1 ***** Processing response from /127.0.0.2
127.0.0.1 ***** Enqueuing request to /127.0.0.2
127.0.0.1 ***** Submitted 2 concurrent range requests
127.0.0.2 ***** RANGE_SLICE message received from /127.0.0.1
127.0.0.1 ***** Executing seq scan across 0 sstables for (max(9149440404233844277), max(9221933478161079123)]
127.0.0.2 ***** Executing seq scan across 0 sstables for [max(9221933478161079123), min(-9223372036854775808)]
127.0.0.1 ***** Read 0 live and 0 tombstone cells
127.0.0.2 ***** Scanned 0 rows and matched 0
127.0.0.1 ***** Sending RANGE_SLICE message to /127.0.0.2
127.0.0.2 ***** Enqueuing response to /127.0.0.1
127.0.0.1 ***** REQUEST_RESPONSE message received from /127.0.0.2
127.0.0.2 ***** Sending REQUEST_RESPONSE message to /127.0.0.1
127.0.0.1 ***** Processing response from /127.0.0.2
{noformat}

> paging may returns incomplete results on small page size
> --------------------------------------------------------
>
>                 Key: CASSANDRA-11195
>                 URL: https://issues.apache.org/jira/browse/CASSANDRA-11195
>             Project: Cassandra
>          Issue Type: Bug
>            Reporter: Jim Witschey
>              Labels: dtest
>         Attachments: node1.log, node1_debug.log, node2.log, node2_debug.log
>
>
> This was found through a flapping test, and running that test is still the easiest way to repro the issue. On CI we're seeing a 40-50% failure rate, but locally this test fails much less frequently.
> If I attach a python debugger and re-query the "bad" query, it continues to return incomplete data indefinitely. If I go directly to cqlsh I can see all rows just fine.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)