You are viewing a plain text version of this content. The canonical link for it is here.
Posted to issues@phoenix.apache.org by "Hadoop QA (Jira)" <ji...@apache.org> on 2020/09/16 06:54:00 UTC
[jira] [Commented] (PHOENIX-6136)
javax.servlet.UnavailableException thrown when using Spark connector
[ https://issues.apache.org/jira/browse/PHOENIX-6136?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=17196723#comment-17196723 ]
Hadoop QA commented on PHOENIX-6136:
------------------------------------
{color:red}-1 overall{color}. Here are the results of testing the latest attachment
http://issues.apache.org/jira/secure/attachment/13011592/PHOENIX-6136.master.v1.patch
against master branch at commit .
ATTACHMENT ID: 13011592
{color:green}+1 @author{color}. The patch does not contain any @author tags.
{color:green}+0 tests included{color}. The patch appears to be a documentation, build,
or dev patch that doesn't require tests.
{color:green}+1 javac{color}. The applied patch does not increase the total number of javac compiler warnings.
{color:green}+1 release audit{color}. The applied patch does not increase the total number of release audit warnings.
{color:green}+1 lineLengths{color}. The patch does not introduce lines longer than 100
{color:red}-1 core tests{color}. The patch failed these unit tests:
{color:red}-1 core zombie tests{color}. There are 7 zombie test(s): at org.apache.phoenix.end2end.InListIT.testBaseTableAndIndexTableHaveRightScan(InListIT.java:1580)
at org.apache.phoenix.end2end.EmptyColumnIT.testWhenTableWithIndexAndVariousOptions(EmptyColumnIT.java:492)
at org.apache.phoenix.end2end.InstrFunctionIT.testSingleByteInstrDescendingNoString(InstrFunctionIT.java:92)
at org.apache.phoenix.end2end.IndexToolForDeleteBeforeRebuildIT.testDeleteBeforeRebuildForGlobalIndex(IndexToolForDeleteBeforeRebuildIT.java:144)
at org.apache.phoenix.end2end.DeleteIT.testDeleteForTableWithRowTimestampCol(DeleteIT.java:694)
at org.apache.phoenix.end2end.DeleteIT.testDeleteForTableWithRowTimestampColClient(DeleteIT.java:684)
Test results: https://ci-hadoop.apache.org/job/PreCommit-PHOENIX-Build/93//testReport/
Code Coverage results: https://ci-hadoop.apache.org/job/PreCommit-PHOENIX-Build/93//artifact/phoenix-core/target/site/jacoco/index.html
Console output: https://ci-hadoop.apache.org/job/PreCommit-PHOENIX-Build/93//console
This message is automatically generated.
> javax.servlet.UnavailableException thrown when using Spark connector
> --------------------------------------------------------------------
>
> Key: PHOENIX-6136
> URL: https://issues.apache.org/jira/browse/PHOENIX-6136
> Project: Phoenix
> Issue Type: Bug
> Components: core, spark-connector
> Affects Versions: 5.1.0
> Reporter: Tamas Adami
> Assignee: Istvan Toth
> Priority: Major
> Attachments: PHOENIX-6136.master.v1.patch
>
> Time Spent: 10m
> Remaining Estimate: 0h
>
> We get en exception from Spark when using the phoenix-spark connector.
> {noformat}
> 2020-09-11 13:29:28,252 WARN [main] component.AbstractLifeCycle: FAILED org.glassfish.jersey.servlet.ServletContainer-62c42a3@f5c8fa49==org.glassfish.jersey.servlet.ServletContainer,jsp=null,order=-1,inst=true,async=true: javax.servlet.UnavailableException: Servlet class org.glassfish.jersey.servlet.ServletContainer is not a javax.servlet.Servlet
> javax.servlet.UnavailableException: Servlet class org.glassfish.jersey.servlet.ServletContainer is not a javax.servlet.Servlet
> at org.spark_project.jetty.servlet.ServletHolder.checkServletType(ServletHolder.java:505)
> This seems to be caused by a conflict between the servlet implementation of jetty, and jersey.{noformat}
> Shading both Jersey libraries in phoenix-client has solved the problem for us.
--
This message was sent by Atlassian Jira
(v8.3.4#803005)