You are viewing a plain text version of this content. The canonical link for it is here.
Posted to issues@flink.apache.org by "Gyula Fora (Jira)" <ji...@apache.org> on 2023/03/06 16:12:00 UTC

[jira] [Assigned] (FLINK-31345) Trim autoscaler configMap to not exceed 1mb size limit

     [ https://issues.apache.org/jira/browse/FLINK-31345?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ]

Gyula Fora reassigned FLINK-31345:
----------------------------------

    Assignee: Gyula Fora

> Trim autoscaler configMap to not exceed 1mb size limit
> ------------------------------------------------------
>
>                 Key: FLINK-31345
>                 URL: https://issues.apache.org/jira/browse/FLINK-31345
>             Project: Flink
>          Issue Type: Bug
>          Components: Autoscaler, Kubernetes Operator
>    Affects Versions: kubernetes-operator-1.4.0
>            Reporter: Maximilian Michels
>            Assignee: Gyula Fora
>            Priority: Major
>             Fix For: kubernetes-operator-1.5.0
>
>
> When the {{autoscaler-<deployment_name>}} ConfigMap which is used to persist scaling decisions and metrics becomes too large, the following error is thrown consistently:
> {noformat}
> io.fabric8.kubernetes.client.KubernetesClientException: Operation: [replace]  for kind: [ConfigMap]  with name: [deployment]  in namespace: [namespace]  failed.
>     at io.fabric8.kubernetes.client.KubernetesClientException.launderThrowable(KubernetesClientException.java:159)
>     at io.fabric8.kubernetes.client.dsl.internal.HasMetadataOperation.lambda$replace$0(HasMetadataOperation.java:169)
>     at io.fabric8.kubernetes.client.dsl.internal.HasMetadataOperation.replace(HasMetadataOperation.java:172)
>     at io.fabric8.kubernetes.client.dsl.internal.HasMetadataOperation.replace(HasMetadataOperation.java:113)
>     at io.fabric8.kubernetes.client.dsl.internal.HasMetadataOperation.replace(HasMetadataOperation.java:41)
>     at io.fabric8.kubernetes.client.extension.ResourceAdapter.replace(ResourceAdapter.java:252)
>     at org.apache.flink.kubernetes.operator.autoscaler.AutoScalerInfo.replaceInKubernetes(AutoScalerInfo.java:167)
>     at org.apache.flink.kubernetes.operator.autoscaler.JobAutoScalerImpl.scale(JobAutoScalerImpl.java:113)
>     at org.apache.flink.kubernetes.operator.reconciler.deployment.AbstractFlinkResourceReconciler.reconcile(AbstractFlinkResourceReconciler.java:178)
>     at org.apache.flink.kubernetes.operator.controller.FlinkDeploymentController.reconcile(FlinkDeploymentController.java:130)
>     at org.apache.flink.kubernetes.operator.controller.FlinkDeploymentController.reconcile(FlinkDeploymentController.java:56)
>     at io.javaoperatorsdk.operator.processing.Controller$1.execute(Controller.java:145)
>     at io.javaoperatorsdk.operator.processing.Controller$1.execute(Controller.java:103)
>     at org.apache.flink.kubernetes.operator.metrics.OperatorJosdkMetrics.timeControllerExecution(OperatorJosdkMetrics.java:80)
>     at io.javaoperatorsdk.operator.processing.Controller.reconcile(Controller.java:102)
>     at io.javaoperatorsdk.operator.processing.event.ReconciliationDispatcher.reconcileExecution(ReconciliationDispatcher.java:139)
>     at io.javaoperatorsdk.operator.processing.event.ReconciliationDispatcher.handleReconcile(ReconciliationDispatcher.java:119)
>     at io.javaoperatorsdk.operator.processing.event.ReconciliationDispatcher.handleDispatch(ReconciliationDispatcher.java:89)
>     at io.javaoperatorsdk.operator.processing.event.ReconciliationDispatcher.handleExecution(ReconciliationDispatcher.java:62)
>     at io.javaoperatorsdk.operator.processing.event.EventProcessor$ReconcilerExecutor.run(EventProcessor.java:406)
>     at java.base/java.util.concurrent.ThreadPoolExecutor.runWorker(Unknown Source)
>     at java.base/java.util.concurrent.ThreadPoolExecutor$Worker.run(Unknown Source)
>     at java.base/java.lang.Thread.run(Unknown Source)
> Caused by: java.io.IOException: stream was reset: NO_ERROR
>     at io.fabric8.kubernetes.client.dsl.internal.OperationSupport.waitForResult(OperationSupport.java:514)
>     at io.fabric8.kubernetes.client.dsl.internal.OperationSupport.handleResponse(OperationSupport.java:551)
>     at io.fabric8.kubernetes.client.dsl.internal.OperationSupport.handleUpdate(OperationSupport.java:347)
>     at io.fabric8.kubernetes.client.dsl.internal.BaseOperation.handleUpdate(BaseOperation.java:680)
>     at io.fabric8.kubernetes.client.dsl.internal.HasMetadataOperation.lambda$replace$0(HasMetadataOperation.java:167)
>     ... 21 more
> Caused by: okhttp3.internal.http2.StreamResetException: stream was reset: NO_ERROR
>     at okhttp3.internal.http2.Http2Stream.checkOutNotClosed$okhttp(Http2Stream.kt:646)
>     at okhttp3.internal.http2.Http2Stream$FramingSink.emitFrame(Http2Stream.kt:557)
>     at okhttp3.internal.http2.Http2Stream$FramingSink.write(Http2Stream.kt:532)
>     at okio.ForwardingSink.write(ForwardingSink.kt:29)
>     at okhttp3.internal.connection.Exchange$RequestBodySink.write(Exchange.kt:218)
>     at okio.RealBufferedSink.emitCompleteSegments(RealBufferedSink.kt:255)
>     at okio.RealBufferedSink.write(RealBufferedSink.kt:185)
>     at okhttp3.RequestBody$Companion$toRequestBody$2.writeTo(RequestBody.kt:152)
>     at okhttp3.internal.http.CallServerInterceptor.intercept(CallServerInterceptor.kt:59)
>     at okhttp3.internal.http.RealInterceptorChain.proceed(RealInterceptorChain.kt:109)
>     at okhttp3.internal.connection.ConnectInterceptor.intercept(ConnectInterceptor.kt:34)
>     at okhttp3.internal.http.RealInterceptorChain.proceed(RealInterceptorChain.kt:109)
>     at okhttp3.internal.cache.CacheInterceptor.intercept(CacheInterceptor.kt:95)
>     at okhttp3.internal.http.RealInterceptorChain.proceed(RealInterceptorChain.kt:109)
>     at okhttp3.internal.http.BridgeInterceptor.intercept(BridgeInterceptor.kt:83)
>     at okhttp3.internal.http.RealInterceptorChain.proceed(RealInterceptorChain.kt:109)
>     at okhttp3.internal.http.RetryAndFollowUpInterceptor.intercept(RetryAndFollowUpInterceptor.kt:76)
>     at okhttp3.internal.http.RealInterceptorChain.proceed(RealInterceptorChain.kt:109)
>     at org.apache.flink.kubernetes.operator.metrics.KubernetesClientMetrics.intercept(KubernetesClientMetrics.java:130)
>     at okhttp3.internal.http.RealInterceptorChain.proceed(RealInterceptorChain.kt:109)
>     at io.fabric8.kubernetes.client.okhttp.OkHttpClientBuilderImpl$InteceptorAdapter.intercept(OkHttpClientBuilderImpl.java:70)
>     at okhttp3.internal.http.RealInterceptorChain.proceed(RealInterceptorChain.kt:109)
>     at io.fabric8.kubernetes.client.okhttp.OkHttpClientBuilderImpl$InteceptorAdapter.intercept(OkHttpClientBuilderImpl.java:70)
>     at okhttp3.internal.http.RealInterceptorChain.proceed(RealInterceptorChain.kt:109)
>     at io.fabric8.kubernetes.client.okhttp.OkHttpClientBuilderImpl$InteceptorAdapter.intercept(OkHttpClientBuilderImpl.java:70)
>     at okhttp3.internal.http.RealInterceptorChain.proceed(RealInterceptorChain.kt:109)
>     at okhttp3.internal.connection.RealCall.getResponseWithInterceptorChain$okhttp(RealCall.kt:201)
>     at okhttp3.internal.connection.RealCall$AsyncCall.run(RealCall.kt:517)
>     ... 3 more
>     Suppressed: okhttp3.internal.http2.StreamResetException: stream was reset: NO_ERROR
>         ... 31 more
>  {noformat}
> We should trim the ConfigMap to not exceed the size limit.



--
This message was sent by Atlassian Jira
(v8.20.10#820010)