You are viewing a plain text version of this content. The canonical link for it is here.
Posted to issues@activemq.apache.org by "John Behm (Jira)" <ji...@apache.org> on 2021/02/03 10:52:00 UTC

[jira] [Comment Edited] (ARTEMIS-3042) Docker Multistage Build from Source

    [ https://issues.apache.org/jira/browse/ARTEMIS-3042?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=17277902#comment-17277902 ] 

John Behm edited comment on ARTEMIS-3042 at 2/3/21, 10:51 AM:
--------------------------------------------------------------

There is no real docker integration, as there is no official docker image that can be used to setup either a local instance or a kubernetes instance.

There are no examples for either a docker-compose or Kubernetes manifest (yaml) file.

The multistage build allows users to see the build process. This might not be the important part, but explicitly that you can simply build your own image from an officially supported Dockerfile is a giant plus for an easier deployment of ActiveMQ artemis.

As Artemis somehow does advertise as being cloud native, this should be like the basics that need to be fulfilled to have an easy setup, either of one instance or of a cluster in Kubernetes.

From my point of view the two step startup, generation of a broker xml followed by monkey-patching either the generated file or not even using the calculated values in that file and provide your own broker.xml seems to be a bad design decision.

Configuration should be provided via environment variables and should preferrably be somewhat static.

 

I think this kind of configuration is slightly overkill in order to deploy a cluster:

 
{code:yaml}
apiVersion: v1
kind: ConfigMap
metadata: 
  name: artemis-config
  namespace: artemis
data: 
  bootstrap.xml: |-
    <?xml version="1.0" encoding="UTF-8" standalone="yes"?>
    <!--
    ~ Licensed to the Apache Software Foundation (ASF) under one or more
    ~ contributor license agreements. See the NOTICE file distributed with
    ~ this work for additional information regarding copyright ownership.
    ~ The ASF licenses this file to You under the Apache License, Version 2.0
    ~ (the "License"); you may not use this file except in compliance with
    ~ the License. You may obtain a copy of the License at
    ~
    ~     http://www.apache.org/licenses/LICENSE-2.0
    ~
    ~ Unless required by applicable law or agreed to in writing, software
    ~ distributed under the License is distributed on an "AS IS" BASIS,
    ~ WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
    ~ See the License for the specific language governing permissions and
    ~ limitations under the License.
    -->

    <broker xmlns="http://activemq.org/schema">

       <jaas-security domain="activemq"/>


       <!-- artemis.URI.instance is parsed from artemis.instance by the CLI startup.
          This is to avoid situations where you could have spaces or special characters on this URI -->
       <server configuration="file:/data/broker.xml"/><!-- <<<<<<<<<<<<<<<<<<<<<<<========================= -->

       <!-- The web server is only bound to localhost by default -->
       <web bind="http://localhost:8161" path="web">
          <app url="activemq-branding" war="activemq-branding.war"/>
          <app url="artemis-plugin" war="artemis-plugin.war"/>
          <app url="console" war="console.war"/>
       </web>

    </broker>
  broker.xml: |-
    <?xml version='1.0'?>
      <!--
      Licensed to the Apache Software Foundation (ASF) under one
      or more contributor license agreements.  See the NOTICE file
      distributed with this work for additional information
      regarding copyright ownership.  The ASF licenses this file
      to you under the Apache License, Version 2.0 (the
      "License"); you may not use this file except in compliance
      with the License.  You may obtain a copy of the License at

      http://www.apache.org/licenses/LICENSE-2.0

      Unless required by applicable law or agreed to in writing,
      software distributed under the License is distributed on an
      "AS IS" BASIS, WITHOUT WARRANTIES OR CONDITIONS OF ANY
      KIND, either express or implied.  See the License for the
      specific language governing permissions and limitations
      under the License.
      -->

      <configuration xmlns="urn:activemq"
                     xmlns:xsi="http://www.w3.org/2001/XMLSchema-instance"
                     xmlns:xi="http://www.w3.org/2001/XInclude"
                     xsi:schemaLocation="urn:activemq /schema/artemis-configuration.xsd">

         <core xmlns="urn:activemq:core" xmlns:xsi="http://www.w3.org/2001/XMLSchema-instance"
               xsi:schemaLocation="urn:activemq:core ">

            <name>${POD_NAME}/${POD_IP}</name>


            <persistence-enabled>true</persistence-enabled>

            <!-- this could be ASYNCIO, MAPPED, NIO
               ASYNCIO: Linux Libaio
               MAPPED: mmap files
               NIO: Plain Java Files
            -->
            <journal-type>ASYNCIO</journal-type>

            <paging-directory>data/paging</paging-directory>

            <bindings-directory>data/bindings</bindings-directory>

            <journal-directory>data/journal</journal-directory>

            <large-messages-directory>data/large-messages</large-messages-directory>

            <journal-datasync>true</journal-datasync>

            <journal-min-files>2</journal-min-files>

            <journal-pool-files>10</journal-pool-files>

            <journal-device-block-size>4096</journal-device-block-size>

            <journal-file-size>10M</journal-file-size>
            
            <!--
            This value was determined through a calculation.
            Your system could perform 41.67 writes per millisecond
            on the current journal configuration.
            That translates as a sync write every 24000 nanoseconds.

            Note: If you specify 0 the system will perform writes directly to the disk.
                  We recommend this to be 0 if you are using journalType=MAPPED and journal-datasync=false.
            -->
            <journal-buffer-timeout>24000</journal-buffer-timeout>


            <!--
            When using ASYNCIO, this will determine the writing queue depth for libaio.
            -->
            <journal-max-io>4096</journal-max-io>
            <!--
            You can verify the network health of a particular NIC by specifying the <network-check-NIC> element.
               <network-check-NIC>theNicName</network-check-NIC>
            -->

            <!--
            Use this to use an HTTP server to validate the network
               <network-check-URL-list>http://www.apache.org</network-check-URL-list> -->

            <!-- <network-check-period>10000</network-check-period> -->
            <!-- <network-check-timeout>1000</network-check-timeout> -->

            <!-- this is a comma separated list, no spaces, just DNS or IPs
               it should accept IPV6

               Warning: Make sure you understand your network topology as this is meant to validate if your network is valid.
                        Using IPs that could eventually disappear or be partially visible may defeat the purpose.
                        You can use a list of multiple IPs, and if any successful ping will make the server OK to continue running -->
            <!-- <network-check-list>10.0.0.1</network-check-list> -->

            <!-- use this to customize the ping used for ipv4 addresses -->
            <!-- <network-check-ping-command>ping -c 1 -t %d %s</network-check-ping-command> -->

            <!-- use this to customize the ping used for ipv6 addresses -->
            <!-- <network-check-ping6-command>ping6 -c 1 %2$s</network-check-ping6-command> -->



         <connectors>
            <!-- Connector used to be announced through cluster connections and notifications -->
            <connector name="artemis">tcp://${POD_IP}:61616</connector>
         </connectors>



            <!-- how often we are looking for how many bytes are being used on the disk in ms -->
            <disk-scan-period>5000</disk-scan-period>

            <!-- once the disk hits this limit the system will block, or close the connection in certain protocols
               that won't support flow control. -->
            <max-disk-usage>90</max-disk-usage>

            <!-- should the broker detect dead locks and other issues -->
            <critical-analyzer>true</critical-analyzer>

            <critical-analyzer-timeout>120000</critical-analyzer-timeout>

            <critical-analyzer-check-period>60000</critical-analyzer-check-period>

            <critical-analyzer-policy>HALT</critical-analyzer-policy>

            
            <page-sync-timeout>168000</page-sync-timeout>


                  <!-- the system will enter into page mode once you hit this limit.
               This is an estimate in bytes of how much the messages are using in memory

                  The system will use half of the available memory (-Xmx) by default for the global-max-size.
                  You may specify a different value here if you need to customize it to your needs.

                  <global-max-size>100Mb</global-max-size>

            -->

            <acceptors>

               <!-- useEpoll means: it will use Netty epoll if you are on a system (Linux) that supports it -->
               <!-- amqpCredits: The number of credits sent to AMQP producers -->
               <!-- amqpLowCredits: The server will send the # credits specified at amqpCredits at this low mark -->
               <!-- amqpDuplicateDetection: If you are not using duplicate detection, set this to false
                                          as duplicate detection requires applicationProperties to be parsed on the server. -->
               <!-- amqpMinLargeMessageSize: Determines how many bytes are considered large, so we start using files to hold their data.
                                             default: 102400, -1 would mean to disable large mesasge control -->

               <!-- Note: If an acceptor needs to be compatible with HornetQ and/or Artemis 1.x clients add
                        "anycastPrefix=jms.queue.;multicastPrefix=jms.topic." to the acceptor url.
                        See https://issues.apache.org/jira/browse/ARTEMIS-1644 for more information. -->


               <!-- Acceptor for every supported protocol -->
               <acceptor name="artemis">tcp://0.0.0.0:61616?tcpSendBufferSize=1048576;tcpReceiveBufferSize=1048576;amqpMinLargeMessageSize=102400;protocols=CORE,AMQP,STOMP,HORNETQ,MQTT,OPENWIRE;useEpoll=true;amqpCredits=1000;amqpLowCredits=300;amqpDuplicateDetection=true</acceptor>

               <!-- AMQP Acceptor.  Listens on default AMQP port for AMQP traffic.-->
               <acceptor name="amqp">tcp://0.0.0.0:5672?tcpSendBufferSize=1048576;tcpReceiveBufferSize=1048576;protocols=AMQP;useEpoll=true;amqpCredits=1000;amqpLowCredits=300;amqpMinLargeMessageSize=102400;amqpDuplicateDetection=true</acceptor>

               <!-- STOMP Acceptor. -->
               <acceptor name="stomp">tcp://0.0.0.0:61613?tcpSendBufferSize=1048576;tcpReceiveBufferSize=1048576;protocols=STOMP;useEpoll=true</acceptor>

               <!-- HornetQ Compatibility Acceptor.  Enables HornetQ Core and STOMP for legacy HornetQ clients. -->
               <acceptor name="hornetq">tcp://0.0.0.0:5445?anycastPrefix=jms.queue.;multicastPrefix=jms.topic.;protocols=HORNETQ,STOMP;useEpoll=true</acceptor>

               <!-- MQTT Acceptor -->
               <acceptor name="mqtt">tcp://0.0.0.0:1883?tcpSendBufferSize=1048576;tcpReceiveBufferSize=1048576;protocols=MQTT;useEpoll=true</acceptor>

            </acceptors>

             <!-- IMPORTANT: MUST BE CHANGED-->
            <cluster-user>${CLUSTER_USER}</cluster-user>
            <cluster-password>${CLUSTER_PASSWORD}</cluster-password>

            <broadcast-groups>
               <broadcast-group name="bg-group1">
                  <group-address>${BROADCAST_IP}</group-address>
                  <group-port>${BROADCAST_PORT}</group-port>
                  <broadcast-period>5000</broadcast-period>
                  <connector-ref>artemis</connector-ref>
               </broadcast-group>
            </broadcast-groups>

            <discovery-groups>
               <discovery-group name="dg-group1">
                  <group-address>${BROADCAST_IP}</group-address>
                  <group-port>${BROADCAST_PORT}</group-port>
                  <refresh-timeout>10000</refresh-timeout>
               </discovery-group>
            </discovery-groups>

            <cluster-connections>
                <cluster-connection name="artemis">
                   <address></address>
                   <connector-ref>artemis</connector-ref>
                   <check-period>30000</check-period>
                   <connection-ttl>60000</connection-ttl>
                   <min-large-message-size>102400</min-large-message-size>
                   <call-timeout>30000</call-timeout>
                   <retry-interval>1000</retry-interval>
                   <retry-interval-multiplier>1.0</retry-interval-multiplier>
                   <max-retry-interval>2000</max-retry-interval>
                   <initial-connect-attempts>-1</initial-connect-attempts>
                   <reconnect-attempts>-1</reconnect-attempts>
                   <use-duplicate-detection>true</use-duplicate-detection>
                   <message-load-balancing>ON_DEMAND</message-load-balancing>
                   <max-hops>1</max-hops>
                   <confirmation-window-size>1048576</confirmation-window-size>
                   <call-failover-timeout>-1</call-failover-timeout>
                   <notification-interval>1000</notification-interval>
                   <notification-attempts>100</notification-attempts>
                   <discovery-group-ref discovery-group-name="dg-group1"/>
                </cluster-connection>
            </cluster-connections>


            <ha-policy>
               <replication>
                   <colocated>
                      <request-backup>true</request-backup>
                      <max-backups>1</max-backups>
                      <backup-request-retries>-1</backup-request-retries>
                      <backup-request-retry-interval>5000</backup-request-retry-interval>
                      
                      <master>
                         <vote-on-replication-failure>true</vote-on-replication-failure>
                         <check-for-live-server>true</check-for-live-server>
                      </master>
                      <slave>
                         <scale-down/>

                         <!-- when a master returns, let clients reconnect to the master again -->
                         <!- NEEDS TESTING 
                         <allow-failback>true</allow-failback>
                         <failback-delay>10000</failback-delay>
                         -->
                      </slave>
                   </colocated>
                </replication>
             </ha-policy>

            <security-settings>
               <security-setting match="#">
                  <permission type="createNonDurableQueue" roles="amq"/>
                  <permission type="deleteNonDurableQueue" roles="amq"/>
                  <permission type="createDurableQueue" roles="amq"/>
                  <permission type="deleteDurableQueue" roles="amq"/>
                  <permission type="createAddress" roles="amq"/>
                  <permission type="deleteAddress" roles="amq"/>
                  <permission type="consume" roles="amq"/>
                  <permission type="browse" roles="amq"/>
                  <permission type="send" roles="amq"/>
                  <!-- we need this otherwise ./artemis data imp wouldn't work -->
                  <permission type="manage" roles="amq"/>
               </security-setting>
            </security-settings>

            <address-settings>
               <!-- if you define auto-create on certain queues, management has to be auto-create -->
               <address-setting match="activemq.management#">
                  <dead-letter-address>DLQ</dead-letter-address>
                  <expiry-address>ExpiryQueue</expiry-address>
                  <redelivery-delay>0</redelivery-delay>
                  <!-- with -1 only the global-max-size is in use for limiting -->
                  <max-size-bytes>-1</max-size-bytes>
                  <message-counter-history-day-limit>10</message-counter-history-day-limit>
                  <address-full-policy>PAGE</address-full-policy>
                  <auto-create-queues>true</auto-create-queues>
                  <auto-create-addresses>true</auto-create-addresses>
               </address-setting>
               <!--default for catch all-->
               <address-setting match="#">
                  <dead-letter-address>DLQ</dead-letter-address>
                  <expiry-address>ExpiryQueue</expiry-address>
                  <redelivery-delay>0</redelivery-delay>
                  <!-- with -1 only the global-max-size is in use for limiting -->
                  <max-size-bytes>-1</max-size-bytes>
                  <message-counter-history-day-limit>10</message-counter-history-day-limit>
                  <address-full-policy>PAGE</address-full-policy>
                  <auto-create-queues>true</auto-create-queues>
                  <auto-create-addresses>true</auto-create-addresses>
                  <auto-create-jms-queues>true</auto-create-jms-queues>
                  <auto-create-jms-topics>true</auto-create-jms-topics>
                  
                  <!-- redestribute messages in a queue if there is no consumer connected to the current node -->
                  <redistribution-delay>0</redistribution-delay>

               </address-setting>
            </address-settings>

            <addresses>
               <address name="DLQ">
                  <anycast>
                     <queue name="DLQ" />
                  </anycast>
               </address>
               <address name="ExpiryQueue">
                  <anycast>
                     <queue name="ExpiryQueue" />
                  </anycast>
               </address>

            </addresses>


            <!-- Uncomment the following if you want to use the Standard LoggingActiveMQServerPlugin pluging to log in events
            <broker-plugins>
               <broker-plugin class-name="org.apache.activemq.artemis.core.server.plugin.impl.LoggingActiveMQServerPlugin">
                  <property key="LOG_ALL_EVENTS" value="true"/>
                  <property key="LOG_CONNECTION_EVENTS" value="true"/>
                  <property key="LOG_SESSION_EVENTS" value="true"/>
                  <property key="LOG_CONSUMER_EVENTS" value="true"/>
                  <property key="LOG_DELIVERING_EVENTS" value="true"/>
                  <property key="LOG_SENDING_EVENTS" value="true"/>
                  <property key="LOG_INTERNAL_EVENTS" value="true"/>
               </broker-plugin>
            </broker-plugins>
            -->

         </core>
      </configuration>
---
apiVersion: v1
kind: ReplicationController
metadata: 
  name: artemis-rc
  namespace: artemis
spec: 
  replicas: 1
  selector: 
    app: artemis-rc
  template: 
    metadata: 
      name: artemis-test-pod
      namespace: artemis
      labels: 
        app: artemis-rc
    spec: 
      containers: 
        - name: artemis
          image: argos-hub.ws.arvato-systems.de/asys-public/activemq-artemis:2.16.0
          # command: ['sh', '-c', 'sleep 36000']
          args: ["run", "--allow-kill", "--", "xml:/data/bootstrap.xml"]
          volumeMounts: 
            - name: artemis-config
              subPath: bootstrap.xml
              mountPath: /data/bootstrap.xml
              readOnly: true
            - name: artemis-config
              subPath: broker.xml
              mountPath: /data/broker.xml
              readOnly: true
          env: 
            - name: POD_NAME
              valueFrom: 
                fieldRef: 
                  fieldPath: metadata.name
            - name: POD_IP
              valueFrom: 
                fieldRef: 
                  fieldPath: status.podIP
            - name: BROADCAST_IP
              value: 231.7.7.7
            - name: BROADCAST_PORT
              value: "9876"
            - name: "EXTRA_ARGS"
              value: "" # extra creation args, to be precise
            - name: CLUSTER_USER
              value: cluster-admin
            - name: CLUSTER_PASSWORD
              value: password-admin
            - name: ARTEMIS_CLUSTER_PROPS
              value: |-
                -DPOD_NAME=$(POD_NAME) 
                -DPOD_IP=$(POD_IP) 
                -DBROADCAST_IP=$(BROADCAST_IP) 
                -DBROADCAST_PORT=$(BROADCAST_PORT)
                -DCLUSTER_USER=$(CLUSTER_USER) 
                -DCLUSTER_PASSWORD=$(CLUSTER_PASSWORD)
      volumes: 
        - name: artemis-config
          configMap: 
            defaultMode: 420
            name: artemis-config
---
apiVersion: v1
kind: Pod
metadata: 
  name: artemis-main
  namespace: artemis
  labels: 
    app: artemis-main
spec: 
  containers: 
    - name: artemis
      image: argos-hub.ws.arvato-systems.de/asys-public/activemq-artemis:2.16.0
      #command: ['sh', '-c', 'sleep 36000']
      args: ["run", "--allow-kill", "--", "xml:/data/bootstrap.xml"]
      volumeMounts: 
        - name: artemis-config
          subPath: bootstrap.xml
          mountPath: /data/bootstrap.xml
          readOnly: true
        - name: artemis-config
          subPath: broker.xml
          mountPath: /data/broker.xml
          readOnly: true
      env: 
        - name: POD_NAME
          valueFrom: 
            fieldRef: 
              fieldPath: metadata.name
        - name: POD_IP
          valueFrom: 
            fieldRef: 
              fieldPath: status.podIP
        - name: BROADCAST_IP
          value: 231.7.7.7
        - name: BROADCAST_PORT
          value: "9876"
        - name: "EXTRA_ARGS"
          value: "" # extra creation args, to be precise
        - name: CLUSTER_USER
          value: cluster-admin
        - name: CLUSTER_PASSWORD
          value: password-admin
        - name: ARTEMIS_CLUSTER_PROPS
          value: |-
            -DPOD_NAME=$(POD_NAME) 
            -DPOD_IP=$(POD_IP) 
            -DBROADCAST_IP=$(BROADCAST_IP) 
            -DBROADCAST_PORT=$(BROADCAST_PORT)
            -DCLUSTER_USER=$(CLUSTER_USER) 
            -DCLUSTER_PASSWORD=$(CLUSTER_PASSWORD)
  volumes: 
    - name: artemis-config
      configMap: 
        defaultMode: 420
        name: artemis-config
{code}
 


was (Author: behm015):
There is no real docker integration, as there is no official docker image that can be used to setup either a local instance or a kubernetes instance.

There are no examples for either a docker-compose or Kubernetes manifest (yaml) file.

The multistage build allows users to see the build process. This might not be the important part, but explicitly that you can simply build your own image from an officially supported Dockerfile is a giant plus for an easier deployment of ActiveMQ artemis.

As Artemis somehow does advertise as being cloud native, this should be like the basics that need to be fulfilled to have an easy setup, either of one instance or of a cluster in Kubernetes.

From my point of view the two step startup, generation of a broker xml followed by monkey-patching either the generated file or not even using the calculated values in that file and provide your own broker.xml seems to be a bad design decision.

Configuration should be provided via environment variables and should preferrably be somewhat static.

 

I thinkthis kind of configuration in order to get a cluster running is slightly overkill in order to deploy a cluster:

 
{code:yaml}
apiVersion: v1
kind: ConfigMap
metadata:
  name: artemis-config
  namespace: artemis
data:
  bootstrap.xml: |-
    <?xml version="1.0" encoding="UTF-8" standalone="yes"?>
    <!--
    ~ Licensed to the Apache Software Foundation (ASF) under one or more
    ~ contributor license agreements. See the NOTICE file distributed with
    ~ this work for additional information regarding copyright ownership.
    ~ The ASF licenses this file to You under the Apache License, Version 2.0
    ~ (the "License"); you may not use this file except in compliance with
    ~ the License. You may obtain a copy of the License at
    ~
    ~     http://www.apache.org/licenses/LICENSE-2.0
    ~
    ~ Unless required by applicable law or agreed to in writing, software
    ~ distributed under the License is distributed on an "AS IS" BASIS,
    ~ WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
    ~ See the License for the specific language governing permissions and
    ~ limitations under the License.
    -->

    <broker xmlns="http://activemq.org/schema">

       <jaas-security domain="activemq"/>


       <!-- artemis.URI.instance is parsed from artemis.instance by the CLI startup.
          This is to avoid situations where you could have spaces or special characters on this URI -->
       <server configuration="file:/data/broker.xml"/><!-- <<<<<<<<<<<<<<<<<<<<<<<========================= -->

       <!-- The web server is only bound to localhost by default -->
       <web bind="http://localhost:8161" path="web">
          <app url="activemq-branding" war="activemq-branding.war"/>
          <app url="artemis-plugin" war="artemis-plugin.war"/>
          <app url="console" war="console.war"/>
       </web>

    </broker>
  broker.xml: |-
    <?xml version='1.0'?>
      <!--
      Licensed to the Apache Software Foundation (ASF) under one
      or more contributor license agreements.  See the NOTICE file
      distributed with this work for additional information
      regarding copyright ownership.  The ASF licenses this file
      to you under the Apache License, Version 2.0 (the
      "License"); you may not use this file except in compliance
      with the License.  You may obtain a copy of the License at

      http://www.apache.org/licenses/LICENSE-2.0

      Unless required by applicable law or agreed to in writing,
      software distributed under the License is distributed on an
      "AS IS" BASIS, WITHOUT WARRANTIES OR CONDITIONS OF ANY
      KIND, either express or implied.  See the License for the
      specific language governing permissions and limitations
      under the License.
      -->

      <configuration xmlns="urn:activemq"
                     xmlns:xsi="http://www.w3.org/2001/XMLSchema-instance"
                     xmlns:xi="http://www.w3.org/2001/XInclude"
                     xsi:schemaLocation="urn:activemq /schema/artemis-configuration.xsd">

         <core xmlns="urn:activemq:core" xmlns:xsi="http://www.w3.org/2001/XMLSchema-instance"
               xsi:schemaLocation="urn:activemq:core ">

            <name>${POD_NAME}/${POD_IP}</name>


            <persistence-enabled>true</persistence-enabled>

            <!-- this could be ASYNCIO, MAPPED, NIO
               ASYNCIO: Linux Libaio
               MAPPED: mmap files
               NIO: Plain Java Files
            -->
            <journal-type>ASYNCIO</journal-type>

            <paging-directory>data/paging</paging-directory>

            <bindings-directory>data/bindings</bindings-directory>

            <journal-directory>data/journal</journal-directory>

            <large-messages-directory>data/large-messages</large-messages-directory>

            <journal-datasync>true</journal-datasync>

            <journal-min-files>2</journal-min-files>

            <journal-pool-files>10</journal-pool-files>

            <journal-device-block-size>4096</journal-device-block-size>

            <journal-file-size>10M</journal-file-size>
            
            <!--
            This value was determined through a calculation.
            Your system could perform 41.67 writes per millisecond
            on the current journal configuration.
            That translates as a sync write every 24000 nanoseconds.

            Note: If you specify 0 the system will perform writes directly to the disk.
                  We recommend this to be 0 if you are using journalType=MAPPED and journal-datasync=false.
            -->
            <journal-buffer-timeout>24000</journal-buffer-timeout>


            <!--
            When using ASYNCIO, this will determine the writing queue depth for libaio.
            -->
            <journal-max-io>4096</journal-max-io>
            <!--
            You can verify the network health of a particular NIC by specifying the <network-check-NIC> element.
               <network-check-NIC>theNicName</network-check-NIC>
            -->

            <!--
            Use this to use an HTTP server to validate the network
               <network-check-URL-list>http://www.apache.org</network-check-URL-list> -->

            <!-- <network-check-period>10000</network-check-period> -->
            <!-- <network-check-timeout>1000</network-check-timeout> -->

            <!-- this is a comma separated list, no spaces, just DNS or IPs
               it should accept IPV6

               Warning: Make sure you understand your network topology as this is meant to validate if your network is valid.
                        Using IPs that could eventually disappear or be partially visible may defeat the purpose.
                        You can use a list of multiple IPs, and if any successful ping will make the server OK to continue running -->
            <!-- <network-check-list>10.0.0.1</network-check-list> -->

            <!-- use this to customize the ping used for ipv4 addresses -->
            <!-- <network-check-ping-command>ping -c 1 -t %d %s</network-check-ping-command> -->

            <!-- use this to customize the ping used for ipv6 addresses -->
            <!-- <network-check-ping6-command>ping6 -c 1 %2$s</network-check-ping6-command> -->



         <connectors>
            <!-- Connector used to be announced through cluster connections and notifications -->
            <connector name="artemis">tcp://${POD_IP}:61616</connector>
         </connectors>



            <!-- how often we are looking for how many bytes are being used on the disk in ms -->
            <disk-scan-period>5000</disk-scan-period>

            <!-- once the disk hits this limit the system will block, or close the connection in certain protocols
               that won't support flow control. -->
            <max-disk-usage>90</max-disk-usage>

            <!-- should the broker detect dead locks and other issues -->
            <critical-analyzer>true</critical-analyzer>

            <critical-analyzer-timeout>120000</critical-analyzer-timeout>

            <critical-analyzer-check-period>60000</critical-analyzer-check-period>

            <critical-analyzer-policy>HALT</critical-analyzer-policy>

            
            <page-sync-timeout>168000</page-sync-timeout>


                  <!-- the system will enter into page mode once you hit this limit.
               This is an estimate in bytes of how much the messages are using in memory

                  The system will use half of the available memory (-Xmx) by default for the global-max-size.
                  You may specify a different value here if you need to customize it to your needs.

                  <global-max-size>100Mb</global-max-size>

            -->

            <acceptors>

               <!-- useEpoll means: it will use Netty epoll if you are on a system (Linux) that supports it -->
               <!-- amqpCredits: The number of credits sent to AMQP producers -->
               <!-- amqpLowCredits: The server will send the # credits specified at amqpCredits at this low mark -->
               <!-- amqpDuplicateDetection: If you are not using duplicate detection, set this to false
                                          as duplicate detection requires applicationProperties to be parsed on the server. -->
               <!-- amqpMinLargeMessageSize: Determines how many bytes are considered large, so we start using files to hold their data.
                                             default: 102400, -1 would mean to disable large mesasge control -->

               <!-- Note: If an acceptor needs to be compatible with HornetQ and/or Artemis 1.x clients add
                        "anycastPrefix=jms.queue.;multicastPrefix=jms.topic." to the acceptor url.
                        See https://issues.apache.org/jira/browse/ARTEMIS-1644 for more information. -->


               <!-- Acceptor for every supported protocol -->
               <acceptor name="artemis">tcp://0.0.0.0:61616?tcpSendBufferSize=1048576;tcpReceiveBufferSize=1048576;amqpMinLargeMessageSize=102400;protocols=CORE,AMQP,STOMP,HORNETQ,MQTT,OPENWIRE;useEpoll=true;amqpCredits=1000;amqpLowCredits=300;amqpDuplicateDetection=true</acceptor>

               <!-- AMQP Acceptor.  Listens on default AMQP port for AMQP traffic.-->
               <acceptor name="amqp">tcp://0.0.0.0:5672?tcpSendBufferSize=1048576;tcpReceiveBufferSize=1048576;protocols=AMQP;useEpoll=true;amqpCredits=1000;amqpLowCredits=300;amqpMinLargeMessageSize=102400;amqpDuplicateDetection=true</acceptor>

               <!-- STOMP Acceptor. -->
               <acceptor name="stomp">tcp://0.0.0.0:61613?tcpSendBufferSize=1048576;tcpReceiveBufferSize=1048576;protocols=STOMP;useEpoll=true</acceptor>

               <!-- HornetQ Compatibility Acceptor.  Enables HornetQ Core and STOMP for legacy HornetQ clients. -->
               <acceptor name="hornetq">tcp://0.0.0.0:5445?anycastPrefix=jms.queue.;multicastPrefix=jms.topic.;protocols=HORNETQ,STOMP;useEpoll=true</acceptor>

               <!-- MQTT Acceptor -->
               <acceptor name="mqtt">tcp://0.0.0.0:1883?tcpSendBufferSize=1048576;tcpReceiveBufferSize=1048576;protocols=MQTT;useEpoll=true</acceptor>

            </acceptors>

             <!-- IMPORTANT: MUST BE CHANGED-->
            <cluster-user>${CLUSTER_USER}</cluster-user>
            <cluster-password>${CLUSTER_PASSWORD}</cluster-password>

            <broadcast-groups>
               <broadcast-group name="bg-group1">
                  <group-address>${BROADCAST_IP}</group-address>
                  <group-port>${BROADCAST_PORT}</group-port>
                  <broadcast-period>5000</broadcast-period>
                  <connector-ref>artemis</connector-ref>
               </broadcast-group>
            </broadcast-groups>

            <discovery-groups>
               <discovery-group name="dg-group1">
                  <group-address>${BROADCAST_IP}</group-address>
                  <group-port>${BROADCAST_PORT}</group-port>
                  <refresh-timeout>10000</refresh-timeout>
               </discovery-group>
            </discovery-groups>

            <cluster-connections>
                <cluster-connection name="artemis">
                   <address></address>
                   <connector-ref>artemis</connector-ref>
                   <check-period>30000</check-period>
                   <connection-ttl>60000</connection-ttl>
                   <min-large-message-size>102400</min-large-message-size>
                   <call-timeout>30000</call-timeout>
                   <retry-interval>1000</retry-interval>
                   <retry-interval-multiplier>1.0</retry-interval-multiplier>
                   <max-retry-interval>2000</max-retry-interval>
                   <initial-connect-attempts>-1</initial-connect-attempts>
                   <reconnect-attempts>-1</reconnect-attempts>
                   <use-duplicate-detection>true</use-duplicate-detection>
                   <message-load-balancing>ON_DEMAND</message-load-balancing>
                   <max-hops>1</max-hops>
                   <confirmation-window-size>1048576</confirmation-window-size>
                   <call-failover-timeout>-1</call-failover-timeout>
                   <notification-interval>1000</notification-interval>
                   <notification-attempts>100</notification-attempts>
                   <discovery-group-ref discovery-group-name="dg-group1"/>
                </cluster-connection>
            </cluster-connections>


            <ha-policy>
               <replication>
                   <colocated>
                      <request-backup>true</request-backup>
                      <max-backups>1</max-backups>
                      <backup-request-retries>-1</backup-request-retries>
                      <backup-request-retry-interval>5000</backup-request-retry-interval>
                      
                      <master>
                         <vote-on-replication-failure>true</vote-on-replication-failure>
                         <check-for-live-server>true</check-for-live-server>
                      </master>
                      <slave>
                         <scale-down/>

                         <!-- when a master returns, let clients reconnect to the master again -->
                         <!- NEEDS TESTING 
                         <allow-failback>true</allow-failback>
                         <failback-delay>10000</failback-delay>
                         -->
                      </slave>
                   </colocated>
                </replication>
             </ha-policy>

            <security-settings>
               <security-setting match="#">
                  <permission type="createNonDurableQueue" roles="amq"/>
                  <permission type="deleteNonDurableQueue" roles="amq"/>
                  <permission type="createDurableQueue" roles="amq"/>
                  <permission type="deleteDurableQueue" roles="amq"/>
                  <permission type="createAddress" roles="amq"/>
                  <permission type="deleteAddress" roles="amq"/>
                  <permission type="consume" roles="amq"/>
                  <permission type="browse" roles="amq"/>
                  <permission type="send" roles="amq"/>
                  <!-- we need this otherwise ./artemis data imp wouldn't work -->
                  <permission type="manage" roles="amq"/>
               </security-setting>
            </security-settings>

            <address-settings>
               <!-- if you define auto-create on certain queues, management has to be auto-create -->
               <address-setting match="activemq.management#">
                  <dead-letter-address>DLQ</dead-letter-address>
                  <expiry-address>ExpiryQueue</expiry-address>
                  <redelivery-delay>0</redelivery-delay>
                  <!-- with -1 only the global-max-size is in use for limiting -->
                  <max-size-bytes>-1</max-size-bytes>
                  <message-counter-history-day-limit>10</message-counter-history-day-limit>
                  <address-full-policy>PAGE</address-full-policy>
                  <auto-create-queues>true</auto-create-queues>
                  <auto-create-addresses>true</auto-create-addresses>
               </address-setting>
               <!--default for catch all-->
               <address-setting match="#">
                  <dead-letter-address>DLQ</dead-letter-address>
                  <expiry-address>ExpiryQueue</expiry-address>
                  <redelivery-delay>0</redelivery-delay>
                  <!-- with -1 only the global-max-size is in use for limiting -->
                  <max-size-bytes>-1</max-size-bytes>
                  <message-counter-history-day-limit>10</message-counter-history-day-limit>
                  <address-full-policy>PAGE</address-full-policy>
                  <auto-create-queues>true</auto-create-queues>
                  <auto-create-addresses>true</auto-create-addresses>
                  <auto-create-jms-queues>true</auto-create-jms-queues>
                  <auto-create-jms-topics>true</auto-create-jms-topics>
                  
                  <!-- redestribute messages in a queue if there is no consumer connected to the current node -->
                  <redistribution-delay>0</redistribution-delay>

               </address-setting>
            </address-settings>

            <addresses>
               <address name="DLQ">
                  <anycast>
                     <queue name="DLQ" />
                  </anycast>
               </address>
               <address name="ExpiryQueue">
                  <anycast>
                     <queue name="ExpiryQueue" />
                  </anycast>
               </address>

            </addresses>


            <!-- Uncomment the following if you want to use the Standard LoggingActiveMQServerPlugin pluging to log in events
            <broker-plugins>
               <broker-plugin class-name="org.apache.activemq.artemis.core.server.plugin.impl.LoggingActiveMQServerPlugin">
                  <property key="LOG_ALL_EVENTS" value="true"/>
                  <property key="LOG_CONNECTION_EVENTS" value="true"/>
                  <property key="LOG_SESSION_EVENTS" value="true"/>
                  <property key="LOG_CONSUMER_EVENTS" value="true"/>
                  <property key="LOG_DELIVERING_EVENTS" value="true"/>
                  <property key="LOG_SENDING_EVENTS" value="true"/>
                  <property key="LOG_INTERNAL_EVENTS" value="true"/>
               </broker-plugin>
            </broker-plugins>
            -->

         </core>
      </configuration>
---
apiVersion: v1
kind: ReplicationController
metadata:
  name: artemis-rc
  namespace: artemis
spec:
  replicas: 1
  selector:
    app: artemis-rc
  template:
    metadata:
      name: artemis-test-pod
      namespace: artemis
      labels:
        app: artemis-rc
    spec:
      containers:
        - name: artemis
          image: argos-hub.ws.arvato-systems.de/asys-public/activemq-artemis:2.16.0
          # command: ['sh', '-c', 'sleep 36000']
          args: ["run", "--allow-kill", "--", "xml:/data/bootstrap.xml"]
          volumeMounts:
            - name: artemis-config
              subPath: bootstrap.xml
              mountPath: /data/bootstrap.xml
              readOnly: true
            - name: artemis-config
              subPath: broker.xml
              mountPath: /data/broker.xml
              readOnly: true
          env:
            - name: POD_NAME
              valueFrom:
                fieldRef:
                  fieldPath: metadata.name
            - name: POD_IP
              valueFrom:
                fieldRef:
                  fieldPath: status.podIP
            - name: BROADCAST_IP
              value: 231.7.7.7
            - name: BROADCAST_PORT
              value: "9876"
            - name: "EXTRA_ARGS"
              value: "" # extra creation args, to be precise
            - name: CLUSTER_USER
              value: cluster-admin
            - name: CLUSTER_PASSWORD
              value: password-admin
            - name: ARTEMIS_CLUSTER_PROPS
              value: |-
                -DPOD_NAME=$(POD_NAME) 
                -DPOD_IP=$(POD_IP) 
                -DBROADCAST_IP=$(BROADCAST_IP) 
                -DBROADCAST_PORT=$(BROADCAST_PORT)
                -DCLUSTER_USER=$(CLUSTER_USER) 
                -DCLUSTER_PASSWORD=$(CLUSTER_PASSWORD)
      volumes:
        - name: artemis-config
          configMap:
            defaultMode: 420
            name: artemis-config
---
apiVersion: v1
kind: Pod
metadata:
  name: artemis-main
  namespace: artemis
  labels:
    app: artemis-main
spec:
  containers:
    - name: artemis
      image: argos-hub.ws.arvato-systems.de/asys-public/activemq-artemis:2.16.0
      #command: ['sh', '-c', 'sleep 36000']
      args: ["run", "--allow-kill", "--", "xml:/data/bootstrap.xml"]
      volumeMounts:
        - name: artemis-config
          subPath: bootstrap.xml
          mountPath: /data/bootstrap.xml
          readOnly: true
        - name: artemis-config
          subPath: broker.xml
          mountPath: /data/broker.xml
          readOnly: true
      env:
        - name: POD_NAME
          valueFrom:
            fieldRef:
              fieldPath: metadata.name
        - name: POD_IP
          valueFrom:
            fieldRef:
              fieldPath: status.podIP
        - name: BROADCAST_IP
          value: 231.7.7.7
        - name: BROADCAST_PORT
          value: "9876"
        - name: "EXTRA_ARGS"
          value: "" # extra creation args, to be precise
        - name: CLUSTER_USER
          value: cluster-admin
        - name: CLUSTER_PASSWORD
          value: password-admin
        - name: ARTEMIS_CLUSTER_PROPS
          value: |-
            -DPOD_NAME=$(POD_NAME) 
            -DPOD_IP=$(POD_IP) 
            -DBROADCAST_IP=$(BROADCAST_IP) 
            -DBROADCAST_PORT=$(BROADCAST_PORT)
            -DCLUSTER_USER=$(CLUSTER_USER) 
            -DCLUSTER_PASSWORD=$(CLUSTER_PASSWORD)
  volumes:
    - name: artemis-config
      configMap:
        defaultMode: 420
        name: artemis-config
{code}
 

> Docker Multistage Build from Source
> -----------------------------------
>
>                 Key: ARTEMIS-3042
>                 URL: https://issues.apache.org/jira/browse/ARTEMIS-3042
>             Project: ActiveMQ Artemis
>          Issue Type: Improvement
>            Reporter: John Behm
>            Priority: Minor
>              Labels: docker,, dockerfile,
>
> It would be rather convenient to get people up and running with an easy to build or to setup Docker image that automatically builds the project from source, discards the build container and moves the necessary files over to the final container that can simply be started.
> The current docker image build is not really user firendly or convenient at all.
>  
> https://github.com/apache/activemq-artemis/tree/master/artemis-docker



--
This message was sent by Atlassian Jira
(v8.3.4#803005)