You are viewing a plain text version of this content. The canonical link for it is here.
Posted to users@cloudstack.apache.org by ilya <il...@gmail.com> on 2016/01/21 09:32:01 UTC

Re: cloudstack,swift

++ Wido - perhaps he seen it..
----
Hi Yuriy,

I'm going to switch to english as i'm posting this thread to "users"
mailing list

Sorry, i dont have enough experience with ceph + cloudstack.

When you ask on mailing lists, please clearly mention the problem - as
its not clear.

Id think its related to your configuration, since error states
> There is no secondary storage VM for secondary storage host Images

Do you have Secondary Storage VM running?

Regards
ilya

On 1/19/16 10:47 PM, Юрий Карпель wrote:
> Приветствую!
> 
> 
> Собрал тестовый стенд с cloudstack 4.7, kvm, ceph (centos7)
> 
> Кластер пока только для тестов, если коротко то сделал 2-а rgw на
> civetweb и сервер nfs + haproxy для s3,swift :
> 
> [client.rgw.srv-rgw01]
> rgw print continue = false
> host = srv-rgw01
> rgw frontends = civetweb port=8080
> rgw_socket_path = /tmp/radosgw.sock
> 
> [client.rgw.srv-rgw02]
> rgw print continue = false
> host = srv-rgw01
> rgw frontends = civetweb port=8080
> rgw_socket_path = /tmp/radosgw.sock
> 
> Проверяем:
> [ceph@ceph-adm ~]$ swift -A http://ceph-rgw.test.bst.ru:8080/auth/v1.0/
> -U cloudstack:swift -K 'KBDbLt3DJ9hhMCVuPDfX1TwtLVywa2NVtO6ODBnu' list
> images
> [ceph@ceph-adm ~]$
> [ceph@ceph-adm ~]$ swift upload images ceph.log
> ceph.log
> [ceph@ceph-adm ~]$ swift stat images
>                       Account: v1
>                     Container: images
>                       Objects: 1
>                         Bytes: 96355
>                      Read ACL:
>                     Write ACL:
>                       Sync To:
>                      Sync Key:
>                 Accept-Ranges: bytes
>              X-Storage-Policy: default-placement
> X-Container-Bytes-Used-Actual: 98304
>                   X-Timestamp: 1453193667.00000
>                    X-Trans-Id:
> tx000000000000000000208-00569f2d4c-4395-default
>                  Content-Type: text/plain; charset=utf-8
> [ceph@ceph-adm ~]$
> 
> 
> Добавляю в cloudstack:
> Name: Images
> Provider: Swift
> URL: http://ceph-rgw.test.bst.ru:8080/auth/v1.0/
> Account:cloudstack
> Username: swift
> Key: 
> 
> Лог:
> 2016-01-20 09:35:37,688 DEBUG [c.c.a.ApiServlet]
> (catalina-exec-20:ctx-72c3e20f ctx-320404d7) (logid:9e2762af) ===END===
>  192.168.7.29 -- GET
>  command=addImageStore&response=json&name=Images&provider=Swift&url=http%3A%2F%2Fceph-rgw.test.bst.ru
> <http://2Fceph-rgw.test.bst.ru>%3A8080%2Fauth%2Fv1.0%2F&details%5B0%5D.key=account&details%5B0%5D.value=cloudstack&details%5B1%5D.key=username&details%5B1%5D.value=swift&details%5B2%5D.key=key&details%5B2%5D.value=KBDbLt3DJ9hhMCVuPDfX1TwtLVywa2NVtO6ODBnu&_=1453271737545
> 2016-01-20 09:35:43,684 DEBUG [c.c.h.d.HostDaoImpl]
> (ClusteredAgentManager Timer:ctx-96868b17) (logid:73617e92) Resetting
> hosts suitable for reconnect
> 2016-01-20 09:35:43,689 DEBUG [c.c.h.d.HostDaoImpl]
> (ClusteredAgentManager Timer:ctx-96868b17) (logid:73617e92) Completed
> resetting hosts suitable for reconnect
> 2016-01-20 09:35:43,689 DEBUG [c.c.h.d.HostDaoImpl]
> (ClusteredAgentManager Timer:ctx-96868b17) (logid:73617e92) Acquiring
> hosts for clusters already owned by this management server
> 2016-01-20 09:35:43,690 DEBUG [c.c.h.d.HostDaoImpl]
> (ClusteredAgentManager Timer:ctx-96868b17) (logid:73617e92) Completed
> acquiring hosts for clusters already owned by this management server
> 2016-01-20 09:35:43,690 DEBUG [c.c.h.d.HostDaoImpl]
> (ClusteredAgentManager Timer:ctx-96868b17) (logid:73617e92) Acquiring
> hosts for clusters not owned by any management server
> 2016-01-20 09:35:43,692 DEBUG [c.c.h.d.HostDaoImpl]
> (ClusteredAgentManager Timer:ctx-96868b17) (logid:73617e92) Completed
> acquiring hosts for clusters not owned by any management server
> 2016-01-20 09:35:47,073 INFO  [o.a.c.f.j.i.AsyncJobManagerImpl]
> (AsyncJobMgr-Heartbeat-1:ctx-2c2a88d4) (logid:053e95d5) Begin cleanup
> expired async-jobs
> 2016-01-20 09:35:47,100 INFO  [o.a.c.f.j.i.AsyncJobManagerImpl]
> (AsyncJobMgr-Heartbeat-1:ctx-2c2a88d4) (logid:053e95d5) End cleanup
> expired async-jobs
> 2016-01-20 09:35:47,906 INFO  [o.a.c.s.r.NfsSecondaryStorageResource]
> (pool-56-thread-1:ctx-c962141a) (logid:e48a8f70) Determined host
> rgw-lb01.cloud.bstelecom.ru <http://rgw-lb01.cloud.bstelecom.ru>
> corresponds to IP 10.30.15.2
> 2016-01-20 09:35:47,906 DEBUG [o.a.c.s.r.NfsSecondaryStorageResource]
> (pool-56-thread-1:ctx-c962141a) (logid:e48a8f70) Mounting device with
> nfs-style path of 10.30.15.2:/nfs
> 2016-01-20 09:35:47,906 DEBUG [o.a.c.s.r.NfsSecondaryStorageResource]
> (pool-56-thread-1:ctx-c962141a) (logid:e48a8f70) making available
> /var/cloudstack/mnt/secStorage/c6c692a0-265d-3109-93d4-f0f65f524d84 on
> nfs://rgw-lb01.test.bst.ru/nfs <http://rgw-lb01.test.bst.ru/nfs>
> 2016-01-20 09:35:47,906 DEBUG [o.a.c.s.r.NfsSecondaryStorageResource]
> (pool-56-thread-1:ctx-c962141a) (logid:e48a8f70) local folder for mount
> will be /var/cloudstack/mnt/secStorage/c6c692a0-265d-3109-93d4-f0f65f524d84
> 2016-01-20 09:35:47,909 DEBUG [o.a.c.s.r.NfsSecondaryStorageResource]
> (pool-56-thread-1:ctx-c962141a) (logid:e48a8f70) Executing: sudo mount
> 2016-01-20 09:35:47,940 DEBUG [o.a.c.s.r.NfsSecondaryStorageResource]
> (pool-56-thread-1:ctx-c962141a) (logid:e48a8f70) Execution is successful.
> 2016-01-20 09:35:47,945 DEBUG [o.a.c.s.r.NfsSecondaryStorageResource]
> (pool-56-thread-1:ctx-c962141a) (logid:e48a8f70) Some device already
> mounted at
> /var/cloudstack/mnt/secStorage/c6c692a0-265d-3109-93d4-f0f65f524d84, no
> need to mount nfs://rgw-lb01test.bst.ru/nfs <http://rgw-lb01test.bst.ru/nfs>
> 2016-01-20 09:35:47,951 DEBUG [o.a.c.s.r.NfsSecondaryStorageResource]
> (pool-56-thread-1:ctx-c962141a) (logid:e48a8f70) Faild to get
> url:http://cloudstack.apt-get.eu/systemvm/4.6/systemvm64template-4.6.0-kvm.qcow2.bz2,
> due to java.io.IOException: access denied
> 
> 
> Как итог не загружены шаблоны systemvm:
>  [c.c.s.StatsCollector] (StatsCollector-4:ctx-4e2b563c) (logid:0523bff9)
> There is no secondary storage VM for secondary storage host Images
> 
> 
> 
> 
>  
> 
> 

Re: cloudstack,swift

Posted by Pierre-Luc Dion <pd...@cloudops.com>.
If  you are using Swift as Secondary Storage, there is still few bugs like
the CLOUDSTACK-9248 remaining. Syed pushed some fixes that are now in +4.7
if I'm correct.
We have some fix going on related to CLOUDSTACK-9248, PR should be in the
list soon as we complete all tests.

But if it's backed by Ceph, as Wido is saying, why use the Swift API
instead of S3 ?


PL


On Fri, Jan 22, 2016 at 6:08 AM, Wido den Hollander <wi...@widodh.nl> wrote:

> Why not use S3? The RADOS Gateway speaks S3 just fine.
>
> I've been using CloudStack with Ceph S3 without any problems.
>
> Wido
>
> On 21-01-16 23:06, ilya wrote:
> > You might find this relevant
> >
> > https://issues.apache.org/jira/browse/CLOUDSTACK-9248
> >
> >
> >
> > On 1/21/16 12:32 AM, ilya wrote:
> >> ++ Wido - perhaps he seen it..
> >> ----
> >> Hi Yuriy,
> >>
> >> I'm going to switch to english as i'm posting this thread to "users"
> >> mailing list
> >>
> >> Sorry, i dont have enough experience with ceph + cloudstack.
> >>
> >> When you ask on mailing lists, please clearly mention the problem - as
> >> its not clear.
> >>
> >> Id think its related to your configuration, since error states
> >>> There is no secondary storage VM for secondary storage host Images
> >>
> >> Do you have Secondary Storage VM running?
> >>
> >> Regards
> >> ilya
> >>
> >> On 1/19/16 10:47 PM, Юрий Карпель wrote:
> >>> Приветствую!
> >>>
> >>>
> >>> Собрал тестовый стенд с cloudstack 4.7, kvm, ceph (centos7)
> >>>
> >>> Кластер пока только для тестов, если коротко то сделал 2-а rgw на
> >>> civetweb и сервер nfs + haproxy для s3,swift :
> >>>
> >>> [client.rgw.srv-rgw01]
> >>> rgw print continue = false
> >>> host = srv-rgw01
> >>> rgw frontends = civetweb port=8080
> >>> rgw_socket_path = /tmp/radosgw.sock
> >>>
> >>> [client.rgw.srv-rgw02]
> >>> rgw print continue = false
> >>> host = srv-rgw01
> >>> rgw frontends = civetweb port=8080
> >>> rgw_socket_path = /tmp/radosgw.sock
> >>>
> >>> Проверяем:
> >>> [ceph@ceph-adm ~]$ swift -A
> http://ceph-rgw.test.bst.ru:8080/auth/v1.0/
> >>> -U cloudstack:swift -K 'KBDbLt3DJ9hhMCVuPDfX1TwtLVywa2NVtO6ODBnu' list
> >>> images
> >>> [ceph@ceph-adm ~]$
> >>> [ceph@ceph-adm ~]$ swift upload images ceph.log
> >>> ceph.log
> >>> [ceph@ceph-adm ~]$ swift stat images
> >>>                       Account: v1
> >>>                     Container: images
> >>>                       Objects: 1
> >>>                         Bytes: 96355
> >>>                      Read ACL:
> >>>                     Write ACL:
> >>>                       Sync To:
> >>>                      Sync Key:
> >>>                 Accept-Ranges: bytes
> >>>              X-Storage-Policy: default-placement
> >>> X-Container-Bytes-Used-Actual: 98304
> >>>                   X-Timestamp: 1453193667.00000
> >>>                    X-Trans-Id:
> >>> tx000000000000000000208-00569f2d4c-4395-default
> >>>                  Content-Type: text/plain; charset=utf-8
> >>> [ceph@ceph-adm ~]$
> >>>
> >>>
> >>> Добавляю в cloudstack:
> >>> Name: Images
> >>> Provider: Swift
> >>> URL: http://ceph-rgw.test.bst.ru:8080/auth/v1.0/
> >>> Account:cloudstack
> >>> Username: swift
> >>> Key:
> >>>
> >>> Лог:
> >>> 2016-01-20 09:35:37,688 DEBUG [c.c.a.ApiServlet]
> >>> (catalina-exec-20:ctx-72c3e20f ctx-320404d7) (logid:9e2762af) ===END===
> >>>  192.168.7.29 -- GET
> >>>
> command=addImageStore&response=json&name=Images&provider=Swift&url=http%3A%2F%
> 2Fceph-rgw.test.bst.ru
> >>> <http://2Fceph-rgw.test.bst.ru
> >%3A8080%2Fauth%2Fv1.0%2F&details%5B0%5D.key=account&details%5B0%5D.value=cloudstack&details%5B1%5D.key=username&details%5B1%5D.value=swift&details%5B2%5D.key=key&details%5B2%5D.value=KBDbLt3DJ9hhMCVuPDfX1TwtLVywa2NVtO6ODBnu&_=1453271737545
> >>> 2016-01-20 09:35:43,684 DEBUG [c.c.h.d.HostDaoImpl]
> >>> (ClusteredAgentManager Timer:ctx-96868b17) (logid:73617e92) Resetting
> >>> hosts suitable for reconnect
> >>> 2016-01-20 09:35:43,689 DEBUG [c.c.h.d.HostDaoImpl]
> >>> (ClusteredAgentManager Timer:ctx-96868b17) (logid:73617e92) Completed
> >>> resetting hosts suitable for reconnect
> >>> 2016-01-20 09:35:43,689 DEBUG [c.c.h.d.HostDaoImpl]
> >>> (ClusteredAgentManager Timer:ctx-96868b17) (logid:73617e92) Acquiring
> >>> hosts for clusters already owned by this management server
> >>> 2016-01-20 09:35:43,690 DEBUG [c.c.h.d.HostDaoImpl]
> >>> (ClusteredAgentManager Timer:ctx-96868b17) (logid:73617e92) Completed
> >>> acquiring hosts for clusters already owned by this management server
> >>> 2016-01-20 09:35:43,690 DEBUG [c.c.h.d.HostDaoImpl]
> >>> (ClusteredAgentManager Timer:ctx-96868b17) (logid:73617e92) Acquiring
> >>> hosts for clusters not owned by any management server
> >>> 2016-01-20 09:35:43,692 DEBUG [c.c.h.d.HostDaoImpl]
> >>> (ClusteredAgentManager Timer:ctx-96868b17) (logid:73617e92) Completed
> >>> acquiring hosts for clusters not owned by any management server
> >>> 2016-01-20 09:35:47,073 INFO  [o.a.c.f.j.i.AsyncJobManagerImpl]
> >>> (AsyncJobMgr-Heartbeat-1:ctx-2c2a88d4) (logid:053e95d5) Begin cleanup
> >>> expired async-jobs
> >>> 2016-01-20 09:35:47,100 INFO  [o.a.c.f.j.i.AsyncJobManagerImpl]
> >>> (AsyncJobMgr-Heartbeat-1:ctx-2c2a88d4) (logid:053e95d5) End cleanup
> >>> expired async-jobs
> >>> 2016-01-20 09:35:47,906 INFO  [o.a.c.s.r.NfsSecondaryStorageResource]
> >>> (pool-56-thread-1:ctx-c962141a) (logid:e48a8f70) Determined host
> >>> rgw-lb01.cloud.bstelecom.ru <http://rgw-lb01.cloud.bstelecom.ru>
> >>> corresponds to IP 10.30.15.2
> >>> 2016-01-20 09:35:47,906 DEBUG [o.a.c.s.r.NfsSecondaryStorageResource]
> >>> (pool-56-thread-1:ctx-c962141a) (logid:e48a8f70) Mounting device with
> >>> nfs-style path of 10.30.15.2:/nfs
> >>> 2016-01-20 09:35:47,906 DEBUG [o.a.c.s.r.NfsSecondaryStorageResource]
> >>> (pool-56-thread-1:ctx-c962141a) (logid:e48a8f70) making available
> >>> /var/cloudstack/mnt/secStorage/c6c692a0-265d-3109-93d4-f0f65f524d84 on
> >>> nfs://rgw-lb01.test.bst.ru/nfs <http://rgw-lb01.test.bst.ru/nfs>
> >>> 2016-01-20 09:35:47,906 DEBUG [o.a.c.s.r.NfsSecondaryStorageResource]
> >>> (pool-56-thread-1:ctx-c962141a) (logid:e48a8f70) local folder for mount
> >>> will be
> /var/cloudstack/mnt/secStorage/c6c692a0-265d-3109-93d4-f0f65f524d84
> >>> 2016-01-20 09:35:47,909 DEBUG [o.a.c.s.r.NfsSecondaryStorageResource]
> >>> (pool-56-thread-1:ctx-c962141a) (logid:e48a8f70) Executing: sudo mount
> >>> 2016-01-20 09:35:47,940 DEBUG [o.a.c.s.r.NfsSecondaryStorageResource]
> >>> (pool-56-thread-1:ctx-c962141a) (logid:e48a8f70) Execution is
> successful.
> >>> 2016-01-20 09:35:47,945 DEBUG [o.a.c.s.r.NfsSecondaryStorageResource]
> >>> (pool-56-thread-1:ctx-c962141a) (logid:e48a8f70) Some device already
> >>> mounted at
> >>> /var/cloudstack/mnt/secStorage/c6c692a0-265d-3109-93d4-f0f65f524d84, no
> >>> need to mount nfs://rgw-lb01test.bst.ru/nfs <
> http://rgw-lb01test.bst.ru/nfs>
> >>> 2016-01-20 09:35:47,951 DEBUG [o.a.c.s.r.NfsSecondaryStorageResource]
> >>> (pool-56-thread-1:ctx-c962141a) (logid:e48a8f70) Faild to get
> >>> url:
> http://cloudstack.apt-get.eu/systemvm/4.6/systemvm64template-4.6.0-kvm.qcow2.bz2
> ,
> >>> due to java.io.IOException: access denied
> >>>
> >>>
> >>> Как итог не загружены шаблоны systemvm:
> >>>  [c.c.s.StatsCollector] (StatsCollector-4:ctx-4e2b563c)
> (logid:0523bff9)
> >>> There is no secondary storage VM for secondary storage host Images
> >>>
> >>>
> >>>
> >>>
> >>>
> >>>
> >>>
>

Re: cloudstack,swift

Posted by Wido den Hollander <wi...@widodh.nl>.
Why not use S3? The RADOS Gateway speaks S3 just fine.

I've been using CloudStack with Ceph S3 without any problems.

Wido

On 21-01-16 23:06, ilya wrote:
> You might find this relevant
> 
> https://issues.apache.org/jira/browse/CLOUDSTACK-9248
> 
> 
> 
> On 1/21/16 12:32 AM, ilya wrote:
>> ++ Wido - perhaps he seen it..
>> ----
>> Hi Yuriy,
>>
>> I'm going to switch to english as i'm posting this thread to "users"
>> mailing list
>>
>> Sorry, i dont have enough experience with ceph + cloudstack.
>>
>> When you ask on mailing lists, please clearly mention the problem - as
>> its not clear.
>>
>> Id think its related to your configuration, since error states
>>> There is no secondary storage VM for secondary storage host Images
>>
>> Do you have Secondary Storage VM running?
>>
>> Regards
>> ilya
>>
>> On 1/19/16 10:47 PM, Юрий Карпель wrote:
>>> Приветствую!
>>>
>>>
>>> Собрал тестовый стенд с cloudstack 4.7, kvm, ceph (centos7)
>>>
>>> Кластер пока только для тестов, если коротко то сделал 2-а rgw на
>>> civetweb и сервер nfs + haproxy для s3,swift :
>>>
>>> [client.rgw.srv-rgw01]
>>> rgw print continue = false
>>> host = srv-rgw01
>>> rgw frontends = civetweb port=8080
>>> rgw_socket_path = /tmp/radosgw.sock
>>>
>>> [client.rgw.srv-rgw02]
>>> rgw print continue = false
>>> host = srv-rgw01
>>> rgw frontends = civetweb port=8080
>>> rgw_socket_path = /tmp/radosgw.sock
>>>
>>> Проверяем:
>>> [ceph@ceph-adm ~]$ swift -A http://ceph-rgw.test.bst.ru:8080/auth/v1.0/
>>> -U cloudstack:swift -K 'KBDbLt3DJ9hhMCVuPDfX1TwtLVywa2NVtO6ODBnu' list
>>> images
>>> [ceph@ceph-adm ~]$
>>> [ceph@ceph-adm ~]$ swift upload images ceph.log
>>> ceph.log
>>> [ceph@ceph-adm ~]$ swift stat images
>>>                       Account: v1
>>>                     Container: images
>>>                       Objects: 1
>>>                         Bytes: 96355
>>>                      Read ACL:
>>>                     Write ACL:
>>>                       Sync To:
>>>                      Sync Key:
>>>                 Accept-Ranges: bytes
>>>              X-Storage-Policy: default-placement
>>> X-Container-Bytes-Used-Actual: 98304
>>>                   X-Timestamp: 1453193667.00000
>>>                    X-Trans-Id:
>>> tx000000000000000000208-00569f2d4c-4395-default
>>>                  Content-Type: text/plain; charset=utf-8
>>> [ceph@ceph-adm ~]$
>>>
>>>
>>> Добавляю в cloudstack:
>>> Name: Images
>>> Provider: Swift
>>> URL: http://ceph-rgw.test.bst.ru:8080/auth/v1.0/
>>> Account:cloudstack
>>> Username: swift
>>> Key: 
>>>
>>> Лог:
>>> 2016-01-20 09:35:37,688 DEBUG [c.c.a.ApiServlet]
>>> (catalina-exec-20:ctx-72c3e20f ctx-320404d7) (logid:9e2762af) ===END===
>>>  192.168.7.29 -- GET
>>>  command=addImageStore&response=json&name=Images&provider=Swift&url=http%3A%2F%2Fceph-rgw.test.bst.ru
>>> <http://2Fceph-rgw.test.bst.ru>%3A8080%2Fauth%2Fv1.0%2F&details%5B0%5D.key=account&details%5B0%5D.value=cloudstack&details%5B1%5D.key=username&details%5B1%5D.value=swift&details%5B2%5D.key=key&details%5B2%5D.value=KBDbLt3DJ9hhMCVuPDfX1TwtLVywa2NVtO6ODBnu&_=1453271737545
>>> 2016-01-20 09:35:43,684 DEBUG [c.c.h.d.HostDaoImpl]
>>> (ClusteredAgentManager Timer:ctx-96868b17) (logid:73617e92) Resetting
>>> hosts suitable for reconnect
>>> 2016-01-20 09:35:43,689 DEBUG [c.c.h.d.HostDaoImpl]
>>> (ClusteredAgentManager Timer:ctx-96868b17) (logid:73617e92) Completed
>>> resetting hosts suitable for reconnect
>>> 2016-01-20 09:35:43,689 DEBUG [c.c.h.d.HostDaoImpl]
>>> (ClusteredAgentManager Timer:ctx-96868b17) (logid:73617e92) Acquiring
>>> hosts for clusters already owned by this management server
>>> 2016-01-20 09:35:43,690 DEBUG [c.c.h.d.HostDaoImpl]
>>> (ClusteredAgentManager Timer:ctx-96868b17) (logid:73617e92) Completed
>>> acquiring hosts for clusters already owned by this management server
>>> 2016-01-20 09:35:43,690 DEBUG [c.c.h.d.HostDaoImpl]
>>> (ClusteredAgentManager Timer:ctx-96868b17) (logid:73617e92) Acquiring
>>> hosts for clusters not owned by any management server
>>> 2016-01-20 09:35:43,692 DEBUG [c.c.h.d.HostDaoImpl]
>>> (ClusteredAgentManager Timer:ctx-96868b17) (logid:73617e92) Completed
>>> acquiring hosts for clusters not owned by any management server
>>> 2016-01-20 09:35:47,073 INFO  [o.a.c.f.j.i.AsyncJobManagerImpl]
>>> (AsyncJobMgr-Heartbeat-1:ctx-2c2a88d4) (logid:053e95d5) Begin cleanup
>>> expired async-jobs
>>> 2016-01-20 09:35:47,100 INFO  [o.a.c.f.j.i.AsyncJobManagerImpl]
>>> (AsyncJobMgr-Heartbeat-1:ctx-2c2a88d4) (logid:053e95d5) End cleanup
>>> expired async-jobs
>>> 2016-01-20 09:35:47,906 INFO  [o.a.c.s.r.NfsSecondaryStorageResource]
>>> (pool-56-thread-1:ctx-c962141a) (logid:e48a8f70) Determined host
>>> rgw-lb01.cloud.bstelecom.ru <http://rgw-lb01.cloud.bstelecom.ru>
>>> corresponds to IP 10.30.15.2
>>> 2016-01-20 09:35:47,906 DEBUG [o.a.c.s.r.NfsSecondaryStorageResource]
>>> (pool-56-thread-1:ctx-c962141a) (logid:e48a8f70) Mounting device with
>>> nfs-style path of 10.30.15.2:/nfs
>>> 2016-01-20 09:35:47,906 DEBUG [o.a.c.s.r.NfsSecondaryStorageResource]
>>> (pool-56-thread-1:ctx-c962141a) (logid:e48a8f70) making available
>>> /var/cloudstack/mnt/secStorage/c6c692a0-265d-3109-93d4-f0f65f524d84 on
>>> nfs://rgw-lb01.test.bst.ru/nfs <http://rgw-lb01.test.bst.ru/nfs>
>>> 2016-01-20 09:35:47,906 DEBUG [o.a.c.s.r.NfsSecondaryStorageResource]
>>> (pool-56-thread-1:ctx-c962141a) (logid:e48a8f70) local folder for mount
>>> will be /var/cloudstack/mnt/secStorage/c6c692a0-265d-3109-93d4-f0f65f524d84
>>> 2016-01-20 09:35:47,909 DEBUG [o.a.c.s.r.NfsSecondaryStorageResource]
>>> (pool-56-thread-1:ctx-c962141a) (logid:e48a8f70) Executing: sudo mount
>>> 2016-01-20 09:35:47,940 DEBUG [o.a.c.s.r.NfsSecondaryStorageResource]
>>> (pool-56-thread-1:ctx-c962141a) (logid:e48a8f70) Execution is successful.
>>> 2016-01-20 09:35:47,945 DEBUG [o.a.c.s.r.NfsSecondaryStorageResource]
>>> (pool-56-thread-1:ctx-c962141a) (logid:e48a8f70) Some device already
>>> mounted at
>>> /var/cloudstack/mnt/secStorage/c6c692a0-265d-3109-93d4-f0f65f524d84, no
>>> need to mount nfs://rgw-lb01test.bst.ru/nfs <http://rgw-lb01test.bst.ru/nfs>
>>> 2016-01-20 09:35:47,951 DEBUG [o.a.c.s.r.NfsSecondaryStorageResource]
>>> (pool-56-thread-1:ctx-c962141a) (logid:e48a8f70) Faild to get
>>> url:http://cloudstack.apt-get.eu/systemvm/4.6/systemvm64template-4.6.0-kvm.qcow2.bz2,
>>> due to java.io.IOException: access denied
>>>
>>>
>>> Как итог не загружены шаблоны systemvm:
>>>  [c.c.s.StatsCollector] (StatsCollector-4:ctx-4e2b563c) (logid:0523bff9)
>>> There is no secondary storage VM for secondary storage host Images
>>>
>>>
>>>
>>>
>>>  
>>>
>>>

Re: cloudstack,swift

Posted by ilya <il...@gmail.com>.
You might find this relevant

https://issues.apache.org/jira/browse/CLOUDSTACK-9248



On 1/21/16 12:32 AM, ilya wrote:
> ++ Wido - perhaps he seen it..
> ----
> Hi Yuriy,
> 
> I'm going to switch to english as i'm posting this thread to "users"
> mailing list
> 
> Sorry, i dont have enough experience with ceph + cloudstack.
> 
> When you ask on mailing lists, please clearly mention the problem - as
> its not clear.
> 
> Id think its related to your configuration, since error states
>> There is no secondary storage VM for secondary storage host Images
> 
> Do you have Secondary Storage VM running?
> 
> Regards
> ilya
> 
> On 1/19/16 10:47 PM, Юрий Карпель wrote:
>> Приветствую!
>>
>>
>> Собрал тестовый стенд с cloudstack 4.7, kvm, ceph (centos7)
>>
>> Кластер пока только для тестов, если коротко то сделал 2-а rgw на
>> civetweb и сервер nfs + haproxy для s3,swift :
>>
>> [client.rgw.srv-rgw01]
>> rgw print continue = false
>> host = srv-rgw01
>> rgw frontends = civetweb port=8080
>> rgw_socket_path = /tmp/radosgw.sock
>>
>> [client.rgw.srv-rgw02]
>> rgw print continue = false
>> host = srv-rgw01
>> rgw frontends = civetweb port=8080
>> rgw_socket_path = /tmp/radosgw.sock
>>
>> Проверяем:
>> [ceph@ceph-adm ~]$ swift -A http://ceph-rgw.test.bst.ru:8080/auth/v1.0/
>> -U cloudstack:swift -K 'KBDbLt3DJ9hhMCVuPDfX1TwtLVywa2NVtO6ODBnu' list
>> images
>> [ceph@ceph-adm ~]$
>> [ceph@ceph-adm ~]$ swift upload images ceph.log
>> ceph.log
>> [ceph@ceph-adm ~]$ swift stat images
>>                       Account: v1
>>                     Container: images
>>                       Objects: 1
>>                         Bytes: 96355
>>                      Read ACL:
>>                     Write ACL:
>>                       Sync To:
>>                      Sync Key:
>>                 Accept-Ranges: bytes
>>              X-Storage-Policy: default-placement
>> X-Container-Bytes-Used-Actual: 98304
>>                   X-Timestamp: 1453193667.00000
>>                    X-Trans-Id:
>> tx000000000000000000208-00569f2d4c-4395-default
>>                  Content-Type: text/plain; charset=utf-8
>> [ceph@ceph-adm ~]$
>>
>>
>> Добавляю в cloudstack:
>> Name: Images
>> Provider: Swift
>> URL: http://ceph-rgw.test.bst.ru:8080/auth/v1.0/
>> Account:cloudstack
>> Username: swift
>> Key: 
>>
>> Лог:
>> 2016-01-20 09:35:37,688 DEBUG [c.c.a.ApiServlet]
>> (catalina-exec-20:ctx-72c3e20f ctx-320404d7) (logid:9e2762af) ===END===
>>  192.168.7.29 -- GET
>>  command=addImageStore&response=json&name=Images&provider=Swift&url=http%3A%2F%2Fceph-rgw.test.bst.ru
>> <http://2Fceph-rgw.test.bst.ru>%3A8080%2Fauth%2Fv1.0%2F&details%5B0%5D.key=account&details%5B0%5D.value=cloudstack&details%5B1%5D.key=username&details%5B1%5D.value=swift&details%5B2%5D.key=key&details%5B2%5D.value=KBDbLt3DJ9hhMCVuPDfX1TwtLVywa2NVtO6ODBnu&_=1453271737545
>> 2016-01-20 09:35:43,684 DEBUG [c.c.h.d.HostDaoImpl]
>> (ClusteredAgentManager Timer:ctx-96868b17) (logid:73617e92) Resetting
>> hosts suitable for reconnect
>> 2016-01-20 09:35:43,689 DEBUG [c.c.h.d.HostDaoImpl]
>> (ClusteredAgentManager Timer:ctx-96868b17) (logid:73617e92) Completed
>> resetting hosts suitable for reconnect
>> 2016-01-20 09:35:43,689 DEBUG [c.c.h.d.HostDaoImpl]
>> (ClusteredAgentManager Timer:ctx-96868b17) (logid:73617e92) Acquiring
>> hosts for clusters already owned by this management server
>> 2016-01-20 09:35:43,690 DEBUG [c.c.h.d.HostDaoImpl]
>> (ClusteredAgentManager Timer:ctx-96868b17) (logid:73617e92) Completed
>> acquiring hosts for clusters already owned by this management server
>> 2016-01-20 09:35:43,690 DEBUG [c.c.h.d.HostDaoImpl]
>> (ClusteredAgentManager Timer:ctx-96868b17) (logid:73617e92) Acquiring
>> hosts for clusters not owned by any management server
>> 2016-01-20 09:35:43,692 DEBUG [c.c.h.d.HostDaoImpl]
>> (ClusteredAgentManager Timer:ctx-96868b17) (logid:73617e92) Completed
>> acquiring hosts for clusters not owned by any management server
>> 2016-01-20 09:35:47,073 INFO  [o.a.c.f.j.i.AsyncJobManagerImpl]
>> (AsyncJobMgr-Heartbeat-1:ctx-2c2a88d4) (logid:053e95d5) Begin cleanup
>> expired async-jobs
>> 2016-01-20 09:35:47,100 INFO  [o.a.c.f.j.i.AsyncJobManagerImpl]
>> (AsyncJobMgr-Heartbeat-1:ctx-2c2a88d4) (logid:053e95d5) End cleanup
>> expired async-jobs
>> 2016-01-20 09:35:47,906 INFO  [o.a.c.s.r.NfsSecondaryStorageResource]
>> (pool-56-thread-1:ctx-c962141a) (logid:e48a8f70) Determined host
>> rgw-lb01.cloud.bstelecom.ru <http://rgw-lb01.cloud.bstelecom.ru>
>> corresponds to IP 10.30.15.2
>> 2016-01-20 09:35:47,906 DEBUG [o.a.c.s.r.NfsSecondaryStorageResource]
>> (pool-56-thread-1:ctx-c962141a) (logid:e48a8f70) Mounting device with
>> nfs-style path of 10.30.15.2:/nfs
>> 2016-01-20 09:35:47,906 DEBUG [o.a.c.s.r.NfsSecondaryStorageResource]
>> (pool-56-thread-1:ctx-c962141a) (logid:e48a8f70) making available
>> /var/cloudstack/mnt/secStorage/c6c692a0-265d-3109-93d4-f0f65f524d84 on
>> nfs://rgw-lb01.test.bst.ru/nfs <http://rgw-lb01.test.bst.ru/nfs>
>> 2016-01-20 09:35:47,906 DEBUG [o.a.c.s.r.NfsSecondaryStorageResource]
>> (pool-56-thread-1:ctx-c962141a) (logid:e48a8f70) local folder for mount
>> will be /var/cloudstack/mnt/secStorage/c6c692a0-265d-3109-93d4-f0f65f524d84
>> 2016-01-20 09:35:47,909 DEBUG [o.a.c.s.r.NfsSecondaryStorageResource]
>> (pool-56-thread-1:ctx-c962141a) (logid:e48a8f70) Executing: sudo mount
>> 2016-01-20 09:35:47,940 DEBUG [o.a.c.s.r.NfsSecondaryStorageResource]
>> (pool-56-thread-1:ctx-c962141a) (logid:e48a8f70) Execution is successful.
>> 2016-01-20 09:35:47,945 DEBUG [o.a.c.s.r.NfsSecondaryStorageResource]
>> (pool-56-thread-1:ctx-c962141a) (logid:e48a8f70) Some device already
>> mounted at
>> /var/cloudstack/mnt/secStorage/c6c692a0-265d-3109-93d4-f0f65f524d84, no
>> need to mount nfs://rgw-lb01test.bst.ru/nfs <http://rgw-lb01test.bst.ru/nfs>
>> 2016-01-20 09:35:47,951 DEBUG [o.a.c.s.r.NfsSecondaryStorageResource]
>> (pool-56-thread-1:ctx-c962141a) (logid:e48a8f70) Faild to get
>> url:http://cloudstack.apt-get.eu/systemvm/4.6/systemvm64template-4.6.0-kvm.qcow2.bz2,
>> due to java.io.IOException: access denied
>>
>>
>> Как итог не загружены шаблоны systemvm:
>>  [c.c.s.StatsCollector] (StatsCollector-4:ctx-4e2b563c) (logid:0523bff9)
>> There is no secondary storage VM for secondary storage host Images
>>
>>
>>
>>
>>  
>>
>>