Login | Register For Free | Help
Search for: (Advanced)

Mailing List Archive: OpenStack: Operators

Shared storage HA question

 

 

First page Previous page 1 2 Next page Last page  View All OpenStack operators RSS feed   Index | Next | Previous | View Threaded


razique.mahroua at gmail

Jul 25, 2013, 2:34 AM

Post #26 of 35 (101 views)
Permalink
Re: Shared storage HA question [In reply to]

I'll try the boot from volume feature and see if it makes any significant difference. (fingers crossed) - I'm wondering though two things :

- How come I never had such issues (to validate) with MooseFS while it's the mfsmount is FUSE-based?
- Does that mean the issue is not si much about Gluster, rather all FUSE-based FS?

regards,

Razique Mahroua - Nuage & Co
razique.mahroua [at] gmail
Tel : +33 9 72 37 94 15



Le 25 juil. 2013 à 09:22, Sylvain Bauza <sylvain.bauza [at] bull> a écrit :

> Hi Denis,
>
> As per my short testings, I would assume anything but FUSE-mounted would match your needs.
>
> On the performance glance, here is what I could suggest :
> - if using GlusterFS, wait for this BP [1] to be implemented. I do agree with Razique on the issues you could face with GlusterFS, this is mainly due to the Windows caching system mixed with QCOW2 copy-on-write images relying on a FUSE mountpoint.
> - if using Ceph, use RADOS to boot from Cinder volumes. Don't use FUSE mountpoint, again.
>
> -Sylvain
>
> [1] https://blueprints.launchpad.net/nova/+spec/glusterfs-native-support
>
>
>
> Le 25/07/2013 08:16, Denis Loshakov a écrit :
>> So, first i'm going to try Ceph.
>> Thanks for advices and lets RTFM begin :)
>>
>> On 24.07.2013 23:18, Razique Mahroua wrote:
>>> +1 :)
>>>
>>>
>>> Le 24 juil. 2013 à 21:08, Joe Topjian <joe.topjian [at] cybera
>>> <mailto:joe.topjian [at] cybera>> a écrit :
>>>
>>>> Hi Jacob,
>>>>
>>>> Are you using SAS or SSD drives for Gluster? Also, do you have one
>>>> large Gluster volume across your entire cloud or is it broke up into a
>>>> few different ones? I've wondered if there's a benefit to doing the
>>>> latter so distribution activity is isolated to only a few nodes. The
>>>> downside to that, of course, is you're limited to what compute nodes
>>>> instances can migrate to.
>>>>
>>>> I use Gluster for instance storage in all of my "controlled"
>>>> environments like internal and sandbox clouds, but I'm hesitant to
>>>> introduce it into production environments as I've seen the same issues
>>>> that Razique describes -- especially with Windows instances. My guess
>>>> is due to how NTFS writes to disk.
>>>>
>>>> I'm curious if you could report the results of the following test: in
>>>> a Windows instance running on Gluster, copy a 3-4gb file to it from
>>>> the local network so it comes in at a very high speed. When I do this,
>>>> the first few gigs come in very fast, but then slows to a crawl and
>>>> the Gluster processes on all nodes spike.
>>>>
>>>> Thanks,
>>>> Joe
>>>>
>>>>
>>>>
>>>> On Wed, Jul 24, 2013 at 12:37 PM, Jacob Godin <jacobgodin [at] gmail
>>>> <mailto:jacobgodin [at] gmail>> wrote:
>>>>
>>>> Oh really, you've done away with Gluster all together? The fast
>>>> backbone is definitely needed, but I would think that was the case
>>>> with any distributed filesystem.
>>>>
>>>> MooseFS looks promising, but apparently it has a few reliability
>>>> problems.
>>>>
>>>>
>>>> On Wed, Jul 24, 2013 at 3:31 PM, Razique Mahroua
>>>> <razique.mahroua [at] gmail <mailto:razique.mahroua [at] gmail>> wrote:
>>>>
>>>> :-)
>>>> Actually I had to remove all my instances running on it
>>>> (especially the windows ones), yah unfortunately my network
>>>> backbone wasn't fast enough to support the load induced by GFS
>>>> - especially the numerous operations performed by the
>>>> self-healing agents :(
>>>>
>>>> I'm currently considering MooseFS, it has the advantage to
>>>> have a pretty long list of companies using it in production
>>>>
>>>> take care
>>>>
>>>>
>>>> Le 24 juil. 2013 à 16:40, Jacob Godin <jacobgodin [at] gmail
>>>> <mailto:jacobgodin [at] gmail>> a écrit :
>>>>
>>>>> A few things I found were key for I/O performance:
>>>>>
>>>>> 1. Make sure your network can sustain the traffic. We are
>>>>> using a 10G backbone with 2 bonded interfaces per node.
>>>>> 2. Use high speed drives. SATA will not cut it.
>>>>> 3. Look into tuning settings. Razique, thanks for sending
>>>>> these along to me a little while back. A couple that I
>>>>> found were useful:
>>>>> * KVM cache=writeback (a little risky, but WAY faster)
>>>>> * Gluster write-behind-window-size (set to 4MB in our
>>>>> setup)
>>>>> * Gluster cache-size (ideal values in our setup were
>>>>> 96MB-128MB)
>>>>>
>>>>> Hope that helps!
>>>>>
>>>>>
>>>>>
>>>>> On Wed, Jul 24, 2013 at 11:32 AM, Razique Mahroua
>>>>> <razique.mahroua [at] gmail
>>>>> <mailto:razique.mahroua [at] gmail>> wrote:
>>>>>
>>>>> I had much performance issues myself with Windows
>>>>> instances, and I/O demanding instances. Make sure it fits
>>>>> your env. first before deploying it in production
>>>>>
>>>>> Regards,
>>>>> Razique
>>>>>
>>>>> *Razique Mahroua** - **Nuage & Co*
>>>>> razique.mahroua [at] gmail <mailto:razique.mahroua [at] gmail>
>>>>> Tel : +33 9 72 37 94 15 <tel:%2B33%209%2072%2037%2094%2015>
>>>>>
>>>>> <NUAGECO-LOGO-Fblan_petit.jpg>
>>>>>
>>>>> Le 24 juil. 2013 à 16:25, Jacob Godin
>>>>> <jacobgodin [at] gmail <mailto:jacobgodin [at] gmail>> a
>>>>> écrit :
>>>>>
>>>>>> Hi Denis,
>>>>>>
>>>>>> I would take a look into GlusterFS with a distributed,
>>>>>> replicated volume. We have been using it for several
>>>>>> months now, and it has been stable. Nova will need to
>>>>>> have the volume mounted to its instances directory
>>>>>> (default /var/lib/nova/instances), and Cinder has direct
>>>>>> support for Gluster as of Grizzly I believe.
>>>>>>
>>>>>>
>>>>>>
>>>>>> On Wed, Jul 24, 2013 at 11:11 AM, Denis Loshakov
>>>>>> <dloshakov [at] gmail <mailto:dloshakov [at] gmail>> wrote:
>>>>>>
>>>>>> Hi all,
>>>>>>
>>>>>> I have issue with creating shared storage for
>>>>>> Openstack. Main idea is to create 100% redundant
>>>>>> shared storage from two servers (kind of network
>>>>>> RAID from two servers).
>>>>>> I have two identical servers with many disks inside.
>>>>>> What solution can any one provide for such schema? I
>>>>>> need shared storage for running VMs (so live
>>>>>> migration can work) and also for cinder-volumes.
>>>>>>
>>>>>> One solution is to install Linux on both servers and
>>>>>> use DRBD + OCFS2, any comments on this?
>>>>>> Also I heard about Quadstor software and it can
>>>>>> create network RAID and present it via iSCSI.
>>>>>>
>>>>>> Thanks.
>>>>>>
>>>>>> P.S. Glance uses swift and is setuped on another servers
>>>>>>
>>>>>> _________________________________________________
>>>>>> OpenStack-operators mailing list
>>>>>> OpenStack-operators [at] lists
>>>>>> <mailto:OpenStack-operators [at] lists>
>>>>>> http://lists.openstack.org/__cgi-bin/mailman/listinfo/__openstack-operators
>>>>>> <http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack-operators>
>>>>>>
>>>>>>
>>>>>> _______________________________________________
>>>>>> OpenStack-operators mailing list
>>>>>> OpenStack-operators [at] lists
>>>>>> <mailto:OpenStack-operators [at] lists>
>>>>>> http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack-operators
>>>>>
>>>>>
>>>>
>>>>
>>>>
>>>> _______________________________________________
>>>> OpenStack-operators mailing list
>>>> OpenStack-operators [at] lists
>>>> <mailto:OpenStack-operators [at] lists>
>>>> http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack-operators
>>>>
>>>>
>>>>
>>>>
>>>> --
>>>> Joe Topjian
>>>> Systems Architect
>>>> Cybera Inc.
>>>>
>>>> www.cybera.ca <http://www.cybera.ca/>
>>>>
>>>> Cybera is a not-for-profit organization that works to spur and support
>>>> innovation, for the economic benefit of Alberta, through the use
>>>> of cyberinfrastructure.
>>>
>>>
>>>
>>> _______________________________________________
>>> OpenStack-operators mailing list
>>> OpenStack-operators [at] lists
>>> http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack-operators
>>>
>>
>> _______________________________________________
>> OpenStack-operators mailing list
>> OpenStack-operators [at] lists
>> http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack-operators
>
Attachments: NUAGECO-LOGO-Fblan_petit.jpg (9.88 KB)


jacobgodin at gmail

Jul 25, 2013, 2:48 AM

Post #27 of 35 (101 views)
Permalink
Re: Shared storage HA question [In reply to]

Hi all,

Will try and get some time to do that Windows testing today. I haven't
attempted that one, have just used benchmarking tools.

Razique, I think the moosefs client does some local buffering? I'm not 100%
on that, but I know that it isn't truly synchronous. Moosefs doesn't do
direct writes and then wait for a response from each replica. Gluster does
this, making it slower, but technically more reliable.

Again, this is my understanding from reading about MFS, you would probably
know better than me :)
On Jul 25, 2013 6:39 AM, "Razique Mahroua" <razique.mahroua [at] gmail>
wrote:

> I'll try the boot from volume feature and see if it makes any significant
> difference. (fingers crossed) - I'm wondering though two things :
>
> - How come I never had such issues (to validate) with MooseFS while it's
> the mfsmount is FUSE-based?
> - Does that mean the issue is not si much about Gluster, rather all
> FUSE-based FS?
>
> regards,
>
> *Razique Mahroua** - **Nuage & Co*
> razique.mahroua [at] gmail
> Tel : +33 9 72 37 94 15
>
>
> Le 25 juil. 2013 à 09:22, Sylvain Bauza <sylvain.bauza [at] bull> a écrit :
>
> Hi Denis,
>
> As per my short testings, I would assume anything but FUSE-mounted would
> match your needs.
>
> On the performance glance, here is what I could suggest :
> - if using GlusterFS, wait for this BP [1] to be implemented. I do agree
> with Razique on the issues you could face with GlusterFS, this is mainly
> due to the Windows caching system mixed with QCOW2 copy-on-write images
> relying on a FUSE mountpoint.
> - if using Ceph, use RADOS to boot from Cinder volumes. Don't use FUSE
> mountpoint, again.
>
> -Sylvain
>
> [1] https://blueprints.launchpad.net/nova/+spec/glusterfs-native-support
>
>
>
> Le 25/07/2013 08:16, Denis Loshakov a écrit :
>
> So, first i'm going to try Ceph.
> Thanks for advices and lets RTFM begin :)
>
> On 24.07.2013 23:18, Razique Mahroua wrote:
>
> +1 :)
>
>
> Le 24 juil. 2013 à 21:08, Joe Topjian <joe.topjian [at] cybera
> <mailto:joe.topjian [at] cybera> <joe.topjian [at] cybera>> a écrit :
>
> Hi Jacob,
>
> Are you using SAS or SSD drives for Gluster? Also, do you have one
> large Gluster volume across your entire cloud or is it broke up into a
> few different ones? I've wondered if there's a benefit to doing the
> latter so distribution activity is isolated to only a few nodes. The
> downside to that, of course, is you're limited to what compute nodes
> instances can migrate to.
>
> I use Gluster for instance storage in all of my "controlled"
> environments like internal and sandbox clouds, but I'm hesitant to
> introduce it into production environments as I've seen the same issues
> that Razique describes -- especially with Windows instances. My guess
> is due to how NTFS writes to disk.
>
> I'm curious if you could report the results of the following test: in
> a Windows instance running on Gluster, copy a 3-4gb file to it from
> the local network so it comes in at a very high speed. When I do this,
> the first few gigs come in very fast, but then slows to a crawl and
> the Gluster processes on all nodes spike.
>
> Thanks,
> Joe
>
>
>
> On Wed, Jul 24, 2013 at 12:37 PM, Jacob Godin <jacobgodin [at] gmail
> <mailto:jacobgodin [at] gmail> <jacobgodin [at] gmail>> wrote:
>
> Oh really, you've done away with Gluster all together? The fast
> backbone is definitely needed, but I would think that was the case
> with any distributed filesystem.
>
> MooseFS looks promising, but apparently it has a few reliability
> problems.
>
>
> On Wed, Jul 24, 2013 at 3:31 PM, Razique Mahroua
> <razique.mahroua [at] gmail <mailto:razique.mahroua [at] gmail><razique.mahroua [at] gmail>>
> wrote:
>
> :-)
> Actually I had to remove all my instances running on it
> (especially the windows ones), yah unfortunately my network
> backbone wasn't fast enough to support the load induced by GFS
> - especially the numerous operations performed by the
> self-healing agents :(
>
> I'm currently considering MooseFS, it has the advantage to
> have a pretty long list of companies using it in production
>
> take care
>
>
> Le 24 juil. 2013 à 16:40, Jacob Godin <jacobgodin [at] gmail
> <mailto:jacobgodin [at] gmail> <jacobgodin [at] gmail>> a écrit :
>
> A few things I found were key for I/O performance:
>
> 1. Make sure your network can sustain the traffic. We are
> using a 10G backbone with 2 bonded interfaces per node.
> 2. Use high speed drives. SATA will not cut it.
> 3. Look into tuning settings. Razique, thanks for sending
> these along to me a little while back. A couple that I
> found were useful:
> * KVM cache=writeback (a little risky, but WAY faster)
> * Gluster write-behind-window-size (set to 4MB in our
> setup)
> * Gluster cache-size (ideal values in our setup were
> 96MB-128MB)
>
> Hope that helps!
>
>
>
> On Wed, Jul 24, 2013 at 11:32 AM, Razique Mahroua
> <razique.mahroua [at] gmail
> <mailto:razique.mahroua [at] gmail> <razique.mahroua [at] gmail>>
> wrote:
>
> I had much performance issues myself with Windows
> instances, and I/O demanding instances. Make sure it fits
> your env. first before deploying it in production
>
> Regards,
> Razique
>
> *Razique Mahroua** - **Nuage & Co*
> razique.mahroua [at] gmail <mailto:razique.mahroua [at] gmail><razique.mahroua [at] gmail>
> Tel : +33 9 72 37 94 15 <tel:%2B33%209%2072%2037%2094%2015>
>
> <NUAGECO-LOGO-Fblan_petit.jpg>
>
> Le 24 juil. 2013 à 16:25, Jacob Godin
> <jacobgodin [at] gmail <mailto:jacobgodin [at] gmail><jacobgodin [at] gmail>>
> a
> écrit :
>
> Hi Denis,
>
> I would take a look into GlusterFS with a distributed,
> replicated volume. We have been using it for several
> months now, and it has been stable. Nova will need to
> have the volume mounted to its instances directory
> (default /var/lib/nova/instances), and Cinder has direct
> support for Gluster as of Grizzly I believe.
>
>
>
> On Wed, Jul 24, 2013 at 11:11 AM, Denis Loshakov
> <dloshakov [at] gmail <mailto:dloshakov [at] gmail><dloshakov [at] gmail>>
> wrote:
>
> Hi all,
>
> I have issue with creating shared storage for
> Openstack. Main idea is to create 100% redundant
> shared storage from two servers (kind of network
> RAID from two servers).
> I have two identical servers with many disks inside.
> What solution can any one provide for such schema? I
> need shared storage for running VMs (so live
> migration can work) and also for cinder-volumes.
>
> One solution is to install Linux on both servers and
> use DRBD + OCFS2, any comments on this?
> Also I heard about Quadstor software and it can
> create network RAID and present it via iSCSI.
>
> Thanks.
>
> P.S. Glance uses swift and is setuped on another servers
>
> _________________________________________________
> OpenStack-operators mailing list
> OpenStack-operators [at] lists
> <mailto:OpenStack-operators [at] lists><OpenStack-operators [at] lists>
>
> http://lists.openstack.org/__cgi-bin/mailman/listinfo/__openstack-operators
>
> <http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack-operators><http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack-operators>
>
>
> _______________________________________________
> OpenStack-operators mailing list
> OpenStack-operators [at] lists
> <mailto:OpenStack-operators [at] lists><OpenStack-operators [at] lists>
>
> http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack-operators
>
>
>
>
>
>
> _______________________________________________
> OpenStack-operators mailing list
> OpenStack-operators [at] lists
> <mailto:OpenStack-operators [at] lists><OpenStack-operators [at] lists>
>
> http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack-operators
>
>
>
>
> --
> Joe Topjian
> Systems Architect
> Cybera Inc.
>
> www.cybera.ca <http://www.cybera.ca/> <http://www.cybera.ca/>
>
> Cybera is a not-for-profit organization that works to spur and support
> innovation, for the economic benefit of Alberta, through the use
> of cyberinfrastructure.
>
>
>
>
> _______________________________________________
> OpenStack-operators mailing list
> OpenStack-operators [at] lists
> http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack-operators
>
>
> _______________________________________________
> OpenStack-operators mailing list
> OpenStack-operators [at] lists
> http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack-operators
>
>
>
>
> _______________________________________________
> OpenStack-operators mailing list
> OpenStack-operators [at] lists
> http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack-operators
>
>
Attachments: NUAGECO-LOGO-Fblan_petit.jpg (9.88 KB)


sylvain.bauza at bull

Jul 25, 2013, 3:35 AM

Post #28 of 35 (101 views)
Permalink
Re: Shared storage HA question [In reply to]

Le 25/07/2013 11:34, Razique Mahroua a écrit :
>
> - How come I never had such issues (to validate) with MooseFS while
> it's the mfsmount is FUSE-based?

I'm not saying FUSE is killing perfs :-) My point was more likely : be
careful, if you mix up GlusterFS + Windows (caching system on C:) +
FUSE, then you *could* have something like race condition which could
affect perfs.

I haven't tested out MFS, happy if this is not reproducing, then :-)

> - Does that mean the issue is not si much about Gluster, rather all
> FUSE-based FS?
>

Indeed. As said though, CephFS is also not perfect. If people want to
choose Ceph, then it's better to boot from volume with Cinder-based RBD
(RADOS Block Device) driver.

-Sylvain

_______________________________________________
OpenStack-operators mailing list
OpenStack-operators [at] lists
http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack-operators


razique.mahroua at gmail

Jul 25, 2013, 6:15 AM

Post #29 of 35 (100 views)
Permalink
Re: Shared storage HA question [In reply to]

Good to know, that what I thought as well
thanks for the insight.
thanks Jabob as well, curious too see what your tests will bring up :)

Razique Mahroua - Nuage & Co
razique.mahroua [at] gmail
Tel : +33 9 72 37 94 15



Le 25 juil. 2013 à 12:35, Sylvain Bauza <sylvain.bauza [at] bull> a écrit :

> Le 25/07/2013 11:34, Razique Mahroua a écrit :
>>
>> - How come I never had such issues (to validate) with MooseFS while it's the mfsmount is FUSE-based?
>
> I'm not saying FUSE is killing perfs :-) My point was more likely : be careful, if you mix up GlusterFS + Windows (caching system on C:) + FUSE, then you *could* have something like race condition which could affect perfs.
>
> I haven't tested out MFS, happy if this is not reproducing, then :-)
>
>> - Does that mean the issue is not si much about Gluster, rather all FUSE-based FS?
>>
>
> Indeed. As said though, CephFS is also not perfect. If people want to choose Ceph, then it's better to boot from volume with Cinder-based RBD (RADOS Block Device) driver.
>
> -Sylvain
Attachments: NUAGECO-LOGO-Fblan_petit.jpg (9.88 KB)


jacobgodin at gmail

Jul 26, 2013, 4:09 AM

Post #30 of 35 (101 views)
Permalink
Re: Shared storage HA question [In reply to]

Hi Joe,

I'm using SAS drives currently. Gluster is configured as one volume, with
the thought being that the more spindles the better. If we were using SSDs,
this would probably be configured differently.

I performed a speed test on a Windows instance and write speeds seemed
consistent. Essentially, I spun up a Linux instance, made a 5GB file and
popped it into Apache. Then, went to my Windows instance (on the same
virtual network) and grabbed the file. It downloaded consistently between
240-300Mbps.


On Wed, Jul 24, 2013 at 4:08 PM, Joe Topjian <joe.topjian [at] cybera> wrote:

> Hi Jacob,
>
> Are you using SAS or SSD drives for Gluster? Also, do you have one large
> Gluster volume across your entire cloud or is it broke up into a few
> different ones? I've wondered if there's a benefit to doing the latter so
> distribution activity is isolated to only a few nodes. The downside to
> that, of course, is you're limited to what compute nodes instances can
> migrate to.
>
> I use Gluster for instance storage in all of my "controlled" environments
> like internal and sandbox clouds, but I'm hesitant to introduce it into
> production environments as I've seen the same issues that Razique describes
> -- especially with Windows instances. My guess is due to how NTFS writes to
> disk.
>
> I'm curious if you could report the results of the following test: in a
> Windows instance running on Gluster, copy a 3-4gb file to it from the local
> network so it comes in at a very high speed. When I do this, the first few
> gigs come in very fast, but then slows to a crawl and the Gluster processes
> on all nodes spike.
>
> Thanks,
> Joe
>
>
>
> On Wed, Jul 24, 2013 at 12:37 PM, Jacob Godin <jacobgodin [at] gmail>wrote:
>
>> Oh really, you've done away with Gluster all together? The fast backbone
>> is definitely needed, but I would think that was the case with any
>> distributed filesystem.
>>
>> MooseFS looks promising, but apparently it has a few reliability problems.
>>
>>
>> On Wed, Jul 24, 2013 at 3:31 PM, Razique Mahroua <
>> razique.mahroua [at] gmail> wrote:
>>
>>> :-)
>>> Actually I had to remove all my instances running on it (especially the
>>> windows ones), yah unfortunately my network backbone wasn't fast enough to
>>> support the load induced by GFS - especially the numerous operations
>>> performed by the self-healing agents :(
>>>
>>> I'm currently considering MooseFS, it has the advantage to have a pretty
>>> long list of companies using it in production
>>>
>>> take care
>>>
>>>
>>> Le 24 juil. 2013 à 16:40, Jacob Godin <jacobgodin [at] gmail> a écrit :
>>>
>>> A few things I found were key for I/O performance:
>>>
>>> 1. Make sure your network can sustain the traffic. We are using a
>>> 10G backbone with 2 bonded interfaces per node.
>>> 2. Use high speed drives. SATA will not cut it.
>>> 3. Look into tuning settings. Razique, thanks for sending these
>>> along to me a little while back. A couple that I found were useful:
>>> - KVM cache=writeback (a little risky, but WAY faster)
>>> - Gluster write-behind-window-size (set to 4MB in our setup)
>>> - Gluster cache-size (ideal values in our setup were 96MB-128MB)
>>>
>>> Hope that helps!
>>>
>>>
>>>
>>> On Wed, Jul 24, 2013 at 11:32 AM, Razique Mahroua <
>>> razique.mahroua [at] gmail> wrote:
>>>
>>>> I had much performance issues myself with Windows instances, and I/O
>>>> demanding instances. Make sure it fits your env. first before deploying it
>>>> in production
>>>>
>>>> Regards,
>>>> Razique
>>>>
>>>> *Razique Mahroua** - **Nuage & Co*
>>>> razique.mahroua [at] gmail
>>>> Tel : +33 9 72 37 94 15
>>>>
>>>> <NUAGECO-LOGO-Fblan_petit.jpg>
>>>>
>>>> Le 24 juil. 2013 à 16:25, Jacob Godin <jacobgodin [at] gmail> a écrit :
>>>>
>>>> Hi Denis,
>>>>
>>>> I would take a look into GlusterFS with a distributed, replicated
>>>> volume. We have been using it for several months now, and it has been
>>>> stable. Nova will need to have the volume mounted to its instances
>>>> directory (default /var/lib/nova/instances), and Cinder has direct support
>>>> for Gluster as of Grizzly I believe.
>>>>
>>>>
>>>>
>>>> On Wed, Jul 24, 2013 at 11:11 AM, Denis Loshakov <dloshakov [at] gmail>wrote:
>>>>
>>>>> Hi all,
>>>>>
>>>>> I have issue with creating shared storage for Openstack. Main idea is
>>>>> to create 100% redundant shared storage from two servers (kind of network
>>>>> RAID from two servers).
>>>>> I have two identical servers with many disks inside. What solution can
>>>>> any one provide for such schema? I need shared storage for running VMs (so
>>>>> live migration can work) and also for cinder-volumes.
>>>>>
>>>>> One solution is to install Linux on both servers and use DRBD + OCFS2,
>>>>> any comments on this?
>>>>> Also I heard about Quadstor software and it can create network RAID
>>>>> and present it via iSCSI.
>>>>>
>>>>> Thanks.
>>>>>
>>>>> P.S. Glance uses swift and is setuped on another servers
>>>>>
>>>>> ______________________________**_________________
>>>>> OpenStack-operators mailing list
>>>>> OpenStack-operators [at] lists**openstack.org<OpenStack-operators [at] lists>
>>>>> http://lists.openstack.org/**cgi-bin/mailman/listinfo/**
>>>>> openstack-operators<http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack-operators>
>>>>>
>>>>
>>>> _______________________________________________
>>>> OpenStack-operators mailing list
>>>> OpenStack-operators [at] lists
>>>> http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack-operators
>>>>
>>>>
>>>>
>>>
>>>
>>
>> _______________________________________________
>> OpenStack-operators mailing list
>> OpenStack-operators [at] lists
>> http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack-operators
>>
>>
>
>
> --
> Joe Topjian
> Systems Architect
> Cybera Inc.
>
> www.cybera.ca
>
> Cybera is a not-for-profit organization that works to spur and support
> innovation, for the economic benefit of Alberta, through the use
> of cyberinfrastructure.
>


song_tia at yahoo

Jul 28, 2013, 10:04 AM

Post #31 of 35 (96 views)
Permalink
Shared storage HA question [In reply to]

>> I have issue with creating shared storage for Openstack. Main idea is to
>> create 100% redundant shared storage from two servers (kind of network
>> RAID from two servers).
>> I have two identical servers with many disks inside. What solution can
>> any one provide for such schema? I need shared storage for running VMs
>> (so live migration can work) and also for cinder-volumes.

>> One solution is to install Linux on both servers and use DRBD + OCFS2,
>> any comments on this?
>> Also I heard about Quadstor software and it can create network RAID and
>> present it via iSCSI.

My two cents worth from a generic concept point view of a network mirror

Key factors to evaluate
1. Write latency
2. Active-active or use active-passive with automatic ip address failover (using CARP etc)
3. Split brain handling
4. How soon after a node failure are the two node brought in sync.

quadstor shines in many aspects and DRBD + SCST is a good alternative. You should evaluate both in my opinion.

-Tia

_______________________________________________
OpenStack-operators mailing list
OpenStack-operators [at] lists
http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack-operators


song_tia at yahoo

Jul 28, 2013, 12:06 PM

Post #32 of 35 (95 views)
Permalink
Re: Shared storage HA question [In reply to]

--------------------------------------------
On Sun, 7/28/13, Tia Song <song_tia [at] yahoo> wrote:

Subject: [Openstack-operators] Shared storage HA question
To: openstack-operators [at] lists
Date: Sunday, July 28, 2013, 10:04 AM

>> I have issue with creating
shared storage for Openstack. Main idea is to
>> create 100% redundant shared storage from two
servers (kind of network
>> RAID from two servers).
>> I have two identical servers with many disks
inside. What solution can
>> any one provide for such schema? I need shared
storage for running VMs
>> (so live migration can work) and also for
cinder-volumes.

>> One solution is to install Linux on both servers
and use DRBD + OCFS2,
>> any comments on this?
>> Also I heard about Quadstor software and it can
create network RAID and
>> present it via iSCSI.

>> My two cents worth from a generic concept point view of a
network mirror

>> Key factors to evaluate
>> 1. Write latency
>> 2. Active-active or use active-passive with automatic ip
address failover (using CARP etc)
>> 3. Split brain handling
>> 4. How soon after a node failure are the two node brought in
sync.

>> quadstor shines in many aspects and  DRBD + SCST is a
>> good alternative. You should evaluate both in my opinion.

Forgot one major aspect. Fencing.
1. How easy it is to setup fencing ?
2. Reliability. Definitely hardware fencing ILO, IPMI etc
3. Test, test, test. Simulate heavy IO, bring down the network between the two nodes etc.

-Tia




_______________________________________________
OpenStack-operators mailing list
OpenStack-operators [at] lists
http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack-operators


_______________________________________________
OpenStack-operators mailing list
OpenStack-operators [at] lists
http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack-operators


dloshakov at gmail

Jul 29, 2013, 3:42 AM

Post #33 of 35 (89 views)
Permalink
Re: Shared storage HA question [In reply to]

Thanks for TIP, I'll start with Ceph, then Quadstor. I think when I will
have some results, i'll post here.

On 28.07.2013 20:04, Tia Song wrote:
>>> I have issue with creating shared storage for Openstack. Main idea is to
>>> create 100% redundant shared storage from two servers (kind of network
>>> RAID from two servers).
>>> I have two identical servers with many disks inside. What solution can
>>> any one provide for such schema? I need shared storage for running VMs
>>> (so live migration can work) and also for cinder-volumes.
>
>>> One solution is to install Linux on both servers and use DRBD + OCFS2,
>>> any comments on this?
>>> Also I heard about Quadstor software and it can create network RAID and
>>> present it via iSCSI.
>
> My two cents worth from a generic concept point view of a network mirror
>
> Key factors to evaluate
> 1. Write latency
> 2. Active-active or use active-passive with automatic ip address failover (using CARP etc)
> 3. Split brain handling
> 4. How soon after a node failure are the two node brought in sync.
>
> quadstor shines in many aspects and DRBD + SCST is a good alternative. You should evaluate both in my opinion.
>
> -Tia
>
> _______________________________________________
> OpenStack-operators mailing list
> OpenStack-operators [at] lists
> http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack-operators
>

_______________________________________________
OpenStack-operators mailing list
OpenStack-operators [at] lists
http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack-operators


joe.topjian at cybera

Jul 29, 2013, 6:35 AM

Post #34 of 35 (87 views)
Permalink
Re: Shared storage HA question [In reply to]

Thanks, Jacob! Good to know that an all around high-speed Gluster
environment should handle spikes of activity.


On Fri, Jul 26, 2013 at 5:09 AM, Jacob Godin <jacobgodin [at] gmail> wrote:

> Hi Joe,
>
> I'm using SAS drives currently. Gluster is configured as one volume, with
> the thought being that the more spindles the better. If we were using SSDs,
> this would probably be configured differently.
>
> I performed a speed test on a Windows instance and write speeds seemed
> consistent. Essentially, I spun up a Linux instance, made a 5GB file and
> popped it into Apache. Then, went to my Windows instance (on the same
> virtual network) and grabbed the file. It downloaded consistently between
> 240-300Mbps.
>
>
> On Wed, Jul 24, 2013 at 4:08 PM, Joe Topjian <joe.topjian [at] cybera>wrote:
>
>> Hi Jacob,
>>
>> Are you using SAS or SSD drives for Gluster? Also, do you have one large
>> Gluster volume across your entire cloud or is it broke up into a few
>> different ones? I've wondered if there's a benefit to doing the latter so
>> distribution activity is isolated to only a few nodes. The downside to
>> that, of course, is you're limited to what compute nodes instances can
>> migrate to.
>>
>> I use Gluster for instance storage in all of my "controlled" environments
>> like internal and sandbox clouds, but I'm hesitant to introduce it into
>> production environments as I've seen the same issues that Razique describes
>> -- especially with Windows instances. My guess is due to how NTFS writes to
>> disk.
>>
>> I'm curious if you could report the results of the following test: in a
>> Windows instance running on Gluster, copy a 3-4gb file to it from the local
>> network so it comes in at a very high speed. When I do this, the first few
>> gigs come in very fast, but then slows to a crawl and the Gluster processes
>> on all nodes spike.
>>
>> Thanks,
>> Joe
>>
>>
>>
>> On Wed, Jul 24, 2013 at 12:37 PM, Jacob Godin <jacobgodin [at] gmail>wrote:
>>
>>> Oh really, you've done away with Gluster all together? The fast backbone
>>> is definitely needed, but I would think that was the case with any
>>> distributed filesystem.
>>>
>>> MooseFS looks promising, but apparently it has a few reliability
>>> problems.
>>>
>>>
>>> On Wed, Jul 24, 2013 at 3:31 PM, Razique Mahroua <
>>> razique.mahroua [at] gmail> wrote:
>>>
>>>> :-)
>>>> Actually I had to remove all my instances running on it (especially the
>>>> windows ones), yah unfortunately my network backbone wasn't fast enough to
>>>> support the load induced by GFS - especially the numerous operations
>>>> performed by the self-healing agents :(
>>>>
>>>> I'm currently considering MooseFS, it has the advantage to have a
>>>> pretty long list of companies using it in production
>>>>
>>>> take care
>>>>
>>>>
>>>> Le 24 juil. 2013 à 16:40, Jacob Godin <jacobgodin [at] gmail> a écrit :
>>>>
>>>> A few things I found were key for I/O performance:
>>>>
>>>> 1. Make sure your network can sustain the traffic. We are using a
>>>> 10G backbone with 2 bonded interfaces per node.
>>>> 2. Use high speed drives. SATA will not cut it.
>>>> 3. Look into tuning settings. Razique, thanks for sending these
>>>> along to me a little while back. A couple that I found were useful:
>>>> - KVM cache=writeback (a little risky, but WAY faster)
>>>> - Gluster write-behind-window-size (set to 4MB in our setup)
>>>> - Gluster cache-size (ideal values in our setup were 96MB-128MB)
>>>>
>>>> Hope that helps!
>>>>
>>>>
>>>>
>>>> On Wed, Jul 24, 2013 at 11:32 AM, Razique Mahroua <
>>>> razique.mahroua [at] gmail> wrote:
>>>>
>>>>> I had much performance issues myself with Windows instances, and I/O
>>>>> demanding instances. Make sure it fits your env. first before deploying it
>>>>> in production
>>>>>
>>>>> Regards,
>>>>> Razique
>>>>>
>>>>> *Razique Mahroua** - **Nuage & Co*
>>>>> razique.mahroua [at] gmail
>>>>> Tel : +33 9 72 37 94 15
>>>>>
>>>>> <NUAGECO-LOGO-Fblan_petit.jpg>
>>>>>
>>>>> Le 24 juil. 2013 à 16:25, Jacob Godin <jacobgodin [at] gmail> a écrit :
>>>>>
>>>>> Hi Denis,
>>>>>
>>>>> I would take a look into GlusterFS with a distributed, replicated
>>>>> volume. We have been using it for several months now, and it has been
>>>>> stable. Nova will need to have the volume mounted to its instances
>>>>> directory (default /var/lib/nova/instances), and Cinder has direct support
>>>>> for Gluster as of Grizzly I believe.
>>>>>
>>>>>
>>>>>
>>>>> On Wed, Jul 24, 2013 at 11:11 AM, Denis Loshakov <dloshakov [at] gmail>wrote:
>>>>>
>>>>>> Hi all,
>>>>>>
>>>>>> I have issue with creating shared storage for Openstack. Main idea is
>>>>>> to create 100% redundant shared storage from two servers (kind of network
>>>>>> RAID from two servers).
>>>>>> I have two identical servers with many disks inside. What solution
>>>>>> can any one provide for such schema? I need shared storage for running VMs
>>>>>> (so live migration can work) and also for cinder-volumes.
>>>>>>
>>>>>> One solution is to install Linux on both servers and use DRBD +
>>>>>> OCFS2, any comments on this?
>>>>>> Also I heard about Quadstor software and it can create network RAID
>>>>>> and present it via iSCSI.
>>>>>>
>>>>>> Thanks.
>>>>>>
>>>>>> P.S. Glance uses swift and is setuped on another servers
>>>>>>
>>>>>> ______________________________**_________________
>>>>>> OpenStack-operators mailing list
>>>>>> OpenStack-operators [at] lists**openstack.org<OpenStack-operators [at] lists>
>>>>>> http://lists.openstack.org/**cgi-bin/mailman/listinfo/**
>>>>>> openstack-operators<http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack-operators>
>>>>>>
>>>>>
>>>>> _______________________________________________
>>>>> OpenStack-operators mailing list
>>>>> OpenStack-operators [at] lists
>>>>> http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack-operators
>>>>>
>>>>>
>>>>>
>>>>
>>>>
>>>
>>> _______________________________________________
>>> OpenStack-operators mailing list
>>> OpenStack-operators [at] lists
>>> http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack-operators
>>>
>>>
>>
>>
>> --
>> Joe Topjian
>> Systems Architect
>> Cybera Inc.
>>
>> www.cybera.ca
>>
>> Cybera is a not-for-profit organization that works to spur and support
>> innovation, for the economic benefit of Alberta, through the use
>> of cyberinfrastructure.
>>
>
>


--
Joe Topjian
Systems Architect
Cybera Inc.

www.cybera.ca

Cybera is a not-for-profit organization that works to spur and support
innovation, for the economic benefit of Alberta, through the use
of cyberinfrastructure.


jacobgodin at gmail

Jul 29, 2013, 6:36 AM

Post #35 of 35 (87 views)
Permalink
Re: Shared storage HA question [In reply to]

Not a problem! It wasn't without its battles :P and I would still love to
see performance come up even more. Might start looking into SSDs..


On Mon, Jul 29, 2013 at 10:35 AM, Joe Topjian <joe.topjian [at] cybera> wrote:

> Thanks, Jacob! Good to know that an all around high-speed Gluster
> environment should handle spikes of activity.
>
>
> On Fri, Jul 26, 2013 at 5:09 AM, Jacob Godin <jacobgodin [at] gmail> wrote:
>
>> Hi Joe,
>>
>> I'm using SAS drives currently. Gluster is configured as one volume, with
>> the thought being that the more spindles the better. If we were using SSDs,
>> this would probably be configured differently.
>>
>> I performed a speed test on a Windows instance and write speeds seemed
>> consistent. Essentially, I spun up a Linux instance, made a 5GB file and
>> popped it into Apache. Then, went to my Windows instance (on the same
>> virtual network) and grabbed the file. It downloaded consistently between
>> 240-300Mbps.
>>
>>
>> On Wed, Jul 24, 2013 at 4:08 PM, Joe Topjian <joe.topjian [at] cybera>wrote:
>>
>>> Hi Jacob,
>>>
>>> Are you using SAS or SSD drives for Gluster? Also, do you have one large
>>> Gluster volume across your entire cloud or is it broke up into a few
>>> different ones? I've wondered if there's a benefit to doing the latter so
>>> distribution activity is isolated to only a few nodes. The downside to
>>> that, of course, is you're limited to what compute nodes instances can
>>> migrate to.
>>>
>>> I use Gluster for instance storage in all of my "controlled"
>>> environments like internal and sandbox clouds, but I'm hesitant to
>>> introduce it into production environments as I've seen the same issues that
>>> Razique describes -- especially with Windows instances. My guess is due to
>>> how NTFS writes to disk.
>>>
>>> I'm curious if you could report the results of the following test: in a
>>> Windows instance running on Gluster, copy a 3-4gb file to it from the local
>>> network so it comes in at a very high speed. When I do this, the first few
>>> gigs come in very fast, but then slows to a crawl and the Gluster processes
>>> on all nodes spike.
>>>
>>> Thanks,
>>> Joe
>>>
>>>
>>>
>>> On Wed, Jul 24, 2013 at 12:37 PM, Jacob Godin <jacobgodin [at] gmail>wrote:
>>>
>>>> Oh really, you've done away with Gluster all together? The fast
>>>> backbone is definitely needed, but I would think that was the case with any
>>>> distributed filesystem.
>>>>
>>>> MooseFS looks promising, but apparently it has a few reliability
>>>> problems.
>>>>
>>>>
>>>> On Wed, Jul 24, 2013 at 3:31 PM, Razique Mahroua <
>>>> razique.mahroua [at] gmail> wrote:
>>>>
>>>>> :-)
>>>>> Actually I had to remove all my instances running on it (especially
>>>>> the windows ones), yah unfortunately my network backbone wasn't fast enough
>>>>> to support the load induced by GFS - especially the numerous operations
>>>>> performed by the self-healing agents :(
>>>>>
>>>>> I'm currently considering MooseFS, it has the advantage to have a
>>>>> pretty long list of companies using it in production
>>>>>
>>>>> take care
>>>>>
>>>>>
>>>>> Le 24 juil. 2013 à 16:40, Jacob Godin <jacobgodin [at] gmail> a écrit :
>>>>>
>>>>> A few things I found were key for I/O performance:
>>>>>
>>>>> 1. Make sure your network can sustain the traffic. We are using a
>>>>> 10G backbone with 2 bonded interfaces per node.
>>>>> 2. Use high speed drives. SATA will not cut it.
>>>>> 3. Look into tuning settings. Razique, thanks for sending these
>>>>> along to me a little while back. A couple that I found were useful:
>>>>> - KVM cache=writeback (a little risky, but WAY faster)
>>>>> - Gluster write-behind-window-size (set to 4MB in our setup)
>>>>> - Gluster cache-size (ideal values in our setup were 96MB-128MB)
>>>>>
>>>>> Hope that helps!
>>>>>
>>>>>
>>>>>
>>>>> On Wed, Jul 24, 2013 at 11:32 AM, Razique Mahroua <
>>>>> razique.mahroua [at] gmail> wrote:
>>>>>
>>>>>> I had much performance issues myself with Windows instances, and I/O
>>>>>> demanding instances. Make sure it fits your env. first before deploying it
>>>>>> in production
>>>>>>
>>>>>> Regards,
>>>>>> Razique
>>>>>>
>>>>>> *Razique Mahroua** - **Nuage & Co*
>>>>>> razique.mahroua [at] gmail
>>>>>> Tel : +33 9 72 37 94 15
>>>>>>
>>>>>> <NUAGECO-LOGO-Fblan_petit.jpg>
>>>>>>
>>>>>> Le 24 juil. 2013 à 16:25, Jacob Godin <jacobgodin [at] gmail> a écrit
>>>>>> :
>>>>>>
>>>>>> Hi Denis,
>>>>>>
>>>>>> I would take a look into GlusterFS with a distributed, replicated
>>>>>> volume. We have been using it for several months now, and it has been
>>>>>> stable. Nova will need to have the volume mounted to its instances
>>>>>> directory (default /var/lib/nova/instances), and Cinder has direct support
>>>>>> for Gluster as of Grizzly I believe.
>>>>>>
>>>>>>
>>>>>>
>>>>>> On Wed, Jul 24, 2013 at 11:11 AM, Denis Loshakov <dloshakov [at] gmail
>>>>>> > wrote:
>>>>>>
>>>>>>> Hi all,
>>>>>>>
>>>>>>> I have issue with creating shared storage for Openstack. Main idea
>>>>>>> is to create 100% redundant shared storage from two servers (kind of
>>>>>>> network RAID from two servers).
>>>>>>> I have two identical servers with many disks inside. What solution
>>>>>>> can any one provide for such schema? I need shared storage for running VMs
>>>>>>> (so live migration can work) and also for cinder-volumes.
>>>>>>>
>>>>>>> One solution is to install Linux on both servers and use DRBD +
>>>>>>> OCFS2, any comments on this?
>>>>>>> Also I heard about Quadstor software and it can create network RAID
>>>>>>> and present it via iSCSI.
>>>>>>>
>>>>>>> Thanks.
>>>>>>>
>>>>>>> P.S. Glance uses swift and is setuped on another servers
>>>>>>>
>>>>>>> ______________________________**_________________
>>>>>>> OpenStack-operators mailing list
>>>>>>> OpenStack-operators [at] lists**openstack.org<OpenStack-operators [at] lists>
>>>>>>> http://lists.openstack.org/**cgi-bin/mailman/listinfo/**
>>>>>>> openstack-operators<http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack-operators>
>>>>>>>
>>>>>>
>>>>>> _______________________________________________
>>>>>> OpenStack-operators mailing list
>>>>>> OpenStack-operators [at] lists
>>>>>>
>>>>>> http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack-operators
>>>>>>
>>>>>>
>>>>>>
>>>>>
>>>>>
>>>>
>>>> _______________________________________________
>>>> OpenStack-operators mailing list
>>>> OpenStack-operators [at] lists
>>>> http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack-operators
>>>>
>>>>
>>>
>>>
>>> --
>>> Joe Topjian
>>> Systems Architect
>>> Cybera Inc.
>>>
>>> www.cybera.ca
>>>
>>> Cybera is a not-for-profit organization that works to spur and support
>>> innovation, for the economic benefit of Alberta, through the use
>>> of cyberinfrastructure.
>>>
>>
>>
>
>
> --
> Joe Topjian
> Systems Architect
> Cybera Inc.
>
> www.cybera.ca
>
> Cybera is a not-for-profit organization that works to spur and support
> innovation, for the economic benefit of Alberta, through the use
> of cyberinfrastructure.
>

First page Previous page 1 2 Next page Last page  View All OpenStack operators RSS feed   Index | Next | Previous | View Threaded
 
 


Interested in having your list archived? Contact Gossamer Threads
 
  Web Applications & Managed Hosting Powered by Gossamer Threads Inc.