
sylvain.bauza at bull
Jul 25, 2013, 12:22 AM
Post #25 of 35
(179 views)
Permalink
|
Hi Denis, As per my short testings, I would assume anything but FUSE-mounted would match your needs. On the performance glance, here is what I could suggest : - if using GlusterFS, wait for this BP [1] to be implemented. I do agree with Razique on the issues you could face with GlusterFS, this is mainly due to the Windows caching system mixed with QCOW2 copy-on-write images relying on a FUSE mountpoint. - if using Ceph, use RADOS to boot from Cinder volumes. Don't use FUSE mountpoint, again. -Sylvain [1] https://blueprints.launchpad.net/nova/+spec/glusterfs-native-support Le 25/07/2013 08:16, Denis Loshakov a écrit : > So, first i'm going to try Ceph. > Thanks for advices and lets RTFM begin :) > > On 24.07.2013 23:18, Razique Mahroua wrote: >> +1 :) >> >> >> Le 24 juil. 2013 à 21:08, Joe Topjian <joe.topjian [at] cybera >> <mailto:joe.topjian [at] cybera>> a écrit : >> >>> Hi Jacob, >>> >>> Are you using SAS or SSD drives for Gluster? Also, do you have one >>> large Gluster volume across your entire cloud or is it broke up into a >>> few different ones? I've wondered if there's a benefit to doing the >>> latter so distribution activity is isolated to only a few nodes. The >>> downside to that, of course, is you're limited to what compute nodes >>> instances can migrate to. >>> >>> I use Gluster for instance storage in all of my "controlled" >>> environments like internal and sandbox clouds, but I'm hesitant to >>> introduce it into production environments as I've seen the same issues >>> that Razique describes -- especially with Windows instances. My guess >>> is due to how NTFS writes to disk. >>> >>> I'm curious if you could report the results of the following test: in >>> a Windows instance running on Gluster, copy a 3-4gb file to it from >>> the local network so it comes in at a very high speed. When I do this, >>> the first few gigs come in very fast, but then slows to a crawl and >>> the Gluster processes on all nodes spike. >>> >>> Thanks, >>> Joe >>> >>> >>> >>> On Wed, Jul 24, 2013 at 12:37 PM, Jacob Godin <jacobgodin [at] gmail >>> <mailto:jacobgodin [at] gmail>> wrote: >>> >>> Oh really, you've done away with Gluster all together? The fast >>> backbone is definitely needed, but I would think that was the case >>> with any distributed filesystem. >>> >>> MooseFS looks promising, but apparently it has a few reliability >>> problems. >>> >>> >>> On Wed, Jul 24, 2013 at 3:31 PM, Razique Mahroua >>> <razique.mahroua [at] gmail <mailto:razique.mahroua [at] gmail>> >>> wrote: >>> >>> :-) >>> Actually I had to remove all my instances running on it >>> (especially the windows ones), yah unfortunately my network >>> backbone wasn't fast enough to support the load induced by GFS >>> - especially the numerous operations performed by the >>> self-healing agents :( >>> >>> I'm currently considering MooseFS, it has the advantage to >>> have a pretty long list of companies using it in production >>> >>> take care >>> >>> >>> Le 24 juil. 2013 à 16:40, Jacob Godin <jacobgodin [at] gmail >>> <mailto:jacobgodin [at] gmail>> a écrit : >>> >>>> A few things I found were key for I/O performance: >>>> >>>> 1. Make sure your network can sustain the traffic. We are >>>> using a 10G backbone with 2 bonded interfaces per node. >>>> 2. Use high speed drives. SATA will not cut it. >>>> 3. Look into tuning settings. Razique, thanks for sending >>>> these along to me a little while back. A couple that I >>>> found were useful: >>>> * KVM cache=writeback (a little risky, but WAY faster) >>>> * Gluster write-behind-window-size (set to 4MB in our >>>> setup) >>>> * Gluster cache-size (ideal values in our setup were >>>> 96MB-128MB) >>>> >>>> Hope that helps! >>>> >>>> >>>> >>>> On Wed, Jul 24, 2013 at 11:32 AM, Razique Mahroua >>>> <razique.mahroua [at] gmail >>>> <mailto:razique.mahroua [at] gmail>> wrote: >>>> >>>> I had much performance issues myself with Windows >>>> instances, and I/O demanding instances. Make sure it fits >>>> your env. first before deploying it in production >>>> >>>> Regards, >>>> Razique >>>> >>>> *Razique Mahroua** - **Nuage & Co* >>>> razique.mahroua [at] gmail >>>> <mailto:razique.mahroua [at] gmail> >>>> Tel : +33 9 72 37 94 15 >>>> <tel:%2B33%209%2072%2037%2094%2015> >>>> >>>> <NUAGECO-LOGO-Fblan_petit.jpg> >>>> >>>> Le 24 juil. 2013 à 16:25, Jacob Godin >>>> <jacobgodin [at] gmail <mailto:jacobgodin [at] gmail>> a >>>> écrit : >>>> >>>>> Hi Denis, >>>>> >>>>> I would take a look into GlusterFS with a distributed, >>>>> replicated volume. We have been using it for several >>>>> months now, and it has been stable. Nova will need to >>>>> have the volume mounted to its instances directory >>>>> (default /var/lib/nova/instances), and Cinder has direct >>>>> support for Gluster as of Grizzly I believe. >>>>> >>>>> >>>>> >>>>> On Wed, Jul 24, 2013 at 11:11 AM, Denis Loshakov >>>>> <dloshakov [at] gmail <mailto:dloshakov [at] gmail>> wrote: >>>>> >>>>> Hi all, >>>>> >>>>> I have issue with creating shared storage for >>>>> Openstack. Main idea is to create 100% redundant >>>>> shared storage from two servers (kind of network >>>>> RAID from two servers). >>>>> I have two identical servers with many disks inside. >>>>> What solution can any one provide for such schema? I >>>>> need shared storage for running VMs (so live >>>>> migration can work) and also for cinder-volumes. >>>>> >>>>> One solution is to install Linux on both servers and >>>>> use DRBD + OCFS2, any comments on this? >>>>> Also I heard about Quadstor software and it can >>>>> create network RAID and present it via iSCSI. >>>>> >>>>> Thanks. >>>>> >>>>> P.S. Glance uses swift and is setuped on another >>>>> servers >>>>> >>>>> _________________________________________________ >>>>> OpenStack-operators mailing list >>>>> OpenStack-operators [at] lists >>>>> <mailto:OpenStack-operators [at] lists> >>>>> http://lists.openstack.org/__cgi-bin/mailman/listinfo/__openstack-operators >>>>> <http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack-operators> >>>>> >>>>> >>>>> _______________________________________________ >>>>> OpenStack-operators mailing list >>>>> OpenStack-operators [at] lists >>>>> <mailto:OpenStack-operators [at] lists> >>>>> http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack-operators >>>>> >>>> >>>> >>> >>> >>> >>> _______________________________________________ >>> OpenStack-operators mailing list >>> OpenStack-operators [at] lists >>> <mailto:OpenStack-operators [at] lists> >>> http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack-operators >>> >>> >>> >>> >>> -- >>> Joe Topjian >>> Systems Architect >>> Cybera Inc. >>> >>> www.cybera.ca <http://www.cybera.ca/> >>> >>> Cybera is a not-for-profit organization that works to spur and support >>> innovation, for the economic benefit of Alberta, through the use >>> of cyberinfrastructure. >> >> >> >> _______________________________________________ >> OpenStack-operators mailing list >> OpenStack-operators [at] lists >> http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack-operators >> > > _______________________________________________ > OpenStack-operators mailing list > OpenStack-operators [at] lists > http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack-operators
|