Login | Register For Free | Help
Search for: (Advanced)

Mailing List Archive: DRBD: Users

NFS not starting with heartbeat

 

 

DRBD users RSS feed   Index | Next | Previous | View Threaded


marcel at kraan

May 16, 2012, 12:40 PM

Post #1 of 8 (1494 views)
Permalink
NFS not starting with heartbeat

Hello,

I use 2 servers with CentOS 6.2

But on 1 server (kvmstorage1) ifs is not starting after a restart or when i shutdown heartbeat and restart is later.

# this is my resources file? it start not all services i have given in the haresources file?
Do i something wrong?


#kvmstorage1
cat /etc/ha.d/haresources
kvmstorage1.localdomain IPaddr::192.168.123.209/24/eth0 drbddisk::main Filesystem::/dev/drbd0::/datastore::ext4 nfs nfslock rpcidmapd mysql

ResourceManager[2506]: 2012/05/16_21:34:40 info: Running /etc/ha.d/resource.d/Filesystem /dev/drbd0 /datastore ext4 start
Filesystem[2933]: 2012/05/16_21:34:40 INFO: Running start for /dev/drbd0 on /datastore
Filesystem[2921]: 2012/05/16_21:34:41 INFO: Success
ResourceManager[2506]: 2012/05/16_21:34:41 info: Running /etc/init.d/nfslock start
ResourceManager[2506]: 2012/05/16_21:34:41 info: Running /etc/init.d/rpcidmapd start
ResourceManager[2506]: 2012/05/16_21:34:42 info: Running /etc/init.d/mysqld start
May 16 21:34:43 kvmstorage1.localdomain heartbeat: [2489]: info: local HA resource acquisition completed (standby).
May 16 21:34:43 kvmstorage1.localdomain heartbeat: [1583]: info: Standby resource acquisition done [foreign].
May 16 21:34:43 kvmstorage1.localdomain heartbeat: [1583]: info: Initial resource acquisition complete (auto_failback)
May 16 21:34:43 kvmstorage1.localdomain heartbeat: [1583]: info: remote resource transition completed.


#kvmstorage2
cat /etc/ha.d/haresources
kvmstorage1.localdomain IPaddr::192.168.123.209/24/eth0 drbddisk::main Filesystem::/dev/drbd0::/datastore::ext4 nfs nfslock rpcidmapd mysql

Filesystem[16037]: 2012/05/16_21:33:53 INFO: Resource is stopped
ResourceManager[15787]: 2012/05/16_21:33:53 info: Running /etc/ha.d/resource.d/Filesystem /dev/drbd0 /datastore ext4 start
Filesystem[16117]: 2012/05/16_21:33:53 INFO: Running start for /dev/drbd0 on /datastore
Filesystem[16109]: 2012/05/16_21:33:53 INFO: Success
ResourceManager[15787]: 2012/05/16_21:33:54 info: Running /etc/init.d/nfs start
ResourceManager[15787]: 2012/05/16_21:33:54 info: Running /etc/init.d/nfslock start
ResourceManager[15787]: 2012/05/16_21:33:54 info: Running /etc/init.d/mysqld start
mach_down[15761]: 2012/05/16_21:33:56 info: /usr/share/heartbeat/mach_down: nice_failback: foreign resources acquired
May 16 21:33:56 kvmstorage2.localdomain heartbeat: [1528]: info: mach_down takeover complete.
mach_down[15761]: 2012/05/16_21:33:56 info: mach_down takeover complete for node kvmstorage1.localdomain.

_______________________________________________
drbd-user mailing list
drbd-user [at] lists
http://lists.linbit.com/mailman/listinfo/drbd-user


marcel at kraan

May 16, 2012, 12:50 PM

Post #2 of 8 (1406 views)
Permalink
Re: NFS not starting with heartbeat [In reply to]

I have a typo... ifs but be NFS

On 16 mei 2012, at 21:40, Marcel Kraan wrote:

> Hello,
>
> I use 2 servers with CentOS 6.2
>
> But on 1 server (kvmstorage1) ifs is not starting after a restart or when i shutdown heartbeat and restart is later.
>
> # this is my resources file? it start not all services i have given in the haresources file?
> Do i something wrong?
>
>
> #kvmstorage1
> cat /etc/ha.d/haresources
> kvmstorage1.localdomain IPaddr::192.168.123.209/24/eth0 drbddisk::main Filesystem::/dev/drbd0::/datastore::ext4 nfs nfslock rpcidmapd mysql
>
> ResourceManager[2506]: 2012/05/16_21:34:40 info: Running /etc/ha.d/resource.d/Filesystem /dev/drbd0 /datastore ext4 start
> Filesystem[2933]: 2012/05/16_21:34:40 INFO: Running start for /dev/drbd0 on /datastore
> Filesystem[2921]: 2012/05/16_21:34:41 INFO: Success
> ResourceManager[2506]: 2012/05/16_21:34:41 info: Running /etc/init.d/nfslock start
> ResourceManager[2506]: 2012/05/16_21:34:41 info: Running /etc/init.d/rpcidmapd start
> ResourceManager[2506]: 2012/05/16_21:34:42 info: Running /etc/init.d/mysqld start
> May 16 21:34:43 kvmstorage1.localdomain heartbeat: [2489]: info: local HA resource acquisition completed (standby).
> May 16 21:34:43 kvmstorage1.localdomain heartbeat: [1583]: info: Standby resource acquisition done [foreign].
> May 16 21:34:43 kvmstorage1.localdomain heartbeat: [1583]: info: Initial resource acquisition complete (auto_failback)
> May 16 21:34:43 kvmstorage1.localdomain heartbeat: [1583]: info: remote resource transition completed.
>
>
> #kvmstorage2
> cat /etc/ha.d/haresources
> kvmstorage1.localdomain IPaddr::192.168.123.209/24/eth0 drbddisk::main Filesystem::/dev/drbd0::/datastore::ext4 nfs nfslock rpcidmapd mysql
>
> Filesystem[16037]: 2012/05/16_21:33:53 INFO: Resource is stopped
> ResourceManager[15787]: 2012/05/16_21:33:53 info: Running /etc/ha.d/resource.d/Filesystem /dev/drbd0 /datastore ext4 start
> Filesystem[16117]: 2012/05/16_21:33:53 INFO: Running start for /dev/drbd0 on /datastore
> Filesystem[16109]: 2012/05/16_21:33:53 INFO: Success
> ResourceManager[15787]: 2012/05/16_21:33:54 info: Running /etc/init.d/nfs start
> ResourceManager[15787]: 2012/05/16_21:33:54 info: Running /etc/init.d/nfslock start
> ResourceManager[15787]: 2012/05/16_21:33:54 info: Running /etc/init.d/mysqld start
> mach_down[15761]: 2012/05/16_21:33:56 info: /usr/share/heartbeat/mach_down: nice_failback: foreign resources acquired
> May 16 21:33:56 kvmstorage2.localdomain heartbeat: [1528]: info: mach_down takeover complete.
> mach_down[15761]: 2012/05/16_21:33:56 info: mach_down takeover complete for node kvmstorage1.localdomain.
>
> _______________________________________________
> drbd-user mailing list
> drbd-user [at] lists
> http://lists.linbit.com/mailman/listinfo/drbd-user


lars.ellenberg at linbit

May 16, 2012, 1:05 PM

Post #3 of 8 (1410 views)
Permalink
Re: NFS not starting with heartbeat [In reply to]

On Wed, May 16, 2012 at 09:40:09PM +0200, Marcel Kraan wrote:
> Hello,
>
> I use 2 servers with CentOS 6.2
>
> But on 1 server (kvmstorage1) ifs is not starting after a restart or when i shutdown heartbeat and restart is later.
>
> # this is my resources file? it start not all services i have given in the haresources file?
> Do i something wrong?

The resource manager of heartbeat haresource mode is a shell script,
named ResourceManager.

It has a function acquireresourcegroup(),
which loops through all resources you defined,
runs "status" on them, and if that does not indicate they are running,
it runs "start" on them.

for j in "$@"
do
if
we_own_resource "$j" || doscript "$j" start
then
: $j start succeeded
else
...

There is no log message coming from "we_own_resource".
We could easily add one, of course, if you like that better.

So all is well, only the "nfs" thingy seemed to have been up "unexpectedly".

May I ask why you chose to use heartbeat haresource mode
instead of RHCS or Pacemaker, or any other potential candidate for the job?

Just curious here. I'm trying to figure out how common it is nowadays
to do new haresources deployments with new projects.


> #kvmstorage1
> cat /etc/ha.d/haresources
> kvmstorage1.localdomain IPaddr::192.168.123.209/24/eth0 drbddisk::main Filesystem::/dev/drbd0::/datastore::ext4 nfs nfslock rpcidmapd mysql
>
> ResourceManager[2506]: 2012/05/16_21:34:40 info: Running /etc/ha.d/resource.d/Filesystem /dev/drbd0 /datastore ext4 start
> Filesystem[2933]: 2012/05/16_21:34:40 INFO: Running start for /dev/drbd0 on /datastore
> Filesystem[2921]: 2012/05/16_21:34:41 INFO: Success
> ResourceManager[2506]: 2012/05/16_21:34:41 info: Running /etc/init.d/nfslock start
> ResourceManager[2506]: 2012/05/16_21:34:41 info: Running /etc/init.d/rpcidmapd start
> ResourceManager[2506]: 2012/05/16_21:34:42 info: Running /etc/init.d/mysqld start
> May 16 21:34:43 kvmstorage1.localdomain heartbeat: [2489]: info: local HA resource acquisition completed (standby).
> May 16 21:34:43 kvmstorage1.localdomain heartbeat: [1583]: info: Standby resource acquisition done [foreign].
> May 16 21:34:43 kvmstorage1.localdomain heartbeat: [1583]: info: Initial resource acquisition complete (auto_failback)
> May 16 21:34:43 kvmstorage1.localdomain heartbeat: [1583]: info: remote resource transition completed.
>
>
> #kvmstorage2
> cat /etc/ha.d/haresources
> kvmstorage1.localdomain IPaddr::192.168.123.209/24/eth0 drbddisk::main Filesystem::/dev/drbd0::/datastore::ext4 nfs nfslock rpcidmapd mysql
>
> Filesystem[16037]: 2012/05/16_21:33:53 INFO: Resource is stopped
> ResourceManager[15787]: 2012/05/16_21:33:53 info: Running /etc/ha.d/resource.d/Filesystem /dev/drbd0 /datastore ext4 start
> Filesystem[16117]: 2012/05/16_21:33:53 INFO: Running start for /dev/drbd0 on /datastore
> Filesystem[16109]: 2012/05/16_21:33:53 INFO: Success
> ResourceManager[15787]: 2012/05/16_21:33:54 info: Running /etc/init.d/nfs start
> ResourceManager[15787]: 2012/05/16_21:33:54 info: Running /etc/init.d/nfslock start
> ResourceManager[15787]: 2012/05/16_21:33:54 info: Running /etc/init.d/mysqld start
> mach_down[15761]: 2012/05/16_21:33:56 info: /usr/share/heartbeat/mach_down: nice_failback: foreign resources acquired
> May 16 21:33:56 kvmstorage2.localdomain heartbeat: [1528]: info: mach_down takeover complete.
> mach_down[15761]: 2012/05/16_21:33:56 info: mach_down takeover complete for node kvmstorage1.localdomain.

--
: Lars Ellenberg
: LINBIT | Your Way to High Availability
: DRBD/HA support and consulting http://www.linbit.com
_______________________________________________
drbd-user mailing list
drbd-user [at] lists
http://lists.linbit.com/mailman/listinfo/drbd-user


marcel at kraan

May 16, 2012, 1:22 PM

Post #4 of 8 (1409 views)
Permalink
Re: NFS not starting with heartbeat [In reply to]

Hello Lars,

Thanks for the response.
I use heartbeat because i used it years ago. (i don't know better)
So i did not know that PaceMaker or RHCS did the same job.

If PaceMaker is a more up2date choice, then i will use this.

marcel


On 16 mei 2012, at 22:05, Lars Ellenberg wrote:

> On Wed, May 16, 2012 at 09:40:09PM +0200, Marcel Kraan wrote:
>> Hello,
>>
>> I use 2 servers with CentOS 6.2
>>
>> But on 1 server (kvmstorage1) ifs is not starting after a restart or when i shutdown heartbeat and restart is later.
>>
>> # this is my resources file? it start not all services i have given in the haresources file?
>> Do i something wrong?
>
> The resource manager of heartbeat haresource mode is a shell script,
> named ResourceManager.
>
> It has a function acquireresourcegroup(),
> which loops through all resources you defined,
> runs "status" on them, and if that does not indicate they are running,
> it runs "start" on them.
>
> for j in "$@"
> do
> if
> we_own_resource "$j" || doscript "$j" start
> then
> : $j start succeeded
> else
> ...
>
> There is no log message coming from "we_own_resource".
> We could easily add one, of course, if you like that better.
>
> So all is well, only the "nfs" thingy seemed to have been up "unexpectedly".
>
> May I ask why you chose to use heartbeat haresource mode
> instead of RHCS or Pacemaker, or any other potential candidate for the job?
>
> Just curious here. I'm trying to figure out how common it is nowadays
> to do new haresources deployments with new projects.
>
>
>> #kvmstorage1
>> cat /etc/ha.d/haresources
>> kvmstorage1.localdomain IPaddr::192.168.123.209/24/eth0 drbddisk::main Filesystem::/dev/drbd0::/datastore::ext4 nfs nfslock rpcidmapd mysql
>>
>> ResourceManager[2506]: 2012/05/16_21:34:40 info: Running /etc/ha.d/resource.d/Filesystem /dev/drbd0 /datastore ext4 start
>> Filesystem[2933]: 2012/05/16_21:34:40 INFO: Running start for /dev/drbd0 on /datastore
>> Filesystem[2921]: 2012/05/16_21:34:41 INFO: Success
>> ResourceManager[2506]: 2012/05/16_21:34:41 info: Running /etc/init.d/nfslock start
>> ResourceManager[2506]: 2012/05/16_21:34:41 info: Running /etc/init.d/rpcidmapd start
>> ResourceManager[2506]: 2012/05/16_21:34:42 info: Running /etc/init.d/mysqld start
>> May 16 21:34:43 kvmstorage1.localdomain heartbeat: [2489]: info: local HA resource acquisition completed (standby).
>> May 16 21:34:43 kvmstorage1.localdomain heartbeat: [1583]: info: Standby resource acquisition done [foreign].
>> May 16 21:34:43 kvmstorage1.localdomain heartbeat: [1583]: info: Initial resource acquisition complete (auto_failback)
>> May 16 21:34:43 kvmstorage1.localdomain heartbeat: [1583]: info: remote resource transition completed.
>>
>>
>> #kvmstorage2
>> cat /etc/ha.d/haresources
>> kvmstorage1.localdomain IPaddr::192.168.123.209/24/eth0 drbddisk::main Filesystem::/dev/drbd0::/datastore::ext4 nfs nfslock rpcidmapd mysql
>>
>> Filesystem[16037]: 2012/05/16_21:33:53 INFO: Resource is stopped
>> ResourceManager[15787]: 2012/05/16_21:33:53 info: Running /etc/ha.d/resource.d/Filesystem /dev/drbd0 /datastore ext4 start
>> Filesystem[16117]: 2012/05/16_21:33:53 INFO: Running start for /dev/drbd0 on /datastore
>> Filesystem[16109]: 2012/05/16_21:33:53 INFO: Success
>> ResourceManager[15787]: 2012/05/16_21:33:54 info: Running /etc/init.d/nfs start
>> ResourceManager[15787]: 2012/05/16_21:33:54 info: Running /etc/init.d/nfslock start
>> ResourceManager[15787]: 2012/05/16_21:33:54 info: Running /etc/init.d/mysqld start
>> mach_down[15761]: 2012/05/16_21:33:56 info: /usr/share/heartbeat/mach_down: nice_failback: foreign resources acquired
>> May 16 21:33:56 kvmstorage2.localdomain heartbeat: [1528]: info: mach_down takeover complete.
>> mach_down[15761]: 2012/05/16_21:33:56 info: mach_down takeover complete for node kvmstorage1.localdomain.
>
> --
> : Lars Ellenberg
> : LINBIT | Your Way to High Availability
> : DRBD/HA support and consulting http://www.linbit.com
> _______________________________________________
> drbd-user mailing list
> drbd-user [at] lists
> http://lists.linbit.com/mailman/listinfo/drbd-user

_______________________________________________
drbd-user mailing list
drbd-user [at] lists
http://lists.linbit.com/mailman/listinfo/drbd-user


danceswithcrows at usa

May 16, 2012, 1:36 PM

Post #5 of 8 (1411 views)
Permalink
Re: NFS not starting with heartbeat [In reply to]

From: Lars Ellenberg <lars.ellenberg [at] linbit>
[snippage]
> May I ask why you chose to use heartbeat haresource mode instead of
> RHCS or Pacemaker, or any other potential candidate for the job?
>
> Just curious here. I'm trying to figure out how common it is nowadays
> to do new haresources deployments with new projects.

The last time I tried pacemaker on a test cluster at work, it didn't work at
all. This was apparently because multicast was required, and the people who
run networking at work said they wouldn't turn multicast on for any of the
switches. heartbeat, though, works reasonably well because it doesn't require
multicast.

FWIW, Marcel's haresources had "nfs nfslock", while a heartbeat-using NFS
cluster here has "nfslock nfs" in the same place. The /etc/init.d/ scripts
have chkconfig comments showing nfslock starting before nfs. So that's
another thing for Marcel to look at.

--
Matt G / Dances With Crows
The Crow202 Blog: http://crow202.org/wordpress/
There is no Darkness in Eternity/But only Light too dim for us to see

_______________________________________________
drbd-user mailing list
drbd-user [at] lists
http://lists.linbit.com/mailman/listinfo/drbd-user


florian at hastexo

May 16, 2012, 1:42 PM

Post #6 of 8 (1407 views)
Permalink
Re: NFS not starting with heartbeat [In reply to]

On Wed, May 16, 2012 at 10:36 PM, Matt Graham <danceswithcrows [at] usa> wrote:
> From: Lars Ellenberg <lars.ellenberg [at] linbit>
> [snippage]
>> May I ask why you chose to use heartbeat haresource mode instead of
>> RHCS or Pacemaker, or any other potential candidate for the job?
>>
>> Just curious here. I'm trying to figure out how common it is nowadays
>> to do new haresources deployments with new projects.
>
> The last time I tried pacemaker on a test cluster at work, it didn't work at
> all.  This was apparently because multicast was required, and the people who
> run networking at work said they wouldn't turn multicast on for any of the
> switches.  heartbeat, though, works reasonably well because it doesn't require
> multicast.

You can run Pacemaker on the Heartbeat CCM which has supported unicast
for ages, and Corosync supports UDP unicast since release 1.3.0
(something like a year and a half ago).

Florian

--
Need help with High Availability?
http://www.hastexo.com/now
_______________________________________________
drbd-user mailing list
drbd-user [at] lists
http://lists.linbit.com/mailman/listinfo/drbd-user


marcel at kraan

May 16, 2012, 1:48 PM

Post #7 of 8 (1432 views)
Permalink
Re: NFS not starting with heartbeat [In reply to]

Hello Matt, Lars,

While installing PaceMaker the "nfslock nfs" solution frm Matt worked when i restart heartbeat
when i reboot the server the ifs is still down? (very strange)

root [at] kvmstorage ~]# ifconfig
eth0 Link encap:Ethernet HWaddr 52:54:00:18:BC:3D
inet addr:192.168.123.211 Bcast:192.168.123.255 Mask:255.255.255.0
inet6 addr: fe80::5054:ff:fe18:bc3d/64 Scope:Link
UP BROADCAST RUNNING MULTICAST MTU:1500 Metric:1
RX packets:92869 errors:0 dropped:15 overruns:0 frame:0
TX packets:97696 errors:0 dropped:0 overruns:0 carrier:0
collisions:0 txqueuelen:1000
RX bytes:113420015 (108.1 MiB) TX bytes:114931967 (109.6 MiB)
Interrupt:10 Base address:0x2000

eth0:0 Link encap:Ethernet HWaddr 52:54:00:18:BC:3D
inet addr:192.168.123.209 Bcast:192.168.123.255 Mask:255.255.255.0
UP BROADCAST RUNNING MULTICAST MTU:1500 Metric:1
Interrupt:10 Base address:0x2000

lo Link encap:Local Loopback
inet addr:127.0.0.1 Mask:255.0.0.0
inet6 addr: ::1/128 Scope:Host
UP LOOPBACK RUNNING MTU:16436 Metric:1
RX packets:12 errors:0 dropped:0 overruns:0 frame:0
TX packets:12 errors:0 dropped:0 overruns:0 carrier:0
collisions:0 txqueuelen:0
RX bytes:840 (840.0 b) TX bytes:840 (840.0 b)

[root [at] kvmstorage ~]# service nfs status
rpc.svcgssd is stopped
rpc.mountd is stopped
nfsd is stopped
rpc.rquotad (pid 1342) is running...


# haresources
kvmstorage1.localdomain IPaddr::192.168.123.209/24/eth0 drbddisk::main Filesystem::/dev/drbd0::/datastore::ext4 nfslock nfs rpcidmapd mysqld

# logfile

May 16 22:44:39 kvmstorage1.localdomain heartbeat: [8177]: info: All HA resources relinquished.
May 16 22:44:40 kvmstorage1.localdomain heartbeat: [6569]: WARN: 1 lost packet(s) for [kvmstorage2.localdomain] [51:53]
May 16 22:44:40 kvmstorage1.localdomain heartbeat: [6569]: info: No pkts missing from kvmstorage2.localdomain!
May 16 22:44:41 kvmstorage1.localdomain heartbeat: [6569]: info: killing HBFIFO process 6572 with signal 15
May 16 22:44:41 kvmstorage1.localdomain heartbeat: [6569]: info: killing HBWRITE process 6573 with signal 15
May 16 22:44:41 kvmstorage1.localdomain heartbeat: [6569]: info: killing HBREAD process 6574 with signal 15
May 16 22:44:41 kvmstorage1.localdomain heartbeat: [6569]: info: Core process 6574 exited. 3 remaining
May 16 22:44:41 kvmstorage1.localdomain heartbeat: [6569]: info: Core process 6572 exited. 2 remaining
May 16 22:44:41 kvmstorage1.localdomain heartbeat: [6569]: info: Core process 6573 exited. 1 remaining
May 16 22:44:41 kvmstorage1.localdomain heartbeat: [6569]: info: kvmstorage1.localdomain Heartbeat shutdown complete.
May 16 22:45:22 kvmstorage1.localdomain heartbeat: [1527]: info: Pacemaker support: false
May 16 22:45:22 kvmstorage1.localdomain heartbeat: [1527]: WARN: Logging daemon is disabled --enabling logging daemon is recommended
May 16 22:45:22 kvmstorage1.localdomain heartbeat: [1527]: info: **************************
May 16 22:45:22 kvmstorage1.localdomain heartbeat: [1527]: info: Configuration validated. Starting heartbeat 3.0.4
May 16 22:45:22 kvmstorage1.localdomain heartbeat: [1528]: info: heartbeat: version 3.0.4
May 16 22:45:22 kvmstorage1.localdomain heartbeat: [1528]: info: Heartbeat generation: 1335851675
May 16 22:45:22 kvmstorage1.localdomain heartbeat: [1528]: info: glib: UDP Broadcast heartbeat started on port 694 (694) interface eth0
May 16 22:45:22 kvmstorage1.localdomain heartbeat: [1528]: info: glib: UDP Broadcast heartbeat closed on port 694 interface eth0 - Status: 1
May 16 22:45:22 kvmstorage1.localdomain heartbeat: [1528]: info: G_main_add_TriggerHandler: Added signal manual handler
May 16 22:45:22 kvmstorage1.localdomain heartbeat: [1528]: info: G_main_add_TriggerHandler: Added signal manual handler
May 16 22:45:22 kvmstorage1.localdomain heartbeat: [1528]: info: G_main_add_SignalHandler: Added signal handler for signal 17
May 16 22:45:22 kvmstorage1.localdomain heartbeat: [1528]: info: Local status now set to: 'up'
May 16 22:45:22 kvmstorage1.localdomain heartbeat: [1528]: info: Link kvmstorage1.localdomain:eth0 up.
May 16 22:45:24 kvmstorage1.localdomain heartbeat: [1528]: info: Link kvmstorage2.localdomain:eth0 up.
May 16 22:45:24 kvmstorage1.localdomain heartbeat: [1528]: info: Status update for node kvmstorage2.localdomain: status active
harc[1630]: 2012/05/16_22:45:24 info: Running /etc/ha.d//rc.d/status status
May 16 22:45:25 kvmstorage1.localdomain heartbeat: [1528]: info: Comm_now_up(): updating status to active
May 16 22:45:25 kvmstorage1.localdomain heartbeat: [1528]: info: Local status now set to: 'active'
May 16 22:45:25 kvmstorage1.localdomain heartbeat: [1528]: info: remote resource transition completed.
May 16 22:45:25 kvmstorage1.localdomain heartbeat: [1528]: info: remote resource transition completed.
May 16 22:45:25 kvmstorage1.localdomain heartbeat: [1528]: info: Local Resource acquisition completed. (none)
May 16 22:45:26 kvmstorage1.localdomain heartbeat: [1528]: info: kvmstorage2.localdomain wants to go standby [foreign]
May 16 22:45:31 kvmstorage1.localdomain heartbeat: [1528]: info: standby: acquire [foreign] resources from kvmstorage2.localdomain
May 16 22:45:31 kvmstorage1.localdomain heartbeat: [2253]: info: acquire local HA resources (standby).
ResourceManager[2275]: 2012/05/16_22:45:31 info: Acquiring resource group: kvmstorage1.localdomain IPaddr::192.168.123.209/24/eth0 drbddisk::main Filesystem::/dev/drbd0::/datastore::ext4 nfslock nfs rpcidmapd mysqld
IPaddr[2321]: 2012/05/16_22:45:32 INFO: Resource is stopped
ResourceManager[2275]: 2012/05/16_22:45:32 info: Running /etc/ha.d/resource.d/IPaddr 192.168.123.209/24/eth0 start
IPaddr[2439]: 2012/05/16_22:45:32 INFO: Using calculated netmask for 192.168.123.209: 255.255.255.0
IPaddr[2439]: 2012/05/16_22:45:32 INFO: eval ifconfig eth0:0 192.168.123.209 netmask 255.255.255.0 broadcast 192.168.123.255
IPaddr[2402]: 2012/05/16_22:45:32 INFO: Success
ResourceManager[2275]: 2012/05/16_22:45:32 info: Running /etc/ha.d/resource.d/drbddisk main start
Filesystem[2627]: 2012/05/16_22:45:33 INFO: Resource is stopped
ResourceManager[2275]: 2012/05/16_22:45:33 info: Running /etc/ha.d/resource.d/Filesystem /dev/drbd0 /datastore ext4 start
Filesystem[2741]: 2012/05/16_22:45:33 INFO: Running start for /dev/drbd0 on /datastore
Filesystem[2729]: 2012/05/16_22:45:33 INFO: Success
ResourceManager[2275]: 2012/05/16_22:45:33 info: Running /etc/init.d/nfslock start
ResourceManager[2275]: 2012/05/16_22:45:34 info: Running /etc/init.d/rpcidmapd start
ResourceManager[2275]: 2012/05/16_22:45:35 info: Running /etc/init.d/mysqld start
May 16 22:45:36 kvmstorage1.localdomain heartbeat: [2253]: info: local HA resource acquisition completed (standby).
May 16 22:45:36 kvmstorage1.localdomain heartbeat: [1528]: info: Standby resource acquisition done [foreign].
May 16 22:45:36 kvmstorage1.localdomain heartbeat: [1528]: info: Initial resource acquisition complete (auto_failback)
May 16 22:45:36 kvmstorage1.localdomain heartbeat: [1528]: info: remote resource transition completed.














On 16 mei 2012, at 22:36, Matt Graham wrote:

> From: Lars Ellenberg <lars.ellenberg [at] linbit>
> [snippage]
>> May I ask why you chose to use heartbeat haresource mode instead of
>> RHCS or Pacemaker, or any other potential candidate for the job?
>>
>> Just curious here. I'm trying to figure out how common it is nowadays
>> to do new haresources deployments with new projects.
>
> The last time I tried pacemaker on a test cluster at work, it didn't work at
> all. This was apparently because multicast was required, and the people who
> run networking at work said they wouldn't turn multicast on for any of the
> switches. heartbeat, though, works reasonably well because it doesn't require
> multicast.
>
> FWIW, Marcel's haresources had "nfs nfslock", while a heartbeat-using NFS
> cluster here has "nfslock nfs" in the same place. The /etc/init.d/ scripts
> have chkconfig comments showing nfslock starting before nfs. So that's
> another thing for Marcel to look at.
>
> --
> Matt G / Dances With Crows
> The Crow202 Blog: http://crow202.org/wordpress/
> There is no Darkness in Eternity/But only Light too dim for us to see
>
> _______________________________________________
> drbd-user mailing list
> drbd-user [at] lists
> http://lists.linbit.com/mailman/listinfo/drbd-user

_______________________________________________
drbd-user mailing list
drbd-user [at] lists
http://lists.linbit.com/mailman/listinfo/drbd-user


marcel at kraan

May 16, 2012, 2:37 PM

Post #8 of 8 (1420 views)
Permalink
Re: NFS not starting with heartbeat [In reply to]

When i turn "off" the services ifs and nfslock at start
chkconfig nfs off
chkconfig nfslock off

then everything is working. also after reboot.
so my problem is solved and i will go into pacemaker next week..



On 16 mei 2012, at 22:48, Marcel Kraan wrote:

> Hello Matt, Lars,
>
> While installing PaceMaker the "nfslock nfs" solution frm Matt worked when i restart heartbeat
> when i reboot the server the ifs is still down? (very strange)
>
> root [at] kvmstorage ~]# ifconfig
> eth0 Link encap:Ethernet HWaddr 52:54:00:18:BC:3D
> inet addr:192.168.123.211 Bcast:192.168.123.255 Mask:255.255.255.0
> inet6 addr: fe80::5054:ff:fe18:bc3d/64 Scope:Link
> UP BROADCAST RUNNING MULTICAST MTU:1500 Metric:1
> RX packets:92869 errors:0 dropped:15 overruns:0 frame:0
> TX packets:97696 errors:0 dropped:0 overruns:0 carrier:0
> collisions:0 txqueuelen:1000
> RX bytes:113420015 (108.1 MiB) TX bytes:114931967 (109.6 MiB)
> Interrupt:10 Base address:0x2000
>
> eth0:0 Link encap:Ethernet HWaddr 52:54:00:18:BC:3D
> inet addr:192.168.123.209 Bcast:192.168.123.255 Mask:255.255.255.0
> UP BROADCAST RUNNING MULTICAST MTU:1500 Metric:1
> Interrupt:10 Base address:0x2000
>
> lo Link encap:Local Loopback
> inet addr:127.0.0.1 Mask:255.0.0.0
> inet6 addr: ::1/128 Scope:Host
> UP LOOPBACK RUNNING MTU:16436 Metric:1
> RX packets:12 errors:0 dropped:0 overruns:0 frame:0
> TX packets:12 errors:0 dropped:0 overruns:0 carrier:0
> collisions:0 txqueuelen:0
> RX bytes:840 (840.0 b) TX bytes:840 (840.0 b)
>
> [root [at] kvmstorage ~]# service nfs status
> rpc.svcgssd is stopped
> rpc.mountd is stopped
> nfsd is stopped
> rpc.rquotad (pid 1342) is running...
>
>
> # haresources
> kvmstorage1.localdomain IPaddr::192.168.123.209/24/eth0 drbddisk::main Filesystem::/dev/drbd0::/datastore::ext4 nfslock nfs rpcidmapd mysqld
>
> # logfile
>
> May 16 22:44:39 kvmstorage1.localdomain heartbeat: [8177]: info: All HA resources relinquished.
> May 16 22:44:40 kvmstorage1.localdomain heartbeat: [6569]: WARN: 1 lost packet(s) for [kvmstorage2.localdomain] [51:53]
> May 16 22:44:40 kvmstorage1.localdomain heartbeat: [6569]: info: No pkts missing from kvmstorage2.localdomain!
> May 16 22:44:41 kvmstorage1.localdomain heartbeat: [6569]: info: killing HBFIFO process 6572 with signal 15
> May 16 22:44:41 kvmstorage1.localdomain heartbeat: [6569]: info: killing HBWRITE process 6573 with signal 15
> May 16 22:44:41 kvmstorage1.localdomain heartbeat: [6569]: info: killing HBREAD process 6574 with signal 15
> May 16 22:44:41 kvmstorage1.localdomain heartbeat: [6569]: info: Core process 6574 exited. 3 remaining
> May 16 22:44:41 kvmstorage1.localdomain heartbeat: [6569]: info: Core process 6572 exited. 2 remaining
> May 16 22:44:41 kvmstorage1.localdomain heartbeat: [6569]: info: Core process 6573 exited. 1 remaining
> May 16 22:44:41 kvmstorage1.localdomain heartbeat: [6569]: info: kvmstorage1.localdomain Heartbeat shutdown complete.
> May 16 22:45:22 kvmstorage1.localdomain heartbeat: [1527]: info: Pacemaker support: false
> May 16 22:45:22 kvmstorage1.localdomain heartbeat: [1527]: WARN: Logging daemon is disabled --enabling logging daemon is recommended
> May 16 22:45:22 kvmstorage1.localdomain heartbeat: [1527]: info: **************************
> May 16 22:45:22 kvmstorage1.localdomain heartbeat: [1527]: info: Configuration validated. Starting heartbeat 3.0.4
> May 16 22:45:22 kvmstorage1.localdomain heartbeat: [1528]: info: heartbeat: version 3.0.4
> May 16 22:45:22 kvmstorage1.localdomain heartbeat: [1528]: info: Heartbeat generation: 1335851675
> May 16 22:45:22 kvmstorage1.localdomain heartbeat: [1528]: info: glib: UDP Broadcast heartbeat started on port 694 (694) interface eth0
> May 16 22:45:22 kvmstorage1.localdomain heartbeat: [1528]: info: glib: UDP Broadcast heartbeat closed on port 694 interface eth0 - Status: 1
> May 16 22:45:22 kvmstorage1.localdomain heartbeat: [1528]: info: G_main_add_TriggerHandler: Added signal manual handler
> May 16 22:45:22 kvmstorage1.localdomain heartbeat: [1528]: info: G_main_add_TriggerHandler: Added signal manual handler
> May 16 22:45:22 kvmstorage1.localdomain heartbeat: [1528]: info: G_main_add_SignalHandler: Added signal handler for signal 17
> May 16 22:45:22 kvmstorage1.localdomain heartbeat: [1528]: info: Local status now set to: 'up'
> May 16 22:45:22 kvmstorage1.localdomain heartbeat: [1528]: info: Link kvmstorage1.localdomain:eth0 up.
> May 16 22:45:24 kvmstorage1.localdomain heartbeat: [1528]: info: Link kvmstorage2.localdomain:eth0 up.
> May 16 22:45:24 kvmstorage1.localdomain heartbeat: [1528]: info: Status update for node kvmstorage2.localdomain: status active
> harc[1630]: 2012/05/16_22:45:24 info: Running /etc/ha.d//rc.d/status status
> May 16 22:45:25 kvmstorage1.localdomain heartbeat: [1528]: info: Comm_now_up(): updating status to active
> May 16 22:45:25 kvmstorage1.localdomain heartbeat: [1528]: info: Local status now set to: 'active'
> May 16 22:45:25 kvmstorage1.localdomain heartbeat: [1528]: info: remote resource transition completed.
> May 16 22:45:25 kvmstorage1.localdomain heartbeat: [1528]: info: remote resource transition completed.
> May 16 22:45:25 kvmstorage1.localdomain heartbeat: [1528]: info: Local Resource acquisition completed. (none)
> May 16 22:45:26 kvmstorage1.localdomain heartbeat: [1528]: info: kvmstorage2.localdomain wants to go standby [foreign]
> May 16 22:45:31 kvmstorage1.localdomain heartbeat: [1528]: info: standby: acquire [foreign] resources from kvmstorage2.localdomain
> May 16 22:45:31 kvmstorage1.localdomain heartbeat: [2253]: info: acquire local HA resources (standby).
> ResourceManager[2275]: 2012/05/16_22:45:31 info: Acquiring resource group: kvmstorage1.localdomain IPaddr::192.168.123.209/24/eth0 drbddisk::main Filesystem::/dev/drbd0::/datastore::ext4 nfslock nfs rpcidmapd mysqld
> IPaddr[2321]: 2012/05/16_22:45:32 INFO: Resource is stopped
> ResourceManager[2275]: 2012/05/16_22:45:32 info: Running /etc/ha.d/resource.d/IPaddr 192.168.123.209/24/eth0 start
> IPaddr[2439]: 2012/05/16_22:45:32 INFO: Using calculated netmask for 192.168.123.209: 255.255.255.0
> IPaddr[2439]: 2012/05/16_22:45:32 INFO: eval ifconfig eth0:0 192.168.123.209 netmask 255.255.255.0 broadcast 192.168.123.255
> IPaddr[2402]: 2012/05/16_22:45:32 INFO: Success
> ResourceManager[2275]: 2012/05/16_22:45:32 info: Running /etc/ha.d/resource.d/drbddisk main start
> Filesystem[2627]: 2012/05/16_22:45:33 INFO: Resource is stopped
> ResourceManager[2275]: 2012/05/16_22:45:33 info: Running /etc/ha.d/resource.d/Filesystem /dev/drbd0 /datastore ext4 start
> Filesystem[2741]: 2012/05/16_22:45:33 INFO: Running start for /dev/drbd0 on /datastore
> Filesystem[2729]: 2012/05/16_22:45:33 INFO: Success
> ResourceManager[2275]: 2012/05/16_22:45:33 info: Running /etc/init.d/nfslock start
> ResourceManager[2275]: 2012/05/16_22:45:34 info: Running /etc/init.d/rpcidmapd start
> ResourceManager[2275]: 2012/05/16_22:45:35 info: Running /etc/init.d/mysqld start
> May 16 22:45:36 kvmstorage1.localdomain heartbeat: [2253]: info: local HA resource acquisition completed (standby).
> May 16 22:45:36 kvmstorage1.localdomain heartbeat: [1528]: info: Standby resource acquisition done [foreign].
> May 16 22:45:36 kvmstorage1.localdomain heartbeat: [1528]: info: Initial resource acquisition complete (auto_failback)
> May 16 22:45:36 kvmstorage1.localdomain heartbeat: [1528]: info: remote resource transition completed.
>
>
>
>
>
>
>
>
>
>
>
>
>
>
> On 16 mei 2012, at 22:36, Matt Graham wrote:
>
>> From: Lars Ellenberg <lars.ellenberg [at] linbit>
>> [snippage]
>>> May I ask why you chose to use heartbeat haresource mode instead of
>>> RHCS or Pacemaker, or any other potential candidate for the job?
>>>
>>> Just curious here. I'm trying to figure out how common it is nowadays
>>> to do new haresources deployments with new projects.
>>
>> The last time I tried pacemaker on a test cluster at work, it didn't work at
>> all. This was apparently because multicast was required, and the people who
>> run networking at work said they wouldn't turn multicast on for any of the
>> switches. heartbeat, though, works reasonably well because it doesn't require
>> multicast.
>>
>> FWIW, Marcel's haresources had "nfs nfslock", while a heartbeat-using NFS
>> cluster here has "nfslock nfs" in the same place. The /etc/init.d/ scripts
>> have chkconfig comments showing nfslock starting before nfs. So that's
>> another thing for Marcel to look at.
>>
>> --
>> Matt G / Dances With Crows
>> The Crow202 Blog: http://crow202.org/wordpress/
>> There is no Darkness in Eternity/But only Light too dim for us to see
>>
>> _______________________________________________
>> drbd-user mailing list
>> drbd-user [at] lists
>> http://lists.linbit.com/mailman/listinfo/drbd-user
>
> _______________________________________________
> drbd-user mailing list
> drbd-user [at] lists
> http://lists.linbit.com/mailman/listinfo/drbd-user

_______________________________________________
drbd-user mailing list
drbd-user [at] lists
http://lists.linbit.com/mailman/listinfo/drbd-user

DRBD users RSS feed   Index | Next | Previous | View Threaded
 
 


Interested in having your list archived? Contact Gossamer Threads
 
  Web Applications & Managed Hosting Powered by Gossamer Threads Inc.