xref: /titanic_50/usr/src/cmd/avs/dscfglockd/dscfglockd.c (revision 4ef97ab2f8bc3015a04caa68a4fb9053546fe1ec)
1 /*
2  * CDDL HEADER START
3  *
4  * The contents of this file are subject to the terms of the
5  * Common Development and Distribution License (the "License").
6  * You may not use this file except in compliance with the License.
7  *
8  * You can obtain a copy of the license at usr/src/OPENSOLARIS.LICENSE
9  * or http://www.opensolaris.org/os/licensing.
10  * See the License for the specific language governing permissions
11  * and limitations under the License.
12  *
13  * When distributing Covered Code, include this CDDL HEADER in each
14  * file and include the License file at usr/src/OPENSOLARIS.LICENSE.
15  * If applicable, add the following below this CDDL HEADER, with the
16  * fields enclosed by brackets "[]" replaced with your own identifying
17  * information: Portions Copyright [yyyy] [name of copyright owner]
18  *
19  * CDDL HEADER END
20  */
21 /*
22  * Copyright 2010 Sun Microsystems, Inc.  All rights reserved.
23  * Use is subject to license terms.
24  */
25 
26 #include <signal.h>
27 #include <sys/types.h>
28 #include <sys/time.h>
29 #include <sys/socket.h>
30 #include <netinet/in.h>
31 #include <netinet/tcp.h>
32 #include <arpa/inet.h>
33 #include <netdb.h>
34 #include <fcntl.h>
35 #include <string.h>
36 #include <memory.h>
37 #include <sys/param.h>
38 #include <sys/pathconf.h>
39 #include <netdir.h>
40 #include <netconfig.h>
41 #include <sys/sockio.h>
42 #include <net/if.h>
43 #include <sys/resource.h>
44 #include <stdio.h>
45 #include <errno.h>
46 #include <assert.h>
47 #include <locale.h>
48 #include <unistd.h>
49 #include <stdlib.h>
50 #include <string.h>
51 #include <strings.h>
52 #include <sys/unistat/spcs_s.h>
53 #include <sys/unistat/spcs_s_u.h>
54 #include <sys/unistat/spcs_errors.h>
55 
56 #include <sys/nsctl/cfg.h>
57 #include <sys/nsctl/cfg_lockd.h>
58 
59 #ifdef DEBUG
60 #define	DPF(m)		if (debug) (void) fprintf m
61 #else
62 #define	DPF(m)
63 #endif
64 
65 #ifdef	TTY_MESSAGES
66 #define	CLOSE_FD	3
67 #else
68 #define	CLOSE_FD	0
69 #endif
70 
71 #define	MAX_LOCKQ	1024
72 #define	MAX_DAEMONS	1024
73 #define	MAX_LOCAL	1024
74 #define	MAX_UNLOCK	32
75 #define	MAX_TIMEOUTS	3
76 #define	TIMEOUT_SECS	5
77 
78 static char program[] = "dscfglockd";
79 static int debug;
80 static int lstate;
81 static int msgtrace;
82 static FILE *debugfile = NULL;
83 
84 struct lock_req {
85 	cfglockd_t	type;	/* read or write */
86 	pid_t	pid;		/* pid of read locker or local writer */
87 	daemonaddr_t	remote;	/* remote machine requesting write lock */
88 	int		state;	/* for write locks */
89 	int32_t		order;	/* who gets priority? */
90 } lock_queue[MAX_LOCKQ];
91 
92 struct unlock_s {
93 	pid_t	pid;		/* pid of locker */
94 	uint8_t seq;		/* seq number of last lock request */
95 } unlock_buf[MAX_UNLOCK];
96 
97 int next_req;
98 int32_t order;
99 
100 #define	lock_wanted	lock_queue[0]
101 long	ticker	= 1l;
102 
103 #define	ALIVE		0x10
104 #define	READ_LOCK	0x11
105 #define	WRITE_LOCK	0x12
106 #define	UNLOCK		0x13
107 #define	GRANTED		0x14
108 
109 int next_q;
110 
111 struct {
112 	cfglockd_t	type;
113 	int		nholders;
114 	int		state;
115 	daemonaddr_t	holder;
116 	struct lockdaemon	*remote_daemon;
117 	pid_t		holding_pid[MAX_LOCAL];
118 } the_lock;
119 
120 daemonaddr_t	thishost;
121 daemonaddr_t	localhost;
122 
123 #define	STATE_CLEAR	0
124 #define	STATE_ASKED	1
125 #define	STATE_OKAYED	2
126 #define	STATE_WANTS	3
127 #define	lockdaemon_dead(ldp)	((ticker - (ldp)->timeout) > MAX_TIMEOUTS)
128 #define	CRIT_BEGIN()	(void) sighold(SIGALRM)
129 #define	CRIT_END()	(void) sigrelse(SIGALRM)
130 
131 #define	NORMAL_UNLOCK	0
132 #define	FORCE_UNLOCK	1
133 
134 struct lockdaemon {
135 	daemonaddr_t	host;
136 	int	up;
137 	long	timeout;
138 	int	inuse;
139 	int	state;
140 	int32_t	order;
141 } daemon_list[MAX_DAEMONS];
142 
143 unsigned short	lock_port = CFG_SERVER_PORT;
144 int	lock_soc = 0;
145 int	pf_inet = PF_INET;
146 #define	dp_addr(p)	inet_ntoa(((struct sockaddr_in *)p)->sin_addr)
147 
148 #define	MAXIFS 32
149 
150 static char *
lockd_type(cfglockd_t type)151 lockd_type(cfglockd_t type)
152 {
153 	switch (type) {
154 	case LOCK_NOTLOCKED:	return "NotLocked";
155 	case LOCK_READ:		return "Read";
156 	case LOCK_WRITE:	return "Write";
157 	case LOCK_LOCKED:	return "Locked";
158 	case LOCK_LOCKEDBY:	return "LockedBy";
159 	case LOCK_STAT:		return "Stat";
160 	case LOCK_ACK:		return "Ack";
161 	default:		return "*unknown*";
162 	}
163 }
164 
165 static char *
lockd_state(int state)166 lockd_state(int state)
167 {
168 	switch (state) {
169 	case STATE_CLEAR:	return "Clear";
170 	case STATE_ASKED:	return "Asked";
171 	case STATE_OKAYED:	return "Okayed";
172 	case STATE_WANTS:	return "Wants";
173 	default:		return "*unknown*";
174 	}
175 }
176 
177 static char *
lockd_msg(int message)178 lockd_msg(int message)
179 {
180 	switch (message) {
181 	case ALIVE:		return "Alive";
182 	case READ_LOCK:		return "ReadLock";
183 	case WRITE_LOCK:	return "WriteLock";
184 	case UNLOCK:		return "Unlock";
185 	case GRANTED:		return "Granted";
186 	default:		return lockd_type((cfglockd_t)message);
187 	}
188 }
189 
190 /*
191  * The following is stolen from autod_nfs.c
192  */
193 static void
getmyaddrs(struct ifconf * ifc)194 getmyaddrs(struct ifconf *ifc)
195 {
196 	int sock;
197 	int numifs;
198 	char *buf;
199 	int family;
200 
201 	ifc->ifc_buf = NULL;
202 	ifc->ifc_len = 0;
203 
204 #ifdef AF_INET6
205 	family = AF_INET6;
206 #else
207 	family = AF_INET;
208 #endif
209 	if ((sock = socket(family, SOCK_DGRAM, 0)) < 0) {
210 #ifdef DEBUG
211 		perror("getmyaddrs(): socket");
212 #endif
213 		return;
214 	}
215 
216 	if (ioctl(sock, SIOCGIFNUM, (char *)&numifs) < 0) {
217 #ifdef DEBUG
218 		perror("getmyaddrs(): SIOCGIFNUM");
219 #endif
220 		numifs = MAXIFS;
221 	}
222 
223 	buf = (char *)malloc(numifs * sizeof (struct ifreq));
224 	if (buf == NULL) {
225 #ifdef DEBUG
226 		(void) fprintf(stderr, "getmyaddrs(): malloc failed\n");
227 #endif
228 		(void) close(sock);
229 		return;
230 	}
231 
232 	ifc->ifc_buf = buf;
233 	ifc->ifc_len = numifs * sizeof (struct ifreq);
234 
235 	if (ioctl(sock, SIOCGIFCONF, (char *)ifc) < 0) {
236 #ifdef DEBUG
237 		perror("getmyaddrs(): SIOCGIFCONF");
238 #endif
239 	}
240 
241 	(void) close(sock);
242 }
243 
244 struct ifconf *ifc;
245 
246 static int
cmp_addr(daemonaddr_t * a,daemonaddr_t * b)247 cmp_addr(daemonaddr_t *a, daemonaddr_t *b)
248 {
249 	int rc;
250 	rc = memcmp(&(a->sin_addr), &(b->sin_addr), sizeof (a->sin_addr));
251 	DPF((stderr, "compare %s %hu with", dp_addr(a), a->sin_port));
252 	DPF((stderr, " %s %hu = %d\n", dp_addr(b), b->sin_port, rc));
253 	return (rc);
254 }
255 
256 static int
addr_is_holder(int32_t order)257 addr_is_holder(int32_t order)
258 {
259 	return ((the_lock.nholders > 0) && the_lock.remote_daemon != NULL &&
260 	    (order == the_lock.remote_daemon->order));
261 }
262 
263 static int
islocalhost(daemonaddr_t * host)264 islocalhost(daemonaddr_t *host)
265 {
266 	int n;
267 	struct sockaddr_in *s1, *s2;
268 	struct ifreq *ifr;
269 	int retval = 0;
270 
271 	ifr = ifc->ifc_req;
272 	n = ifc->ifc_len / sizeof (struct ifreq);
273 	s1 = host;
274 	s2 = NULL;
275 	for (; n > 0; n--, ifr++) {
276 		if (ifr->ifr_addr.sa_family != AF_INET)
277 			continue;
278 
279 		/* LINTED pointer alignment */
280 		s2 = (struct sockaddr_in *)&ifr->ifr_addr;
281 
282 		if (memcmp((char *)&s2->sin_addr,
283 		    (char *)&s1->sin_addr, sizeof (s1->sin_addr)) == 0) {
284 			retval = 1;
285 			/* it's me */
286 			break;
287 		}
288 	}
289 	return (retval);
290 }
291 
292 static void
send_lockmsg(int cmd,pid_t pid,daemonaddr_t * dp,uint8_t seq)293 send_lockmsg(int cmd, pid_t pid, daemonaddr_t *dp, uint8_t seq)
294 {
295 	struct lock_msg message_buf;
296 	int rc;
297 
298 	if (msgtrace && debugfile) {
299 		time_t t = time(0);
300 		(void) fprintf(debugfile, "%19.19s send %-9.9s to   %s\n",
301 		    ctime(&t), lockd_msg(cmd), dp_addr(dp));
302 	}
303 	DPF((stderr, "send %d to %s port %hu\n", cmd,
304 	    dp_addr(dp), dp->sin_port));
305 	message_buf.message = cmd;
306 	message_buf.pid = pid;
307 	message_buf.order = order;
308 	message_buf.seq = seq;
309 	do {
310 		rc = sendto(lock_soc, &message_buf, sizeof (message_buf), 0,
311 		    (struct sockaddr *)dp, sizeof (struct sockaddr));
312 	} while (rc == -1 && errno == EINTR);
313 	if (rc == -1)
314 		spcs_log("cfglockd", NULL, "sendto rc -1 errno %d", errno);
315 }
316 
317 /*
318  * send an alive message to all configured daemons so that they can tell
319  * us if they are holding a write lock.
320  */
321 
322 static void
send_aliveall()323 send_aliveall()
324 {
325 	struct lockdaemon *ldp;
326 	int i;
327 	for (i = 0, ldp = daemon_list; i < MAX_DAEMONS; i++, ldp++) {
328 		if (ldp->inuse == 0)
329 			break;
330 		send_lockmsg(ALIVE, (pid_t)0, &(ldp->host), 0);
331 	}
332 }
333 
334 /* find the lock daemon structure for a give daemon address */
335 
336 static struct lockdaemon *
find_lockdaemon(daemonaddr_t * d)337 find_lockdaemon(daemonaddr_t *d)
338 {
339 	struct lockdaemon *ldp;
340 	int i;
341 	for (i = 0, ldp = daemon_list; i < MAX_DAEMONS; i++, ldp++) {
342 		if (ldp->inuse == 0)
343 			break;
344 		if (cmp_addr(&(ldp->host), d) == 0)
345 			return (ldp);
346 	}
347 	return (NULL);
348 }
349 
350 /*
351  * a messge has been received from daemon, note this and if the daemon
352  * was previously dead  and we have the write lock tell it that we do.
353  */
354 
355 static void
daemon_alive(daemonaddr_t * daemon,int32_t order)356 daemon_alive(daemonaddr_t *daemon, int32_t order)
357 {
358 	struct lockdaemon *ldp;
359 	int i;
360 
361 	for (i = 0, ldp = daemon_list; i < MAX_DAEMONS; i++, ldp++) {
362 		if (ldp->inuse == 0)
363 			break;
364 		if (cmp_addr(&(ldp->host), daemon) == 0) {
365 			ldp->order = order;
366 			ldp->timeout = ticker;
367 			if (ldp->up == 0) {
368 				spcs_log("cfglockd", NULL,
369 				    "daemon restarted on %s\n",
370 				    dp_addr(daemon));
371 				DPF((stderr, "daemon restarted on %s\n",
372 				    dp_addr(daemon)));
373 				ldp->up = 1;
374 				goto come_up;
375 			}
376 			return;
377 		}
378 	}
379 	/* new daemon has announced itself */
380 	if (i < MAX_DAEMONS) {
381 		DPF((stderr, "new daemon on %s\n", dp_addr(daemon)));
382 		spcs_log("cfglockd", NULL,
383 		    "new daemon on %s\n", dp_addr(daemon));
384 		ldp->host = *daemon;
385 		ldp->inuse = 1;
386 		ldp->timeout = ticker;
387 		ldp->order = order;
388 	} else {
389 		/* problem, more daemons than expected */
390 		i++;
391 	}
392 come_up:
393 	if (the_lock.type == LOCK_WRITE && the_lock.remote_daemon == NULL)
394 		send_lockmsg(WRITE_LOCK, (pid_t)0, daemon, 0);
395 }
396 
397 static void
delete_queue_entry(struct lock_req * req)398 delete_queue_entry(struct  lock_req *req)
399 {
400 	int i;
401 
402 	for (i = (req - lock_queue); i++ < next_req; req++)
403 		*req = *(req+1);
404 	next_req--;
405 }
406 
407 static void
take_lock(int ackmessage)408 take_lock(int ackmessage)
409 {
410 	send_lockmsg(ackmessage, (pid_t)0, &lock_wanted.remote, 0);
411 	delete_queue_entry(lock_queue);
412 }
413 
414 static void
check_for_write_lock()415 check_for_write_lock()
416 {
417 	struct lockdaemon *ldp;
418 	int i;
419 	int	wait = 0;
420 
421 	DPF((stderr, "check for lock\n"));
422 	if (lock_wanted.state != STATE_ASKED)
423 		return;
424 	for (i = 0, ldp = daemon_list; i < MAX_DAEMONS; i++, ldp++) {
425 		if (ldp->inuse == 0)
426 			break;
427 		if (ldp->up && ldp->state != STATE_OKAYED) {
428 			wait = 1;
429 			break;
430 		}
431 	}
432 	if (wait == 0 && lock_wanted.type == LOCK_WRITE) {
433 		the_lock.type = LOCK_WRITE;
434 		the_lock.holding_pid[0] = lock_wanted.pid;
435 		the_lock.nholders = 1;
436 		the_lock.state = STATE_CLEAR;
437 		take_lock(LOCK_LOCKED);
438 	}
439 }
440 
441 static void
lock_granted(daemonaddr_t * da)442 lock_granted(daemonaddr_t *da)
443 {
444 	struct lockdaemon *ldp;
445 
446 	if ((ldp = find_lockdaemon(da)) != NULL) {
447 		/* if we already own the lock, throw the msg away */
448 		if (the_lock.remote_daemon == NULL &&
449 		    the_lock.type == LOCK_WRITE) {
450 			return;
451 		}
452 
453 		/*
454 		 * If the current lock isn't a write lock and we're not
455 		 * asking for one
456 		 * -OR-
457 		 * The current lock is a write lock and it's not owned by us
458 		 * -THEN-
459 		 * send back an unlocked message.
460 		 */
461 		if ((the_lock.type != LOCK_WRITE &&
462 		    the_lock.state != STATE_ASKED) ||
463 		    (the_lock.type == LOCK_WRITE &&
464 		    the_lock.remote_daemon != NULL)) {
465 			send_lockmsg(UNLOCK, (pid_t)0, &(ldp->host), 0);
466 			return;
467 		}
468 		ldp->state = STATE_OKAYED;
469 	}
470 	check_for_write_lock();
471 }
472 
473 static int
try_lock()474 try_lock()
475 {
476 	struct lockdaemon *ldp;
477 	int i;
478 
479 	switch (the_lock.type) {
480 	case LOCK_READ:
481 		if (lock_wanted.type == LOCK_READ) {
482 			i = the_lock.nholders++;
483 			the_lock.holding_pid[i] = lock_wanted.pid;
484 			the_lock.state = STATE_CLEAR;
485 			DPF((stderr, "increment read lockers to %d\n",
486 			    the_lock.nholders));
487 			take_lock(LOCK_LOCKED);
488 			break;
489 		}
490 		/* write lock has to wait */
491 		break;
492 	case LOCK_WRITE:
493 		/* lock has to wait until write lock is cleared */
494 		break;
495 	case LOCK_NOTLOCKED:
496 		if (lock_wanted.type == LOCK_READ) {
497 			DPF((stderr, "local locker, 1 lock holder\n"));
498 			the_lock.holding_pid[0] = lock_wanted.pid;
499 			the_lock.nholders = 1;
500 			the_lock.type = LOCK_READ;
501 			the_lock.state = STATE_CLEAR;
502 			the_lock.remote_daemon = NULL;
503 			take_lock(LOCK_LOCKED);
504 			return (1);
505 		}
506 		if (islocalhost(&lock_wanted.remote)) {
507 			DPF((stderr, "local locker, take write lock\n"));
508 			/* tell everyone I'm locking */
509 			if (lock_wanted.state != STATE_ASKED) {
510 				for (i = 0, ldp = daemon_list; i < MAX_DAEMONS;
511 				    i++, ldp++) {
512 					if (ldp->inuse == 0)
513 						break;
514 					ldp->state = STATE_ASKED;
515 					send_lockmsg(WRITE_LOCK, (pid_t)0,
516 					    &(ldp->host), 0);
517 				}
518 			}
519 			lock_wanted.state = STATE_ASKED;
520 			check_for_write_lock();
521 			the_lock.remote_daemon = NULL;
522 			the_lock.state = STATE_ASKED;
523 			return (0);
524 		} else {
525 			DPF((stderr, "remote locker, take write lock\n"));
526 			the_lock.type = LOCK_WRITE;
527 			the_lock.holder = lock_wanted.remote;
528 			the_lock.nholders = 1;
529 			the_lock.remote_daemon =
530 			    find_lockdaemon(&the_lock.holder);
531 			the_lock.state = STATE_CLEAR;
532 			/* okay to remote */
533 			take_lock(GRANTED);
534 		}
535 		break;
536 	default:
537 		DPF((stderr, "weird lock type held - %d\n", the_lock.type));
538 		the_lock.type = LOCK_NOTLOCKED;
539 		break;
540 	}
541 	return (0);
542 }
543 
544 static void
process_queue()545 process_queue()
546 {
547 	if (next_req < 1)
548 		return;		/* no locks queued */
549 	while (try_lock())
550 		;
551 }
552 
553 static int
lock_sort(const void * a,const void * b)554 lock_sort(const void *a, const void *b)
555 {
556 	struct lock_req *left = (struct lock_req *)a;
557 	struct lock_req *right = (struct lock_req *)b;
558 
559 	return (left->order - right->order);
560 }
561 
562 static void
queue_lock(cfglockd_t type,struct lock_msg * msg,daemonaddr_t * addr)563 queue_lock(cfglockd_t type, struct lock_msg *msg, daemonaddr_t *addr)
564 {
565 	int	i;
566 	struct lock_req *lrp;
567 	struct lockdaemon *ldp;
568 
569 	/* first check if new lock matches current lock */
570 	if (the_lock.type == type && addr_is_holder(msg->order)) {
571 		/* remote daemon missed locked message */
572 		send_lockmsg(GRANTED, (pid_t)0, addr, msg->seq);
573 		return;
574 	}
575 
576 	/* next search queue to check for duplicate */
577 	for (i = 0, lrp = lock_queue; i++ < next_req; lrp++) {
578 		if (lrp->type == type && lrp->pid == msg->pid &&
579 		    cmp_addr(addr, &(lrp->remote)) == 0)
580 			return;
581 
582 	}
583 
584 	/*
585 	 * It's a new lock request.  Are we in the middle of
586 	 * obtaining one for ourselves?
587 	 */
588 
589 	if (the_lock.type == LOCK_NOTLOCKED && the_lock.state == STATE_ASKED) {
590 		/* did a higher priority request just come in? */
591 		if (msg->order < order) {
592 			/* requeue our request */
593 			the_lock.state = STATE_CLEAR;
594 			lock_wanted.state = STATE_CLEAR;
595 
596 			/* let the other lockds know */
597 			for (i = 0, ldp = daemon_list; i < MAX_DAEMONS;
598 			    i++, ldp++) {
599 				if (ldp->inuse == 0)
600 					break;
601 				if (ldp->up && ldp->state == STATE_OKAYED) {
602 					send_lockmsg(UNLOCK, (pid_t)0,
603 					    &(ldp->host), 0);
604 				}
605 			}
606 		}
607 	}
608 
609 
610 	lrp = lock_queue;
611 	lrp += (next_req++);
612 	lrp->type = type;
613 	lrp->pid = msg->pid;
614 	lrp->state = STATE_CLEAR;
615 	lrp->order = msg->order;
616 	if (addr) {
617 		lrp->remote = *addr;
618 	}
619 
620 	if (next_req > 1)
621 		qsort(lock_queue, next_req, sizeof (lock_queue[0]), lock_sort);
622 
623 	if (the_lock.type != LOCK_WRITE)
624 		process_queue();
625 }
626 
627 static void
lock_stat()628 lock_stat()
629 {
630 	char *lt = "Unknown";
631 	struct lockdaemon *ldp;
632 	int i;
633 
634 	switch (the_lock.type) {
635 	case LOCK_NOTLOCKED:
636 		lt = "not locked";
637 		break;
638 	case LOCK_READ:
639 		lt = "read locked";
640 		break;
641 	case LOCK_WRITE:
642 		lt = "write locked";
643 		break;
644 	}
645 	spcs_log("cfglockd", NULL, "Lock is %s (%d)", lt, the_lock.type);
646 	spcs_log("cfglockd", NULL, "There are %d holders of the lock",
647 	    the_lock.nholders);
648 	if (the_lock.nholders > 0) {
649 		for (i = 0; i < the_lock.nholders; i++)
650 			spcs_log("cfglockd", NULL, "holding_pid[%d] = %6d", i,
651 			    the_lock.holding_pid[i]);
652 	}
653 	spcs_log("cfglockd", NULL, "holder daemon was %s port %hu, remote %x",
654 	    dp_addr(&the_lock.holder), the_lock.holder.sin_port,
655 	    the_lock.remote_daemon);
656 	spcs_log("cfglockd", NULL, "Lock queue, %d requests", next_req);
657 	for (i = 0; i < next_req; i++) {
658 		spcs_log("cfglockd", NULL, "request %d type %d order %d", i,
659 		    lock_queue[i].type, lock_queue[i].order);
660 		spcs_log("cfglockd", NULL, "  client %s port %hu, pid %d",
661 		    dp_addr(&lock_queue[i].remote),
662 		    lock_queue[i].remote.sin_port, lock_queue[i].pid);
663 	}
664 	spcs_log("cfglockd", NULL, "Daemon list");
665 
666 	for (i = 0, ldp = daemon_list; i < MAX_DAEMONS; i++, ldp++) {
667 		if (ldp->inuse == 0)
668 			break;
669 		spcs_log("cfglockd", NULL, "daemon %d, %s port %hu", i,
670 		    dp_addr(&ldp->host), ldp->host.sin_port);
671 		spcs_log("cfglockd", NULL,
672 		    "  up %d timeout %ld missed %d state %d\n", ldp->up,
673 		    ldp->timeout, ticker - ldp->timeout, ldp->state);
674 	}
675 }
676 
677 static int
is_duplicate(cfglockd_t type,pid_t pid,uint8_t seq)678 is_duplicate(cfglockd_t type, pid_t pid, uint8_t seq)
679 {
680 	struct unlock_s *bufp;
681 	int i;
682 
683 	if (!pid) {
684 		return (0);
685 	}
686 
687 	for (i = 0, bufp = unlock_buf; bufp->pid && i < MAX_UNLOCK;
688 	    i++, bufp++) {
689 		if (bufp->pid == pid && bufp->seq == seq) {
690 			/* throw message away */
691 #ifdef DEBUG
692 			spcs_log("cfglockd", NULL,
693 			    "duplicate '%d' request received from %d",
694 			    type, pid);
695 #endif
696 			return (1);
697 		}
698 	}
699 
700 	/* add it to the list */
701 	bcopy(unlock_buf, &unlock_buf[ 1 ],
702 	    sizeof (unlock_buf) - sizeof (struct unlock_s));
703 	(*unlock_buf).pid = pid;
704 	(*unlock_buf).seq = seq;
705 
706 	return (0);
707 }
708 
709 static void
local_lock(cfglockd_t type,struct lock_msg * msg,daemonaddr_t * client)710 local_lock(cfglockd_t type, struct lock_msg *msg, daemonaddr_t *client)
711 {
712 	if (is_duplicate(type, msg->pid, msg->seq)) {
713 		if (the_lock.remote_daemon == NULL &&
714 		    (the_lock.type == LOCK_WRITE ||
715 		    the_lock.type == LOCK_READ) &&
716 		    the_lock.holding_pid[0] == msg->pid) {
717 			send_lockmsg(LOCK_LOCKED, (pid_t)0, client, msg->seq);
718 		}
719 	} else {
720 		queue_lock(type, msg, client);
721 	}
722 }
723 
724 static void
remote_lock(struct sockaddr_in * remote,struct lock_msg * msg)725 remote_lock(struct sockaddr_in *remote, struct lock_msg *msg)
726 {
727 	/* make sure remote knows we are alive */
728 	send_lockmsg(ALIVE, (pid_t)0, remote, 0);
729 
730 	/* clear out pid as it is meaningless on this node */
731 	msg->pid = (pid_t)0;
732 
733 	queue_lock(LOCK_WRITE, msg, (daemonaddr_t *)remote);
734 }
735 
736 static void
unqueue_lock(daemonaddr_t * d,pid_t pid)737 unqueue_lock(daemonaddr_t *d, pid_t pid)
738 {
739 	int	i;
740 	struct lock_req *lrp, *xrp;
741 	int diff;
742 
743 	/* search queue to delete ungranted locks */
744 	for (i = 0, xrp = lrp = lock_queue; i++ < next_req; lrp++) {
745 		*xrp = *lrp;
746 		diff = 0;
747 		if (pid != (pid_t)0 && lrp->pid != pid)
748 			diff = 1;
749 		if (d != NULL && cmp_addr(d, &(lrp->remote)) != 0)
750 			diff = 1;
751 		if (!diff)
752 			continue;
753 
754 		xrp++;
755 	}
756 	next_req = xrp - lock_queue;
757 }
758 
759 static void
xxunlock()760 xxunlock()
761 {
762 	DPF((stderr, "** UNLOCK **\n"));
763 	the_lock.remote_daemon = NULL;
764 	the_lock.type = LOCK_NOTLOCKED;
765 	the_lock.nholders = 0;
766 	the_lock.state = STATE_CLEAR;
767 	process_queue();
768 }
769 
770 
771 static void
local_unlock(pid_t pid,uint8_t seq,int method)772 local_unlock(pid_t pid, uint8_t seq, int method)
773 {
774 	struct lockdaemon *ldp;
775 	int i;
776 
777 	if (method == NORMAL_UNLOCK && is_duplicate(LOCK_NOTLOCKED, pid, seq)) {
778 		return;
779 	}
780 
781 	if (the_lock.type == LOCK_READ) {
782 		/* delete reference to pid of reading process */
783 		for (i = 0; i < the_lock.nholders; i++) {
784 			if (the_lock.holding_pid[i] == pid) {
785 				DPF((stderr, "decrement lockers from %d\n",
786 				    the_lock.nholders));
787 				--the_lock.nholders;
788 				break;
789 			}
790 		}
791 		for (; i < the_lock.nholders; i++) {
792 			the_lock.holding_pid[i] = the_lock.holding_pid[i+1];
793 		}
794 		if (the_lock.nholders > 0)
795 			return;
796 	} else {
797 		/* LOCK_WRITE */
798 		if (pid != the_lock.holding_pid[0])
799 			return;
800 		the_lock.holding_pid[0] = (pid_t)0;
801 		for (i = 0, ldp = daemon_list; i < MAX_DAEMONS; i++, ldp++) {
802 			if (ldp->inuse == 0)
803 				break;
804 			if (ldp->up)
805 				send_lockmsg(UNLOCK, (pid_t)0, &(ldp->host), 0);
806 		}
807 	}
808 	xxunlock();
809 }
810 
811 static void
remote_unlock(int32_t order,daemonaddr_t * d)812 remote_unlock(int32_t order, daemonaddr_t *d)
813 {
814 	int	i;
815 	struct lock_req *lrp;
816 
817 	DPF((stderr, "remote unlock from %s ", dp_addr(d)));
818 	DPF((stderr, "when %s holds lock\n", dp_addr(&the_lock.holder)));
819 
820 	/* search queue to check for ungranted lock */
821 	for (i = 0, lrp = lock_queue; i++ < next_req; lrp++) {
822 		if (lrp->type == LOCK_WRITE &&
823 		    cmp_addr(d, &(lrp->remote)) == 0) {
824 			delete_queue_entry(lrp);
825 			return;
826 		}
827 
828 	}
829 	if (addr_is_holder(order)) {
830 		xxunlock();
831 	}
832 }
833 
834 static void
lockedby(daemonaddr_t * d,uint8_t seq)835 lockedby(daemonaddr_t *d, uint8_t seq)
836 {
837 	DPF((stderr, "lockby enquiry from %s ", dp_addr(d)));
838 	switch (the_lock.type) {
839 	case LOCK_NOTLOCKED:
840 		send_lockmsg(LOCK_NOTLOCKED, (pid_t)0, d, seq);
841 		break;
842 	case LOCK_READ:
843 		send_lockmsg(LOCK_READ, the_lock.holding_pid[0], d, seq);
844 		break;
845 	case LOCK_WRITE:
846 		send_lockmsg(LOCK_WRITE, the_lock.holding_pid[0], d, seq);
847 		break;
848 	}
849 }
850 
851 /* ARGSUSED */
852 static void
keepalive(int signo)853 keepalive(int signo)
854 {
855 	int i;
856 	struct lock_req *locker;
857 	struct lockdaemon *ldp;
858 
859 	DPF((stderr, "keepalive...\n"));
860 	ticker++;
861 
862 	/*
863 	 * tell any other daemon that has a lock request in our queue that
864 	 * this daemon is still alive.
865 	 */
866 
867 	for (i = 0, locker = lock_queue; i < next_req; i++, locker++) {
868 		if (locker->pid == 0)	/* remote lock request */
869 			send_lockmsg(ALIVE, (pid_t)0, &(locker->remote), 0);
870 	}
871 
872 	/*
873 	 * if a remote daemon holds the lock, check it is still alive and
874 	 * if the remote daemon is sent it a grant message in case the
875 	 * remote daemon missed our original grant.
876 	 */
877 
878 	if (the_lock.remote_daemon) {
879 		if (lockdaemon_dead(the_lock.remote_daemon)) {
880 			DPF((stderr, "lock owner died\n"));
881 			the_lock.remote_daemon->up = 0;
882 			xxunlock();
883 		} else {
884 			send_lockmsg(GRANTED, (pid_t)0, &the_lock.holder, 0);
885 		}
886 	}
887 
888 	/*
889 	 * check for response from daemons preventing this daemon
890 	 * from taking a write lock by not sending a grant message.
891 	 * if the remote daemon is alive send another lock request,
892 	 * otherwise mark it as dead.
893 	 * send alive message to any live remote daemons if this
894 	 * daemon has the write lock.
895 	 */
896 	if (lstate) {
897 		(void) printf("\nlock: %s\n", lockd_type(the_lock.type));
898 		(void) printf("    no. holders: %d\n", the_lock.nholders);
899 		(void) printf("    hold addr  : %s\n", the_lock.remote_daemon?
900 		    dp_addr(the_lock.remote_daemon): "0.0.0.0");
901 		(void) printf("    holding pid:");
902 		for (i = 0; i < the_lock.nholders; i++) {
903 			(void) printf(" %ld", the_lock.holding_pid[ i ]);
904 		}
905 		(void) printf("\n");
906 	}
907 	for (i = 0, ldp = daemon_list; i < MAX_DAEMONS; i++, ldp++) {
908 		if (ldp->inuse == 0)
909 			break;
910 
911 		if (lstate) {
912 			(void) printf("%-15.15s ", dp_addr(&ldp->host));
913 			(void) printf("%-4.4s ", ldp->up? "up" : "down");
914 			(void) printf("%5ld ", ldp->timeout);
915 			(void) printf("%-10.10s ", lockd_state(ldp->state));
916 			(void) printf("%6d\n", ldp->order);
917 		}
918 
919 		if (ldp->state == STATE_ASKED) {
920 			if (lockdaemon_dead(ldp)) {
921 				ldp->up = 0;
922 				ldp->state = STATE_CLEAR;
923 				continue;
924 			}
925 			send_lockmsg(WRITE_LOCK, (pid_t)0, &(ldp->host), 0);
926 			continue;
927 		}
928 		if (the_lock.type == LOCK_WRITE &&
929 		    the_lock.remote_daemon == NULL)
930 			send_lockmsg(ALIVE, (pid_t)0, &(ldp->host), 0);
931 	}
932 }
933 
934 static void
dispatch(struct lock_msg * mp,daemonaddr_t * host)935 dispatch(struct lock_msg *mp, daemonaddr_t *host)
936 {
937 	int message = mp->message;
938 	int localhost;
939 
940 	localhost = islocalhost(host);
941 	if (msgtrace && debugfile) {
942 		time_t t = time(0);
943 		if (localhost) {
944 			(void) fprintf(debugfile,
945 			    "%19.19s recv %-9.9s from %s (%ld)\n", ctime(&t),
946 			    lockd_msg(message), dp_addr(host), mp->pid);
947 		} else {
948 			(void) fprintf(debugfile,
949 			    "%19.19s recv %-9.9s from %s order %d (%ld)\n",
950 			    ctime(&t), lockd_msg(message), dp_addr(host),
951 			    mp->order, mp->pid);
952 		}
953 	}
954 	DPF((stderr, "received message %d\n", message));
955 	DPF((stderr, "from %s port %hu\n", dp_addr(host), host->sin_port));
956 	if (!localhost)
957 		daemon_alive(host, mp->order);
958 	else
959 		mp->order = order;
960 	switch (message) {
961 	case ALIVE:
962 		DPF((stderr, "received ALIVE %s\n", dp_addr(host)));
963 		/* do nothing, general "not localhost" code above does this */
964 		break;
965 	case UNLOCK:
966 		DPF((stderr, "received UNLOCK\n"));
967 		remote_unlock(mp->order, host);
968 		break;
969 	case GRANTED:
970 		DPF((stderr, "received GRANTED\n"));
971 		lock_granted(host);
972 		break;
973 	case WRITE_LOCK:
974 		DPF((stderr, "received WRITE_LOCK\n"));
975 		assert(!localhost);
976 		remote_lock(host, mp);
977 		break;
978 	case READ_LOCK:
979 	case LOCK_READ:
980 		DPF((stderr, "received READ_LOCK\n"));
981 		assert(localhost);
982 		local_lock(LOCK_READ, mp, host);
983 		break;
984 	case LOCK_WRITE:
985 		DPF((stderr, "received LOCK_WRITE\n"));
986 		assert(localhost);
987 		local_lock(LOCK_WRITE, mp, host);
988 		break;
989 	case LOCK_NOTLOCKED:
990 		DPF((stderr, "received LOCK_NOTLOCKED\n"));
991 		send_lockmsg(LOCK_ACK, (pid_t)0, host, mp->seq);
992 		if (the_lock.type != LOCK_NOTLOCKED) {
993 			local_unlock(mp->pid, mp->seq, NORMAL_UNLOCK);
994 		}
995 		break;
996 	case LOCK_LOCKEDBY:
997 		lockedby(host, mp->seq);
998 		break;
999 	case LOCK_STAT:
1000 		lock_stat();
1001 		break;
1002 	case LOCK_ACK:
1003 		/* throw message away -- this is an error to receive */
1004 		break;
1005 	}
1006 }
1007 
1008 /*
1009  * unqueue any locks asked for by pid and unlock any locks held by pid.
1010  */
1011 
1012 static void
purge_pid(pid_t pid)1013 purge_pid(pid_t pid)
1014 {
1015 	DPF((stderr, "purge locks for %ld\n", pid));
1016 	unqueue_lock(NULL, pid);
1017 	if (the_lock.type != LOCK_NOTLOCKED)
1018 		local_unlock(pid, 0, FORCE_UNLOCK);
1019 }
1020 
1021 /*
1022  * Check for exit or exec of client processes.
1023  * The lock protecting the processes pid in the lockfile will
1024  * be removed by the kernel when a client exits or execs.
1025  */
1026 
1027 static void
check_for_dead()1028 check_for_dead()
1029 {
1030 	int i, x;
1031 	pid_t pid;
1032 
1033 	for (i = 0; (x = cfg_filelock(i, 0)) != CFG_LF_EOF; i++) {
1034 		if (x == CFG_LF_AGAIN)
1035 			continue; /* can't take lock, must be still alive */
1036 		cfg_readpid(i, &pid);
1037 		cfg_writepid(i, (pid_t)0);
1038 		(void) cfg_fileunlock(i);
1039 		if (pid != (pid_t)0)
1040 			purge_pid(pid);
1041 	}
1042 }
1043 
1044 static void
build_daemon_list(char * cf_file,int exe)1045 build_daemon_list(char *cf_file, int exe)
1046 {
1047 	FILE *fp;
1048 	char	host[1024];
1049 	int	port;
1050 	int	i;
1051 	struct	hostent *hp;
1052 	struct lockdaemon *ldp;
1053 
1054 	if ((hp = gethostbyname("localhost")) == NULL) {
1055 		(void) fprintf(stderr, "%s: Can't find hostent for %s\n",
1056 		    program, "localhost");
1057 		spcs_log("cfglockd", NULL, "couldn't find localhost");
1058 		exit(1);
1059 	}
1060 
1061 	(void) memcpy(&(localhost.sin_addr.s_addr), *(hp->h_addr_list),
1062 	    sizeof (localhost.sin_addr));
1063 	if (cf_file == NULL) {
1064 		(void) endhostent();
1065 		return;
1066 	}
1067 	if (exe) {
1068 		if ((fp = popen(cf_file, "r")) == NULL) {
1069 			perror(cf_file);
1070 			(void) fprintf(stderr,
1071 			    "%s: Can't open config program\n", program);
1072 			spcs_log("cfglockd", NULL, "couldn't read config");
1073 			exit(1);
1074 		}
1075 	} else {
1076 		if ((fp = fopen(cf_file, "r")) == NULL) {
1077 			perror(cf_file);
1078 			(void) fprintf(stderr, "%s: Can't open config file\n",
1079 			    program);
1080 			spcs_log("cfglockd", NULL, "couldn't read config");
1081 			exit(1);
1082 		}
1083 	}
1084 	ldp = daemon_list;
1085 	while ((i = fscanf(fp, "%s %d\n", host, &port)) != EOF) {
1086 		if (host[0] == '#')	/* line starting with # are comments */
1087 			continue;
1088 		if (i == 1) {
1089 			port = lock_port;
1090 		} else {
1091 			if (strcmp(host, "localhost") == 0) {
1092 				lock_port = port;
1093 				continue;
1094 			}
1095 		}
1096 
1097 		if ((hp = gethostbyname(host)) == NULL) {
1098 			(void) fprintf(stderr,
1099 			    "%s: Can't find hostent for %s\n", program, host);
1100 			continue;
1101 		}
1102 
1103 		(void) memcpy(&(ldp->host.sin_addr.s_addr), *(hp->h_addr_list),
1104 		    sizeof (ldp->host.sin_addr));
1105 		DPF((stderr, "daemon: %s\t%s\n",
1106 		    inet_ntoa(ldp->host.sin_addr), hp->h_name));
1107 		if (islocalhost(&(ldp->host))) {
1108 			DPF((stderr, "is an alias for this host, skipping\n"));
1109 			continue;
1110 		}
1111 		ldp->host.sin_port = htons((short)port);
1112 		ldp->host.sin_family = hp->h_addrtype;
1113 		ldp->inuse = 1;
1114 		ldp->up = 1;
1115 		ldp++;
1116 	}
1117 	if (exe)
1118 		(void) pclose(fp);
1119 	else
1120 		(void) fclose(fp);
1121 	(void) endhostent();
1122 }
1123 
1124 static void
usage()1125 usage()
1126 {
1127 	(void) fprintf(stderr,
1128 	    gettext("usage: %s [-d] [-f file]|[-e program]\n"), program);
1129 	exit(1);
1130 }
1131 
1132 static void
unexpected(int sig)1133 unexpected(int sig)
1134 {
1135 	spcs_log("cfglockd", NULL, "pid %d unexpected signal %d, ignoring",
1136 	    getpid(), sig);
1137 }
1138 
1139 static void
term(int sig)1140 term(int sig)
1141 {
1142 	(void) unlink(CFG_PIDFILE);
1143 	spcs_log("cfglockd", NULL, "pid %d terminate on signal %d", getpid(),
1144 	    sig);
1145 	exit(0);
1146 }
1147 
1148 static void
init(int argc,char * argv[])1149 init(int argc, char *argv[])
1150 {
1151 #if defined(_SunOS_5_6) || defined(_SunOS_5_7) || defined(_SunOS_5_8)
1152 	struct rlimit rl;
1153 #endif
1154 	int	c, i, x;
1155 	int	rc;
1156 	char	*cp = NULL;
1157 	struct	itimerval	tv;
1158 	struct	timeval		tp;
1159 	socklen_t len = sizeof (thishost);
1160 	int	exe = 0;
1161 	pid_t	pid;
1162 	FILE	*fp;
1163 
1164 	lstate = (getenv("LOCKD_STATE") != NULL);
1165 	msgtrace = (getenv("LOCKD_MSG") != NULL);
1166 
1167 	/*
1168 	 * Fork off a child that becomes the daemon.
1169 	 */
1170 
1171 #ifndef TTY_MESSAGES
1172 	if ((rc = fork()) > 0)
1173 		exit(0);
1174 	else if (rc < 0) {
1175 		spcs_log("cfglockd", NULL, "can't fork %d", errno);
1176 		(void) fprintf(stderr, gettext("dscfglockd: cannot fork: %s\n"),
1177 		    strerror(errno));
1178 		exit(1);
1179 	}
1180 #endif
1181 
1182 	/*
1183 	 * In child - become daemon.
1184 	 */
1185 
1186 #if !defined(_SunOS_5_6) && !defined(_SunOS_5_7) && !defined(_SunOS_5_8)
1187 	/* use closefrom(3C) from PSARC/2000/193 when possible */
1188 	closefrom(CLOSE_FD);
1189 #else
1190 	(void) getrlimit(RLIMIT_NOFILE, &rl);
1191 	for (i = CLOSE_FD; i < rl.rlim_max; i++)
1192 		(void) close(i);
1193 #endif
1194 
1195 #ifdef DEBUG
1196 #ifndef	TTY_MESSAGES
1197 	(void) open("/dev/console", O_WRONLY|O_APPEND);
1198 	(void) dup(0);
1199 	(void) dup(0);
1200 #endif
1201 #endif
1202 	(void) close(0);
1203 
1204 	if (msgtrace || lstate) {
1205 		debugfile = fopen("/var/tmp/dscfglockd.out", "a");
1206 		if (debugfile) {
1207 			time_t t = time(0);
1208 			setbuf(debugfile, (char *)0);
1209 			(void) fprintf(debugfile, "%19.19s dscfglockd start\n",
1210 			    ctime(&t));
1211 		}
1212 	}
1213 
1214 	(void) setpgrp();
1215 	spcs_log("cfglockd", NULL, "new lock daemon, pid %d", getpid());
1216 
1217 	/*
1218 	 * Catch as unexpected all signals apart from SIGTERM.
1219 	 */
1220 
1221 	for (i = 1; i < _sys_nsig; i++)
1222 		(void) sigset(i, unexpected);
1223 	(void) sigset(SIGTERM, term);
1224 
1225 	for (i = 0; (c = getopt(argc, argv, "df:e:")) != EOF; i++) {
1226 		switch (c) {
1227 		case 'd':
1228 			debug = 1;
1229 			break;
1230 		case 'e':
1231 			exe = 1;
1232 			if (cp) {
1233 				usage();
1234 			}
1235 			cp = optarg;
1236 			break;
1237 		case 'f':
1238 			if (cp) {
1239 				usage();
1240 			}
1241 			cp = optarg;
1242 			break;
1243 		default:
1244 			usage();
1245 			break;
1246 		}
1247 	}
1248 
1249 	ifc = (struct ifconf *)malloc(sizeof (struct ifconf));
1250 	if (ifc == NULL) {
1251 		perror(CFG_PIDFILE);
1252 		DPF((stderr, "Can't open pid file\n"));
1253 		exit(1);
1254 	}
1255 	(void) memset((char *)ifc, 0, sizeof (struct ifconf));
1256 	getmyaddrs(ifc);
1257 
1258 	/*
1259 	 * if (lockdaemonalive()) {
1260 	 *	(void) fprintf(stderr, "%s: %s\n", program,
1261 	 *		gettext("There is already a live lockdaemon"));
1262 	 *	exit(1);
1263 	 * }
1264 	 */
1265 	if ((fp = fopen(CFG_PIDFILE, "w")) == NULL) {
1266 		perror(CFG_PIDFILE);
1267 		DPF((stderr, "Can't open pid file\n"));
1268 		exit(1);
1269 	}
1270 	(void) fprintf(fp, "%ld\n", getpid());
1271 	(void) fclose(fp);
1272 
1273 	/* order should be set to node number within cluster */
1274 	order = cfg_iscluster();
1275 	cfg_lfinit();
1276 
1277 	if (!order) {
1278 		(void) gettimeofday(&tp, NULL);
1279 		srand48(tp.tv_usec);
1280 		order = lrand48();
1281 		if (debugfile) {
1282 			(void) fprintf(debugfile, "WARNING: order number "
1283 			    "is 0 -- changing randomly to %d\n", order);
1284 		}
1285 	}
1286 	c = 0;
1287 	for (i = 0; (x = cfg_filelock(i, 0)) != CFG_LF_EOF; i++) {
1288 		if (x == CFG_LF_AGAIN) {
1289 			cfg_readpid(i, &pid);
1290 			if (c++ == 0)
1291 				spcs_log("cfglockd", NULL,
1292 				    "init .dscfg.lck slot %d pid %d locked",
1293 				    i, pid);
1294 			DPF((stderr, "client process %ld still alive\n", pid));
1295 			continue; /* can't take lock, must be still alive */
1296 		}
1297 		cfg_writepid(i, 0);
1298 		(void) cfg_fileunlock(i);
1299 	}
1300 
1301 	tv.it_interval.tv_sec = TIMEOUT_SECS;
1302 	tv.it_interval.tv_usec = 0;
1303 	tv.it_value = tv.it_interval;
1304 
1305 	bzero(unlock_buf, sizeof (unlock_buf));
1306 	next_q = 0;
1307 	build_daemon_list(cp, exe);
1308 	if ((lock_soc = socket(pf_inet, SOCK_DGRAM, 0)) < 0) {
1309 		(void) fprintf(stderr, "%s: %s\n", program,
1310 		    gettext("failed to create socket"));
1311 		perror("socket");
1312 		spcs_log("cfglockd", NULL, "couldn't create socket");
1313 		exit(1);
1314 	}
1315 	thishost.sin_family = AF_INET;
1316 	thishost.sin_addr.s_addr = INADDR_ANY;
1317 	thishost.sin_port = htons(lock_port);
1318 	rc = bind(lock_soc, (struct sockaddr *)&thishost, sizeof (thishost));
1319 	if (rc < 0) {
1320 		perror("bind");
1321 		spcs_log("cfglockd", NULL, "couldn't bind");
1322 		exit(1);
1323 	}
1324 	if (getsockname(lock_soc, (struct sockaddr *)&thishost, &len) < 0)
1325 		perror("getsockname");
1326 	send_aliveall();
1327 	(void) sigset(SIGALRM, keepalive);
1328 	(void) setitimer(ITIMER_REAL, &tv, NULL);
1329 	/*
1330 	 * wait 2 time outs before allowing a lock to find if someone else
1331 	 * currently has the lock.
1332 	 */
1333 }
1334 
1335 #ifdef lint
1336 int
lintmain(int argc,char * argv[])1337 lintmain(int argc, char *argv[])
1338 #else
1339 int
1340 main(int argc, char *argv[])
1341 #endif
1342 {
1343 	struct lock_msg message_buf;
1344 	daemonaddr_t from;
1345 	int addrlen;
1346 	int rc;
1347 	int x = 1;		/* kludge to stop warnings from compiler */
1348 
1349 	init(argc, argv);
1350 	CRIT_BEGIN();
1351 	while (x) {
1352 		CRIT_END();
1353 		addrlen = sizeof (from);
1354 		DPF((stderr, "begin recvfrom\n"));
1355 		rc = recvfrom(lock_soc, &message_buf, sizeof (message_buf),
1356 		    0, (struct sockaddr *)&from, &addrlen);
1357 		DPF((stderr, "end recvfrom rc = %d\n", rc));
1358 		CRIT_BEGIN();
1359 		if (rc == sizeof (message_buf))
1360 			dispatch(&message_buf, &from);
1361 		else
1362 			check_for_write_lock();
1363 
1364 		/* if we own the lock, check to see if the process died */
1365 		if (the_lock.type != LOCK_NOTLOCKED &&
1366 		    the_lock.remote_daemon == NULL)
1367 			check_for_dead();
1368 	}
1369 	CRIT_END();
1370 	return (0);
1371 }
1372