xref: /freebsd/sys/netsmb/smb_iod.c (revision 3f0164abf32b9b761e0a2cb4bdca3a8b84f156d4)
1 /*-
2  * Copyright (c) 2000-2001 Boris Popov
3  * All rights reserved.
4  *
5  * Redistribution and use in source and binary forms, with or without
6  * modification, are permitted provided that the following conditions
7  * are met:
8  * 1. Redistributions of source code must retain the above copyright
9  *    notice, this list of conditions and the following disclaimer.
10  * 2. Redistributions in binary form must reproduce the above copyright
11  *    notice, this list of conditions and the following disclaimer in the
12  *    documentation and/or other materials provided with the distribution.
13  *
14  * THIS SOFTWARE IS PROVIDED BY THE AUTHOR AND CONTRIBUTORS ``AS IS'' AND
15  * ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE
16  * IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE
17  * ARE DISCLAIMED.  IN NO EVENT SHALL THE AUTHOR OR CONTRIBUTORS BE LIABLE
18  * FOR ANY DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL
19  * DAMAGES (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS
20  * OR SERVICES; LOSS OF USE, DATA, OR PROFITS; OR BUSINESS INTERRUPTION)
21  * HOWEVER CAUSED AND ON ANY THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT
22  * LIABILITY, OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY
23  * OUT OF THE USE OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF
24  * SUCH DAMAGE.
25  */
26 
27 #include <sys/cdefs.h>
28 __FBSDID("$FreeBSD$");
29 
30 #include <sys/param.h>
31 #include <sys/systm.h>
32 #include <sys/endian.h>
33 #include <sys/proc.h>
34 #include <sys/kernel.h>
35 #include <sys/kthread.h>
36 #include <sys/malloc.h>
37 #include <sys/mbuf.h>
38 #include <sys/unistd.h>
39 
40 #include <netsmb/smb.h>
41 #include <netsmb/smb_conn.h>
42 #include <netsmb/smb_rq.h>
43 #include <netsmb/smb_tran.h>
44 #include <netsmb/smb_trantcp.h>
45 
46 
47 #define SMBIOD_SLEEP_TIMO	2
48 #define	SMBIOD_PING_TIMO	60	/* seconds */
49 
50 #define	SMB_IOD_EVLOCKPTR(iod)	(&((iod)->iod_evlock))
51 #define	SMB_IOD_EVLOCK(iod)	smb_sl_lock(&((iod)->iod_evlock))
52 #define	SMB_IOD_EVUNLOCK(iod)	smb_sl_unlock(&((iod)->iod_evlock))
53 
54 #define	SMB_IOD_RQLOCKPTR(iod)	(&((iod)->iod_rqlock))
55 #define	SMB_IOD_RQLOCK(iod)	smb_sl_lock(&((iod)->iod_rqlock))
56 #define	SMB_IOD_RQUNLOCK(iod)	smb_sl_unlock(&((iod)->iod_rqlock))
57 
58 #define	smb_iod_wakeup(iod)	wakeup(&(iod)->iod_flags)
59 
60 
61 static MALLOC_DEFINE(M_SMBIOD, "SMBIOD", "SMB network io daemon");
62 
63 static int smb_iod_next;
64 
65 static int  smb_iod_sendall(struct smbiod *iod);
66 static int  smb_iod_disconnect(struct smbiod *iod);
67 static void smb_iod_thread(void *);
68 
69 static __inline void
70 smb_iod_rqprocessed(struct smb_rq *rqp, int error)
71 {
72 	SMBRQ_SLOCK(rqp);
73 	rqp->sr_lerror = error;
74 	rqp->sr_rpgen++;
75 	rqp->sr_state = SMBRQ_NOTIFIED;
76 	wakeup(&rqp->sr_state);
77 	SMBRQ_SUNLOCK(rqp);
78 }
79 
80 static void
81 smb_iod_invrq(struct smbiod *iod)
82 {
83 	struct smb_rq *rqp;
84 
85 	/*
86 	 * Invalidate all outstanding requests for this connection
87 	 */
88 	SMB_IOD_RQLOCK(iod);
89 	TAILQ_FOREACH(rqp, &iod->iod_rqlist, sr_link) {
90 		if (rqp->sr_flags & SMBR_INTERNAL)
91 			SMBRQ_SUNLOCK(rqp);
92 		rqp->sr_flags |= SMBR_RESTART;
93 		smb_iod_rqprocessed(rqp, ENOTCONN);
94 	}
95 	SMB_IOD_RQUNLOCK(iod);
96 }
97 
98 static void
99 smb_iod_closetran(struct smbiod *iod)
100 {
101 	struct smb_vc *vcp = iod->iod_vc;
102 	struct thread *td = iod->iod_td;
103 
104 	if (vcp->vc_tdata == NULL)
105 		return;
106 	SMB_TRAN_DISCONNECT(vcp, td);
107 	SMB_TRAN_DONE(vcp, td);
108 	vcp->vc_tdata = NULL;
109 }
110 
111 static void
112 smb_iod_dead(struct smbiod *iod)
113 {
114 	iod->iod_state = SMBIOD_ST_DEAD;
115 	smb_iod_closetran(iod);
116 	smb_iod_invrq(iod);
117 }
118 
119 static int
120 smb_iod_connect(struct smbiod *iod)
121 {
122 	struct smb_vc *vcp = iod->iod_vc;
123 	struct thread *td = iod->iod_td;
124 	int error;
125 
126 	SMBIODEBUG("%d\n", iod->iod_state);
127 	switch(iod->iod_state) {
128 	    case SMBIOD_ST_VCACTIVE:
129 		SMBERROR("called for already opened connection\n");
130 		return EISCONN;
131 	    case SMBIOD_ST_DEAD:
132 		return ENOTCONN;	/* XXX: last error code ? */
133 	    default:
134 		break;
135 	}
136 	vcp->vc_genid++;
137 	error = 0;
138 
139 	error = (int)SMB_TRAN_CREATE(vcp, td);
140 	if (error)
141 		goto fail;
142 	SMBIODEBUG("tcreate\n");
143 	if (vcp->vc_laddr) {
144 		error = (int)SMB_TRAN_BIND(vcp, vcp->vc_laddr, td);
145 		if (error)
146 			goto fail;
147 	}
148 	SMBIODEBUG("tbind\n");
149 	error = (int)SMB_TRAN_CONNECT(vcp, vcp->vc_paddr, td);
150 	if (error)
151 		goto fail;
152 	SMB_TRAN_SETPARAM(vcp, SMBTP_SELECTID, &iod->iod_flags);
153 	iod->iod_state = SMBIOD_ST_TRANACTIVE;
154 	SMBIODEBUG("tconnect\n");
155 	/* vcp->vc_mid = 0;*/
156 	error = (int)smb_smb_negotiate(vcp, &iod->iod_scred);
157 	if (error)
158 		goto fail;
159 	SMBIODEBUG("snegotiate\n");
160 	error = (int)smb_smb_ssnsetup(vcp, &iod->iod_scred);
161 	if (error)
162 		goto fail;
163 	iod->iod_state = SMBIOD_ST_VCACTIVE;
164 	SMBIODEBUG("completed\n");
165 	smb_iod_invrq(iod);
166 	return (0);
167 
168  fail:
169 	smb_iod_dead(iod);
170 	return (error);
171 }
172 
173 static int
174 smb_iod_disconnect(struct smbiod *iod)
175 {
176 	struct smb_vc *vcp = iod->iod_vc;
177 
178 	SMBIODEBUG("\n");
179 	if (iod->iod_state == SMBIOD_ST_VCACTIVE) {
180 		smb_smb_ssnclose(vcp, &iod->iod_scred);
181 		iod->iod_state = SMBIOD_ST_TRANACTIVE;
182 	}
183 	vcp->vc_smbuid = SMB_UID_UNKNOWN;
184 	smb_iod_closetran(iod);
185 	iod->iod_state = SMBIOD_ST_NOTCONN;
186 	return 0;
187 }
188 
189 static int
190 smb_iod_treeconnect(struct smbiod *iod, struct smb_share *ssp)
191 {
192 	int error;
193 
194 	if (iod->iod_state != SMBIOD_ST_VCACTIVE) {
195 		if (iod->iod_state != SMBIOD_ST_DEAD)
196 			return ENOTCONN;
197 		iod->iod_state = SMBIOD_ST_RECONNECT;
198 		error = smb_iod_connect(iod);
199 		if (error)
200 			return error;
201 	}
202 	SMBIODEBUG("tree reconnect\n");
203 	SMBS_ST_LOCK(ssp);
204 	ssp->ss_flags |= SMBS_RECONNECTING;
205 	SMBS_ST_UNLOCK(ssp);
206 	error = smb_smb_treeconnect(ssp, &iod->iod_scred);
207 	SMBS_ST_LOCK(ssp);
208 	ssp->ss_flags &= ~SMBS_RECONNECTING;
209 	SMBS_ST_UNLOCK(ssp);
210 	wakeup(&ssp->ss_vcgenid);
211 	return error;
212 }
213 
214 static int
215 smb_iod_sendrq(struct smbiod *iod, struct smb_rq *rqp)
216 {
217 	struct thread *td = iod->iod_td;
218 	struct smb_vc *vcp = iod->iod_vc;
219 	struct smb_share *ssp = rqp->sr_share;
220 	struct mbuf *m;
221 	int error;
222 
223 	SMBIODEBUG("iod_state = %d\n", iod->iod_state);
224 	switch (iod->iod_state) {
225 	    case SMBIOD_ST_NOTCONN:
226 		smb_iod_rqprocessed(rqp, ENOTCONN);
227 		return 0;
228 	    case SMBIOD_ST_DEAD:
229 		iod->iod_state = SMBIOD_ST_RECONNECT;
230 		return 0;
231 	    case SMBIOD_ST_RECONNECT:
232 		return 0;
233 	    default:
234 		break;
235 	}
236 	if (rqp->sr_sendcnt == 0) {
237 #ifdef movedtoanotherplace
238 		if (vcp->vc_maxmux != 0 && iod->iod_muxcnt >= vcp->vc_maxmux)
239 			return 0;
240 #endif
241 		le16enc(rqp->sr_rqtid, ssp ? ssp->ss_tid : SMB_TID_UNKNOWN);
242 		le16enc(rqp->sr_rquid, vcp ? vcp->vc_smbuid : 0);
243 		mb_fixhdr(&rqp->sr_rq);
244 		if (vcp->vc_hflags2 & SMB_FLAGS2_SECURITY_SIGNATURE)
245 			smb_rq_sign(rqp);
246 	}
247 	if (rqp->sr_sendcnt++ > 5) {
248 		rqp->sr_flags |= SMBR_RESTART;
249 		smb_iod_rqprocessed(rqp, rqp->sr_lerror);
250 		/*
251 		 * If all attempts to send a request failed, then
252 		 * something is seriously hosed.
253 		 */
254 		return ENOTCONN;
255 	}
256 	SMBSDEBUG("M:%04x, P:%04x, U:%04x, T:%04x\n", rqp->sr_mid, 0, 0, 0);
257 	m_dumpm(rqp->sr_rq.mb_top);
258 	m = m_copym(rqp->sr_rq.mb_top, 0, M_COPYALL, M_WAITOK);
259 	error = rqp->sr_lerror = SMB_TRAN_SEND(vcp, m, td);
260 	if (error == 0) {
261 		getnanotime(&rqp->sr_timesent);
262 		iod->iod_lastrqsent = rqp->sr_timesent;
263 		rqp->sr_flags |= SMBR_SENT;
264 		rqp->sr_state = SMBRQ_SENT;
265 		return 0;
266 	}
267 	/*
268 	 * Check for fatal errors
269 	 */
270 	if (SMB_TRAN_FATAL(vcp, error)) {
271 		/*
272 		 * No further attempts should be made
273 		 */
274 		return ENOTCONN;
275 	}
276 	if (smb_rq_intr(rqp))
277 		smb_iod_rqprocessed(rqp, EINTR);
278 	return 0;
279 }
280 
281 /*
282  * Process incoming packets
283  */
284 static int
285 smb_iod_recvall(struct smbiod *iod)
286 {
287 	struct smb_vc *vcp = iod->iod_vc;
288 	struct thread *td = iod->iod_td;
289 	struct smb_rq *rqp;
290 	struct mbuf *m;
291 	u_char *hp;
292 	u_short mid;
293 	int error;
294 
295 	switch (iod->iod_state) {
296 	    case SMBIOD_ST_NOTCONN:
297 	    case SMBIOD_ST_DEAD:
298 	    case SMBIOD_ST_RECONNECT:
299 		return 0;
300 	    default:
301 		break;
302 	}
303 	for (;;) {
304 		m = NULL;
305 		error = SMB_TRAN_RECV(vcp, &m, td);
306 		if (error == EWOULDBLOCK)
307 			break;
308 		if (SMB_TRAN_FATAL(vcp, error)) {
309 			smb_iod_dead(iod);
310 			break;
311 		}
312 		if (error)
313 			break;
314 		if (m == NULL) {
315 			SMBERROR("tran return NULL without error\n");
316 			error = EPIPE;
317 			continue;
318 		}
319 		m = m_pullup(m, SMB_HDRLEN);
320 		if (m == NULL)
321 			continue;	/* wait for a good packet */
322 		/*
323 		 * Now we got an entire and possibly invalid SMB packet.
324 		 * Be careful while parsing it.
325 		 */
326 		m_dumpm(m);
327 		hp = mtod(m, u_char*);
328 		if (bcmp(hp, SMB_SIGNATURE, SMB_SIGLEN) != 0) {
329 			m_freem(m);
330 			continue;
331 		}
332 		mid = SMB_HDRMID(hp);
333 		SMBSDEBUG("mid %04x\n", (u_int)mid);
334 		SMB_IOD_RQLOCK(iod);
335 		TAILQ_FOREACH(rqp, &iod->iod_rqlist, sr_link) {
336 			if (rqp->sr_mid != mid)
337 				continue;
338 			SMBRQ_SLOCK(rqp);
339 			if (rqp->sr_rp.md_top == NULL) {
340 				md_initm(&rqp->sr_rp, m);
341 			} else {
342 				if (rqp->sr_flags & SMBR_MULTIPACKET) {
343 					md_append_record(&rqp->sr_rp, m);
344 				} else {
345 					SMBRQ_SUNLOCK(rqp);
346 					SMBERROR("duplicate response %d (ignored)\n", mid);
347 					break;
348 				}
349 			}
350 			SMBRQ_SUNLOCK(rqp);
351 			smb_iod_rqprocessed(rqp, 0);
352 			break;
353 		}
354 		SMB_IOD_RQUNLOCK(iod);
355 		if (rqp == NULL) {
356 			SMBERROR("drop resp with mid %d\n", (u_int)mid);
357 /*			smb_printrqlist(vcp);*/
358 			m_freem(m);
359 		}
360 	}
361 	/*
362 	 * check for interrupts
363 	 */
364 	SMB_IOD_RQLOCK(iod);
365 	TAILQ_FOREACH(rqp, &iod->iod_rqlist, sr_link) {
366 		if (smb_td_intr(rqp->sr_cred->scr_td)) {
367 			smb_iod_rqprocessed(rqp, EINTR);
368 		}
369 	}
370 	SMB_IOD_RQUNLOCK(iod);
371 	return 0;
372 }
373 
374 int
375 smb_iod_request(struct smbiod *iod, int event, void *ident)
376 {
377 	struct smbiod_event *evp;
378 	int error;
379 
380 	SMBIODEBUG("\n");
381 	evp = smb_zmalloc(sizeof(*evp), M_SMBIOD, M_WAITOK);
382 	evp->ev_type = event;
383 	evp->ev_ident = ident;
384 	SMB_IOD_EVLOCK(iod);
385 	STAILQ_INSERT_TAIL(&iod->iod_evlist, evp, ev_link);
386 	if ((event & SMBIOD_EV_SYNC) == 0) {
387 		SMB_IOD_EVUNLOCK(iod);
388 		smb_iod_wakeup(iod);
389 		return 0;
390 	}
391 	smb_iod_wakeup(iod);
392 	msleep(evp, SMB_IOD_EVLOCKPTR(iod), PWAIT | PDROP, "90evw", 0);
393 	error = evp->ev_error;
394 	free(evp, M_SMBIOD);
395 	return error;
396 }
397 
398 /*
399  * Place request in the queue.
400  * Request from smbiod have a high priority.
401  */
402 int
403 smb_iod_addrq(struct smb_rq *rqp)
404 {
405 	struct smb_vc *vcp = rqp->sr_vc;
406 	struct smbiod *iod = vcp->vc_iod;
407 	int error;
408 
409 	SMBIODEBUG("\n");
410 	if (rqp->sr_cred->scr_td != NULL &&
411 	    rqp->sr_cred->scr_td->td_proc == iod->iod_p) {
412 		rqp->sr_flags |= SMBR_INTERNAL;
413 		SMB_IOD_RQLOCK(iod);
414 		TAILQ_INSERT_HEAD(&iod->iod_rqlist, rqp, sr_link);
415 		SMB_IOD_RQUNLOCK(iod);
416 		for (;;) {
417 			if (smb_iod_sendrq(iod, rqp) != 0) {
418 				smb_iod_dead(iod);
419 				break;
420 			}
421 			/*
422 			 * we don't need to lock state field here
423 			 */
424 			if (rqp->sr_state != SMBRQ_NOTSENT)
425 				break;
426 			tsleep(&iod->iod_flags, PWAIT, "90sndw", hz);
427 		}
428 		if (rqp->sr_lerror)
429 			smb_iod_removerq(rqp);
430 		return rqp->sr_lerror;
431 	}
432 
433 	switch (iod->iod_state) {
434 	    case SMBIOD_ST_NOTCONN:
435 		return ENOTCONN;
436 	    case SMBIOD_ST_DEAD:
437 		error = smb_iod_request(vcp->vc_iod, SMBIOD_EV_CONNECT | SMBIOD_EV_SYNC, NULL);
438 		if (error)
439 			return error;
440 		return EXDEV;
441 	    default:
442 		break;
443 	}
444 
445 	SMB_IOD_RQLOCK(iod);
446 	for (;;) {
447 		if (vcp->vc_maxmux == 0) {
448 			SMBERROR("maxmux == 0\n");
449 			break;
450 		}
451 		if (iod->iod_muxcnt < vcp->vc_maxmux)
452 			break;
453 		iod->iod_muxwant++;
454 		msleep(&iod->iod_muxwant, SMB_IOD_RQLOCKPTR(iod),
455 		    PWAIT, "90mux", 0);
456 	}
457 	iod->iod_muxcnt++;
458 	TAILQ_INSERT_TAIL(&iod->iod_rqlist, rqp, sr_link);
459 	SMB_IOD_RQUNLOCK(iod);
460 	smb_iod_wakeup(iod);
461 	return 0;
462 }
463 
464 int
465 smb_iod_removerq(struct smb_rq *rqp)
466 {
467 	struct smb_vc *vcp = rqp->sr_vc;
468 	struct smbiod *iod = vcp->vc_iod;
469 
470 	SMBIODEBUG("\n");
471 	if (rqp->sr_flags & SMBR_INTERNAL) {
472 		SMB_IOD_RQLOCK(iod);
473 		TAILQ_REMOVE(&iod->iod_rqlist, rqp, sr_link);
474 		SMB_IOD_RQUNLOCK(iod);
475 		return 0;
476 	}
477 	SMB_IOD_RQLOCK(iod);
478 	while (rqp->sr_flags & SMBR_XLOCK) {
479 		rqp->sr_flags |= SMBR_XLOCKWANT;
480 		msleep(rqp, SMB_IOD_RQLOCKPTR(iod), PWAIT, "90xrm", 0);
481 	}
482 	TAILQ_REMOVE(&iod->iod_rqlist, rqp, sr_link);
483 	iod->iod_muxcnt--;
484 	if (iod->iod_muxwant) {
485 		iod->iod_muxwant--;
486 		wakeup(&iod->iod_muxwant);
487 	}
488 	SMB_IOD_RQUNLOCK(iod);
489 	return 0;
490 }
491 
492 int
493 smb_iod_waitrq(struct smb_rq *rqp)
494 {
495 	struct smbiod *iod = rqp->sr_vc->vc_iod;
496 	int error;
497 
498 	SMBIODEBUG("\n");
499 	if (rqp->sr_flags & SMBR_INTERNAL) {
500 		for (;;) {
501 			smb_iod_sendall(iod);
502 			smb_iod_recvall(iod);
503 			if (rqp->sr_rpgen != rqp->sr_rplast)
504 				break;
505 			tsleep(&iod->iod_flags, PWAIT, "90irq", hz);
506 		}
507 		smb_iod_removerq(rqp);
508 		return rqp->sr_lerror;
509 
510 	}
511 	SMBRQ_SLOCK(rqp);
512 	if (rqp->sr_rpgen == rqp->sr_rplast)
513 		msleep(&rqp->sr_state, SMBRQ_SLOCKPTR(rqp), PWAIT, "90wrq", 0);
514 	rqp->sr_rplast++;
515 	SMBRQ_SUNLOCK(rqp);
516 	error = rqp->sr_lerror;
517 	if (rqp->sr_flags & SMBR_MULTIPACKET) {
518 		/*
519 		 * If request should stay in the list, then reinsert it
520 		 * at the end of queue so other waiters have chance to concur
521 		 */
522 		SMB_IOD_RQLOCK(iod);
523 		TAILQ_REMOVE(&iod->iod_rqlist, rqp, sr_link);
524 		TAILQ_INSERT_TAIL(&iod->iod_rqlist, rqp, sr_link);
525 		SMB_IOD_RQUNLOCK(iod);
526 	} else
527 		smb_iod_removerq(rqp);
528 	return error;
529 }
530 
531 
532 static int
533 smb_iod_sendall(struct smbiod *iod)
534 {
535 	struct smb_vc *vcp = iod->iod_vc;
536 	struct smb_rq *rqp;
537 	struct timespec ts, tstimeout;
538 	int herror;
539 
540 	herror = 0;
541 	/*
542 	 * Loop through the list of requests and send them if possible
543 	 */
544 	SMB_IOD_RQLOCK(iod);
545 	TAILQ_FOREACH(rqp, &iod->iod_rqlist, sr_link) {
546 		switch (rqp->sr_state) {
547 		    case SMBRQ_NOTSENT:
548 			rqp->sr_flags |= SMBR_XLOCK;
549 			SMB_IOD_RQUNLOCK(iod);
550 			herror = smb_iod_sendrq(iod, rqp);
551 			SMB_IOD_RQLOCK(iod);
552 			rqp->sr_flags &= ~SMBR_XLOCK;
553 			if (rqp->sr_flags & SMBR_XLOCKWANT) {
554 				rqp->sr_flags &= ~SMBR_XLOCKWANT;
555 				wakeup(rqp);
556 			}
557 			break;
558 		    case SMBRQ_SENT:
559 			SMB_TRAN_GETPARAM(vcp, SMBTP_TIMEOUT, &tstimeout);
560 			timespecadd(&tstimeout, &tstimeout);
561 			getnanotime(&ts);
562 			timespecsub(&ts, &tstimeout);
563 			if (timespeccmp(&ts, &rqp->sr_timesent, >)) {
564 				smb_iod_rqprocessed(rqp, ETIMEDOUT);
565 			}
566 			break;
567 		    default:
568 			break;
569 		}
570 		if (herror)
571 			break;
572 	}
573 	SMB_IOD_RQUNLOCK(iod);
574 	if (herror == ENOTCONN)
575 		smb_iod_dead(iod);
576 	return 0;
577 }
578 
579 /*
580  * "main" function for smbiod daemon
581  */
582 static __inline void
583 smb_iod_main(struct smbiod *iod)
584 {
585 /*	struct smb_vc *vcp = iod->iod_vc;*/
586 	struct smbiod_event *evp;
587 /*	struct timespec tsnow;*/
588 	int error;
589 
590 	SMBIODEBUG("\n");
591 	error = 0;
592 
593 	/*
594 	 * Check all interesting events
595 	 */
596 	for (;;) {
597 		SMB_IOD_EVLOCK(iod);
598 		evp = STAILQ_FIRST(&iod->iod_evlist);
599 		if (evp == NULL) {
600 			SMB_IOD_EVUNLOCK(iod);
601 			break;
602 		}
603 		STAILQ_REMOVE_HEAD(&iod->iod_evlist, ev_link);
604 		evp->ev_type |= SMBIOD_EV_PROCESSING;
605 		SMB_IOD_EVUNLOCK(iod);
606 		switch (evp->ev_type & SMBIOD_EV_MASK) {
607 		    case SMBIOD_EV_CONNECT:
608 			iod->iod_state = SMBIOD_ST_RECONNECT;
609 			evp->ev_error = smb_iod_connect(iod);
610 			break;
611 		    case SMBIOD_EV_DISCONNECT:
612 			evp->ev_error = smb_iod_disconnect(iod);
613 			break;
614 		    case SMBIOD_EV_TREECONNECT:
615 			evp->ev_error = smb_iod_treeconnect(iod, evp->ev_ident);
616 			break;
617 		    case SMBIOD_EV_SHUTDOWN:
618 			iod->iod_flags |= SMBIOD_SHUTDOWN;
619 			break;
620 		    case SMBIOD_EV_NEWRQ:
621 			break;
622 		}
623 		if (evp->ev_type & SMBIOD_EV_SYNC) {
624 			SMB_IOD_EVLOCK(iod);
625 			wakeup(evp);
626 			SMB_IOD_EVUNLOCK(iod);
627 		} else
628 			free(evp, M_SMBIOD);
629 	}
630 #if 0
631 	if (iod->iod_state == SMBIOD_ST_VCACTIVE) {
632 		getnanotime(&tsnow);
633 		timespecsub(&tsnow, &iod->iod_pingtimo);
634 		if (timespeccmp(&tsnow, &iod->iod_lastrqsent, >)) {
635 			smb_smb_echo(vcp, &iod->iod_scred);
636 		}
637 	}
638 #endif
639 	smb_iod_sendall(iod);
640 	smb_iod_recvall(iod);
641 	return;
642 }
643 
644 void
645 smb_iod_thread(void *arg)
646 {
647 	struct smbiod *iod = arg;
648 
649 	mtx_lock(&Giant);
650 
651 	/*
652 	 * Here we assume that the thread structure will be the same
653 	 * for an entire kthread (kproc, to be more precise) life.
654 	 */
655 	iod->iod_td = curthread;
656 	smb_makescred(&iod->iod_scred, iod->iod_td, NULL);
657 	while ((iod->iod_flags & SMBIOD_SHUTDOWN) == 0) {
658 		smb_iod_main(iod);
659 		SMBIODEBUG("going to sleep for %d ticks\n", iod->iod_sleeptimo);
660 		if (iod->iod_flags & SMBIOD_SHUTDOWN)
661 			break;
662 		tsleep(&iod->iod_flags, PWAIT, "90idle", iod->iod_sleeptimo);
663 	}
664 	mtx_unlock(&Giant);
665 	kproc_exit(0);
666 }
667 
668 int
669 smb_iod_create(struct smb_vc *vcp)
670 {
671 	struct smbiod *iod;
672 	int error;
673 
674 	iod = smb_zmalloc(sizeof(*iod), M_SMBIOD, M_WAITOK);
675 	iod->iod_id = smb_iod_next++;
676 	iod->iod_state = SMBIOD_ST_NOTCONN;
677 	iod->iod_vc = vcp;
678 	iod->iod_sleeptimo = hz * SMBIOD_SLEEP_TIMO;
679 	iod->iod_pingtimo.tv_sec = SMBIOD_PING_TIMO;
680 	getnanotime(&iod->iod_lastrqsent);
681 	vcp->vc_iod = iod;
682 	smb_sl_init(&iod->iod_rqlock, "90rql");
683 	TAILQ_INIT(&iod->iod_rqlist);
684 	smb_sl_init(&iod->iod_evlock, "90evl");
685 	STAILQ_INIT(&iod->iod_evlist);
686 	error = kproc_create(smb_iod_thread, iod, &iod->iod_p,
687 	    RFNOWAIT, 0, "smbiod%d", iod->iod_id);
688 	if (error) {
689 		SMBERROR("can't start smbiod: %d", error);
690 		free(iod, M_SMBIOD);
691 		return error;
692 	}
693 	return 0;
694 }
695 
696 int
697 smb_iod_destroy(struct smbiod *iod)
698 {
699 	smb_iod_request(iod, SMBIOD_EV_SHUTDOWN | SMBIOD_EV_SYNC, NULL);
700 	smb_sl_destroy(&iod->iod_rqlock);
701 	smb_sl_destroy(&iod->iod_evlock);
702 	free(iod, M_SMBIOD);
703 	return 0;
704 }
705 
706 int
707 smb_iod_init(void)
708 {
709 	return 0;
710 }
711 
712 int
713 smb_iod_done(void)
714 {
715 	return 0;
716 }
717 
718