xref: /titanic_50/usr/src/uts/common/fs/smbsrv/smb_kshare.c (revision a307732568c3d861c38b0342ae32434226d10e94)
1 /*
2  * CDDL HEADER START
3  *
4  * The contents of this file are subject to the terms of the
5  * Common Development and Distribution License (the "License").
6  * You may not use this file except in compliance with the License.
7  *
8  * You can obtain a copy of the license at usr/src/OPENSOLARIS.LICENSE
9  * or http://www.opensolaris.org/os/licensing.
10  * See the License for the specific language governing permissions
11  * and limitations under the License.
12  *
13  * When distributing Covered Code, include this CDDL HEADER in each
14  * file and include the License file at usr/src/OPENSOLARIS.LICENSE.
15  * If applicable, add the following below this CDDL HEADER, with the
16  * fields enclosed by brackets "[]" replaced with your own identifying
17  * information: Portions Copyright [yyyy] [name of copyright owner]
18  *
19  * CDDL HEADER END
20  */
21 
22 /*
23  * Copyright (c) 2007, 2010, Oracle and/or its affiliates. All rights reserved.
24  */
25 
26 #include <smbsrv/smb_door.h>
27 #include <smbsrv/smb_kproto.h>
28 #include <smbsrv/smb_ktypes.h>
29 
30 typedef struct smb_unshare {
31 	list_node_t	us_lnd;
32 	char		us_sharename[MAXNAMELEN];
33 } smb_unshare_t;
34 
35 static smb_export_t smb_export;
36 
37 static int smb_kshare_cmp(const void *, const void *);
38 static void smb_kshare_hold(const void *);
39 static boolean_t smb_kshare_rele(const void *);
40 static void smb_kshare_destroy(void *);
41 static char *smb_kshare_oemname(const char *);
42 static int smb_kshare_is_special(const char *);
43 static boolean_t smb_kshare_is_admin(const char *);
44 static smb_kshare_t *smb_kshare_decode(nvlist_t *);
45 static uint32_t smb_kshare_decode_bool(nvlist_t *, const char *, uint32_t);
46 static void smb_kshare_unexport_thread(smb_thread_t *, void *);
47 static int smb_kshare_export(smb_kshare_t *);
48 static int smb_kshare_unexport(const char *);
49 static int smb_kshare_export_trans(char *, char *, char *);
50 static void smb_kshare_csc_flags(smb_kshare_t *, const char *);
51 
52 static boolean_t smb_export_isready(void);
53 
54 static int smb_kshare_chk_dsrv_status(int, smb_dr_ctx_t *);
55 
56 static smb_avl_nops_t smb_kshare_avlops = {
57 	smb_kshare_cmp,
58 	smb_kshare_hold,
59 	smb_kshare_rele,
60 	smb_kshare_destroy
61 };
62 
63 /*
64  * This function is not MultiThread safe. The caller has to make sure only one
65  * thread calls this function.
66  */
67 door_handle_t
68 smb_kshare_door_init(int door_id)
69 {
70 	return (door_ki_lookup(door_id));
71 }
72 
73 /*
74  * This function is not MultiThread safe. The caller has to make sure only one
75  * thread calls this function.
76  */
77 void
78 smb_kshare_door_fini(door_handle_t dhdl)
79 {
80 	if (dhdl)
81 		door_ki_rele(dhdl);
82 }
83 
84 /*
85  * This is a special interface that will be utilized by ZFS to cause
86  * a share to be added/removed
87  *
88  * arg is either a smb_share_t or share_name from userspace.
89  * It will need to be copied into the kernel.   It is smb_share_t
90  * for add operations and share_name for delete operations.
91  */
92 int
93 smb_kshare_upcall(door_handle_t dhdl, void *arg, boolean_t add_share)
94 {
95 	door_arg_t	doorarg = { 0 };
96 	char		*buf = NULL;
97 	char		*str = NULL;
98 	int		error;
99 	int		rc;
100 	unsigned int	used;
101 	smb_dr_ctx_t	*dec_ctx;
102 	smb_dr_ctx_t	*enc_ctx;
103 	smb_share_t	*lmshare = NULL;
104 	int		opcode;
105 
106 	opcode = (add_share) ? SMB_SHROP_ADD : SMB_SHROP_DELETE;
107 
108 	buf = kmem_alloc(SMB_SHARE_DSIZE, KM_SLEEP);
109 	enc_ctx = smb_dr_encode_start(buf, SMB_SHARE_DSIZE);
110 	smb_dr_put_uint32(enc_ctx, opcode);
111 
112 	switch (opcode) {
113 	case SMB_SHROP_ADD:
114 		lmshare = kmem_alloc(sizeof (smb_share_t), KM_SLEEP);
115 		if (error = xcopyin(arg, lmshare, sizeof (smb_share_t))) {
116 			kmem_free(lmshare, sizeof (smb_share_t));
117 			kmem_free(buf, SMB_SHARE_DSIZE);
118 			return (error);
119 		}
120 		smb_dr_put_share(enc_ctx, lmshare);
121 		break;
122 
123 	case SMB_SHROP_DELETE:
124 		str = kmem_alloc(MAXPATHLEN, KM_SLEEP);
125 		if (error = copyinstr(arg, str, MAXPATHLEN, NULL)) {
126 			kmem_free(str, MAXPATHLEN);
127 			kmem_free(buf, SMB_SHARE_DSIZE);
128 			return (error);
129 		}
130 		smb_dr_put_string(enc_ctx, str);
131 		kmem_free(str, MAXPATHLEN);
132 		break;
133 	}
134 
135 	if ((error = smb_dr_encode_finish(enc_ctx, &used)) != 0) {
136 		kmem_free(buf, SMB_SHARE_DSIZE);
137 		if (lmshare)
138 			kmem_free(lmshare, sizeof (smb_share_t));
139 		return (NERR_InternalError);
140 	}
141 
142 	doorarg.data_ptr = buf;
143 	doorarg.data_size = used;
144 	doorarg.rbuf = buf;
145 	doorarg.rsize = SMB_SHARE_DSIZE;
146 
147 	error = door_ki_upcall_limited(dhdl, &doorarg, NULL, SIZE_MAX, 0);
148 
149 	if (error) {
150 		kmem_free(buf, SMB_SHARE_DSIZE);
151 		if (lmshare)
152 			kmem_free(lmshare, sizeof (smb_share_t));
153 		return (error);
154 	}
155 
156 	dec_ctx = smb_dr_decode_start(doorarg.data_ptr, doorarg.data_size);
157 	if (smb_kshare_chk_dsrv_status(opcode, dec_ctx) != 0) {
158 		kmem_free(buf, SMB_SHARE_DSIZE);
159 		if (lmshare)
160 			kmem_free(lmshare, sizeof (smb_share_t));
161 		return (NERR_InternalError);
162 	}
163 
164 	rc = smb_dr_get_uint32(dec_ctx);
165 	if (opcode == SMB_SHROP_ADD)
166 		smb_dr_get_share(dec_ctx, lmshare);
167 
168 	if (smb_dr_decode_finish(dec_ctx))
169 		rc = NERR_InternalError;
170 
171 	kmem_free(buf, SMB_SHARE_DSIZE);
172 	if (lmshare)
173 		kmem_free(lmshare, sizeof (smb_share_t));
174 
175 	return ((rc == NERR_DuplicateShare && add_share) ? 0 : rc);
176 }
177 
178 /*
179  * Executes map and unmap command for shares.
180  */
181 int
182 smb_kshare_exec(smb_shr_execinfo_t *execinfo)
183 {
184 	int exec_rc = 0;
185 
186 	(void) smb_kdoor_upcall(SMB_DR_SHR_EXEC,
187 	    execinfo, smb_shr_execinfo_xdr, &exec_rc, xdr_int);
188 
189 	return (exec_rc);
190 }
191 
192 /*
193  * Obtains any host access restriction on the specified
194  * share for the given host (ipaddr) by calling smbd
195  */
196 uint32_t
197 smb_kshare_hostaccess(smb_kshare_t *shr, smb_inaddr_t *ipaddr)
198 {
199 	smb_shr_hostaccess_query_t req;
200 	uint32_t host_access = SMB_SHRF_ACC_OPEN;
201 	uint32_t flag = SMB_SHRF_ACC_OPEN;
202 	uint32_t access;
203 
204 	if (smb_inet_iszero(ipaddr))
205 		return (ACE_ALL_PERMS);
206 
207 	if ((shr->shr_access_none == NULL || *shr->shr_access_none == '\0') &&
208 	    (shr->shr_access_ro == NULL || *shr->shr_access_ro == '\0') &&
209 	    (shr->shr_access_rw == NULL || *shr->shr_access_rw == '\0'))
210 		return (ACE_ALL_PERMS);
211 
212 	if (shr->shr_access_none != NULL)
213 		flag |= SMB_SHRF_ACC_NONE;
214 	if (shr->shr_access_ro != NULL)
215 		flag |= SMB_SHRF_ACC_RO;
216 	if (shr->shr_access_rw != NULL)
217 		flag |= SMB_SHRF_ACC_RW;
218 
219 	req.shq_none = shr->shr_access_none;
220 	req.shq_ro = shr->shr_access_ro;
221 	req.shq_rw = shr->shr_access_rw;
222 	req.shq_flag = flag;
223 	req.shq_ipaddr = *ipaddr;
224 
225 	(void) smb_kdoor_upcall(SMB_DR_SHR_HOSTACCESS,
226 	    &req, smb_shr_hostaccess_query_xdr, &host_access, xdr_uint32_t);
227 
228 	switch (host_access) {
229 	case SMB_SHRF_ACC_RO:
230 		access = ACE_ALL_PERMS & ~ACE_ALL_WRITE_PERMS;
231 		break;
232 	case SMB_SHRF_ACC_OPEN:
233 	case SMB_SHRF_ACC_RW:
234 		access = ACE_ALL_PERMS;
235 		break;
236 	case SMB_SHRF_ACC_NONE:
237 	default:
238 		access = 0;
239 	}
240 
241 	return (access);
242 }
243 
244 /*
245  * This function is called when smb_server_t is
246  * created which means smb/service is ready for
247  * exporting SMB shares
248  */
249 void
250 smb_export_start(void)
251 {
252 	mutex_enter(&smb_export.e_mutex);
253 	if (smb_export.e_ready) {
254 		mutex_exit(&smb_export.e_mutex);
255 		return;
256 	}
257 
258 	smb_export.e_ready = B_TRUE;
259 	mutex_exit(&smb_export.e_mutex);
260 
261 	smb_avl_create(&smb_export.e_share_avl, sizeof (smb_kshare_t),
262 	    offsetof(smb_kshare_t, shr_link), &smb_kshare_avlops);
263 
264 	(void) smb_kshare_export_trans("IPC$",	"IPC$", "Remote IPC");
265 	(void) smb_kshare_export_trans("c$",	SMB_CVOL, "Default Share");
266 	(void) smb_kshare_export_trans("vss$",	SMB_VSS, "VSS");
267 }
268 
269 /*
270  * This function is called when smb_server_t goes
271  * away which means SMB shares should not be made
272  * available to clients
273  */
274 void
275 smb_export_stop(void)
276 {
277 	mutex_enter(&smb_export.e_mutex);
278 	if (!smb_export.e_ready) {
279 		mutex_exit(&smb_export.e_mutex);
280 		return;
281 	}
282 	smb_export.e_ready = B_FALSE;
283 	mutex_exit(&smb_export.e_mutex);
284 
285 	smb_avl_destroy(&smb_export.e_share_avl);
286 	smb_vfs_rele_all(&smb_export);
287 }
288 
289 int
290 smb_kshare_init(void)
291 {
292 	int rc;
293 
294 	smb_export.e_cache_share = kmem_cache_create("smb_share_cache",
295 	    sizeof (smb_kshare_t), 8, NULL, NULL, NULL, NULL, NULL, 0);
296 
297 	smb_export.e_cache_unexport = kmem_cache_create("smb_unexport_cache",
298 	    sizeof (smb_unshare_t), 8, NULL, NULL, NULL, NULL, NULL, 0);
299 
300 	smb_export.e_cache_vfs = kmem_cache_create("smb_vfs_cache",
301 	    sizeof (smb_vfs_t), 8, NULL, NULL, NULL, NULL, NULL, 0);
302 
303 	smb_llist_constructor(&smb_export.e_vfs_list, sizeof (smb_vfs_t),
304 	    offsetof(smb_vfs_t, sv_lnd));
305 
306 	smb_slist_constructor(&smb_export.e_unexport_list,
307 	    sizeof (smb_unshare_t), offsetof(smb_unshare_t, us_lnd));
308 
309 	smb_thread_init(&smb_export.e_unexport_thread, "smb_thread_unexport",
310 	    smb_kshare_unexport_thread, NULL, NULL, NULL);
311 
312 	if ((rc = smb_thread_start(&smb_export.e_unexport_thread)) != 0)
313 		return (rc);
314 
315 	return (0);
316 }
317 
318 void
319 smb_kshare_fini(void)
320 {
321 	smb_unshare_t *ux;
322 
323 	smb_thread_stop(&smb_export.e_unexport_thread);
324 	smb_thread_destroy(&smb_export.e_unexport_thread);
325 
326 	while ((ux = list_head(&smb_export.e_unexport_list.sl_list)) != NULL) {
327 		smb_slist_remove(&smb_export.e_unexport_list, ux);
328 		kmem_cache_free(smb_export.e_cache_unexport, ux);
329 	}
330 	smb_slist_destructor(&smb_export.e_unexport_list);
331 
332 	smb_vfs_rele_all(&smb_export);
333 
334 	smb_llist_destructor(&smb_export.e_vfs_list);
335 
336 	kmem_cache_destroy(smb_export.e_cache_unexport);
337 	kmem_cache_destroy(smb_export.e_cache_share);
338 	kmem_cache_destroy(smb_export.e_cache_vfs);
339 }
340 
341 /*
342  * A list of shares in nvlist format can be sent down
343  * from userspace thourgh the IOCTL interface. The nvlist
344  * is unpacked here and all the shares in the list will
345  * be exported.
346  */
347 int
348 smb_kshare_export_list(smb_ioc_share_t *ioc)
349 {
350 	nvlist_t	*shrlist;
351 	nvlist_t	 *share;
352 	nvpair_t	 *nvp;
353 	smb_kshare_t	 *shr;
354 	char		*shrname;
355 	int		rc;
356 
357 	if (!smb_export_isready())
358 		return (ENOTACTIVE);
359 
360 	if ((rc = nvlist_unpack(ioc->shr, ioc->shrlen, &shrlist, KM_SLEEP))
361 	    != 0)
362 		return (rc);
363 
364 	for (nvp = nvlist_next_nvpair(shrlist, NULL); nvp != NULL;
365 	    nvp = nvlist_next_nvpair(shrlist, nvp)) {
366 		if (nvpair_type(nvp) != DATA_TYPE_NVLIST)
367 			continue;
368 
369 		shrname = nvpair_name(nvp);
370 		ASSERT(shrname);
371 
372 		if ((rc = nvpair_value_nvlist(nvp, &share)) != 0) {
373 			cmn_err(CE_WARN, "export[%s]: failed accessing",
374 			    shrname);
375 			continue;
376 		}
377 
378 		if ((shr = smb_kshare_decode(share)) == NULL) {
379 			cmn_err(CE_WARN, "export[%s]: failed decoding",
380 			    shrname);
381 			continue;
382 		}
383 
384 		if ((rc = smb_kshare_export(shr)) != 0) {
385 			smb_kshare_destroy(shr);
386 			continue;
387 		}
388 	}
389 
390 	nvlist_free(shrlist);
391 
392 	return (0);
393 }
394 
395 /*
396  * This function is invoked when a share is disabled to disconnect trees
397  * and close files.  Cleaning up may involve VOP and/or VFS calls, which
398  * may conflict/deadlock with stuck threads if something is amiss with the
399  * file system.  Queueing the request for asynchronous processing allows the
400  * call to return immediately so that, if the unshare is being done in the
401  * context of a forced unmount, the forced unmount will always be able to
402  * proceed (unblocking stuck I/O and eventually allowing all blocked unshare
403  * processes to complete).
404  *
405  * The path lookup to find the root vnode of the VFS in question and the
406  * release of this vnode are done synchronously prior to any associated
407  * unmount.  Doing these asynchronous to an associated unmount could run
408  * the risk of a spurious EBUSY for a standard unmount or an EIO during
409  * the path lookup due to a forced unmount finishing first.
410  */
411 int
412 smb_kshare_unexport_list(smb_ioc_share_t *ioc)
413 {
414 	smb_unshare_t	*ux;
415 	nvlist_t	*shrlist;
416 	nvpair_t	*nvp;
417 	boolean_t	unexport = B_FALSE;
418 	char		*shrname;
419 	int		rc;
420 
421 	if ((rc = nvlist_unpack(ioc->shr, ioc->shrlen, &shrlist, 0)) != 0)
422 		return (rc);
423 
424 	for (nvp = nvlist_next_nvpair(shrlist, NULL); nvp != NULL;
425 	    nvp = nvlist_next_nvpair(shrlist, nvp)) {
426 		if (nvpair_type(nvp) != DATA_TYPE_NVLIST)
427 			continue;
428 
429 		shrname = nvpair_name(nvp);
430 		ASSERT(shrname);
431 
432 		if ((rc = smb_kshare_unexport(shrname)) != 0)
433 			continue;
434 
435 		ux = kmem_cache_alloc(smb_export.e_cache_unexport, KM_SLEEP);
436 		(void) strlcpy(ux->us_sharename, shrname, MAXNAMELEN);
437 
438 		smb_slist_insert_tail(&smb_export.e_unexport_list, ux);
439 		unexport = B_TRUE;
440 	}
441 
442 	nvlist_free(shrlist);
443 
444 	if (unexport)
445 		smb_thread_signal(&smb_export.e_unexport_thread);
446 
447 	return (0);
448 }
449 
450 /*
451  * This function builds a response for a NetShareEnum RAP request.
452  * List of shares is scanned twice. In the first round the total number
453  * of shares which their OEM name is shorter than 13 chars (esi->es_ntotal)
454  * and also the number of shares that fit in the given buffer are calculated.
455  * In the second round the shares data are encoded in the buffer.
456  *
457  * The data associated with each share has two parts, a fixed size part and
458  * a variable size part which is share's comment. The outline of the response
459  * buffer is so that fixed part for all the shares will appear first and follows
460  * with the comments for all those shares and that's why the data cannot be
461  * encoded in one round without unnecessarily complicating the code.
462  */
463 void
464 smb_kshare_enum(smb_enumshare_info_t *esi)
465 {
466 	smb_avl_t *share_avl;
467 	smb_avl_cursor_t cursor;
468 	smb_kshare_t *shr;
469 	int remained;
470 	uint16_t infolen = 0;
471 	uint16_t cmntlen = 0;
472 	uint16_t sharelen;
473 	uint16_t clen;
474 	uint32_t cmnt_offs;
475 	smb_msgbuf_t info_mb;
476 	smb_msgbuf_t cmnt_mb;
477 	boolean_t autohome_added = B_FALSE;
478 
479 	if (!smb_export_isready()) {
480 		esi->es_ntotal = esi->es_nsent = 0;
481 		esi->es_datasize = 0;
482 		return;
483 	}
484 
485 	esi->es_ntotal = esi->es_nsent = 0;
486 	remained = esi->es_bufsize;
487 	share_avl = &smb_export.e_share_avl;
488 
489 	/* Do the necessary calculations in the first round */
490 	smb_avl_iterinit(share_avl, &cursor);
491 
492 	while ((shr = smb_avl_iterate(share_avl, &cursor)) != NULL) {
493 		if (shr->shr_oemname == NULL) {
494 			smb_avl_release(share_avl, shr);
495 			continue;
496 		}
497 
498 		if ((shr->shr_flags & SMB_SHRF_AUTOHOME) && !autohome_added) {
499 			if (esi->es_posix_uid == shr->shr_uid) {
500 				autohome_added = B_TRUE;
501 			} else {
502 				smb_avl_release(share_avl, shr);
503 				continue;
504 			}
505 		}
506 
507 		esi->es_ntotal++;
508 
509 		if (remained <= 0) {
510 			smb_avl_release(share_avl, shr);
511 			continue;
512 		}
513 
514 		clen = strlen(shr->shr_cmnt) + 1;
515 		sharelen = SHARE_INFO_1_SIZE + clen;
516 
517 		if (sharelen <= remained) {
518 			infolen += SHARE_INFO_1_SIZE;
519 			cmntlen += clen;
520 		}
521 
522 		remained -= sharelen;
523 		smb_avl_release(share_avl, shr);
524 	}
525 
526 	esi->es_datasize = infolen + cmntlen;
527 
528 	smb_msgbuf_init(&info_mb, (uint8_t *)esi->es_buf, infolen, 0);
529 	smb_msgbuf_init(&cmnt_mb, (uint8_t *)esi->es_buf + infolen, cmntlen, 0);
530 	cmnt_offs = infolen;
531 
532 	/* Encode the data in the second round */
533 	smb_avl_iterinit(share_avl, &cursor);
534 	autohome_added = B_FALSE;
535 
536 	while ((shr = smb_avl_iterate(share_avl, &cursor)) != NULL) {
537 		if (shr->shr_oemname == NULL) {
538 			smb_avl_release(share_avl, shr);
539 			continue;
540 		}
541 
542 		if ((shr->shr_flags & SMB_SHRF_AUTOHOME) && !autohome_added) {
543 			if (esi->es_posix_uid == shr->shr_uid) {
544 				autohome_added = B_TRUE;
545 			} else {
546 				smb_avl_release(share_avl, shr);
547 				continue;
548 			}
549 		}
550 
551 		if (smb_msgbuf_encode(&info_mb, "13c.wl",
552 		    shr->shr_oemname, shr->shr_type, cmnt_offs) < 0) {
553 			smb_avl_release(share_avl, shr);
554 			break;
555 		}
556 
557 		if (smb_msgbuf_encode(&cmnt_mb, "s", shr->shr_cmnt) < 0) {
558 			smb_avl_release(share_avl, shr);
559 			break;
560 		}
561 
562 		cmnt_offs += strlen(shr->shr_cmnt) + 1;
563 		esi->es_nsent++;
564 
565 		smb_avl_release(share_avl, shr);
566 	}
567 
568 	smb_msgbuf_term(&info_mb);
569 	smb_msgbuf_term(&cmnt_mb);
570 }
571 
572 /*
573  * Looks up the given share and returns a pointer
574  * to its definition if it's found. A hold on the
575  * object is taken before the pointer is returned
576  * in which case the caller MUST always call
577  * smb_kshare_release().
578  */
579 smb_kshare_t *
580 smb_kshare_lookup(const char *shrname)
581 {
582 	smb_kshare_t key;
583 	smb_kshare_t *shr;
584 
585 	ASSERT(shrname);
586 
587 	if (!smb_export_isready())
588 		return (NULL);
589 
590 	key.shr_name = (char *)shrname;
591 	shr = smb_avl_lookup(&smb_export.e_share_avl, &key);
592 
593 	return (shr);
594 }
595 
596 /*
597  * Releases the hold taken on the specified share object
598  */
599 void
600 smb_kshare_release(smb_kshare_t *shr)
601 {
602 	ASSERT(shr);
603 	ASSERT(shr->shr_magic == SMB_SHARE_MAGIC);
604 
605 	smb_avl_release(&smb_export.e_share_avl, shr);
606 }
607 
608 
609 /*
610  * Add the given share in the specified server.
611  * If the share is a disk share, smb_vfs_hold() is
612  * invoked to ensure that there is a hold on the
613  * corresponding file system before the share is
614  * added to shares AVL.
615  *
616  * If the share is an Autohome share and it is
617  * already in the AVL only a reference count for
618  * that share is incremented.
619  */
620 static int
621 smb_kshare_export(smb_kshare_t *shr)
622 {
623 	smb_avl_t	*share_avl;
624 	smb_kshare_t	*auto_shr;
625 	vnode_t		*vp;
626 	int		rc = 0;
627 
628 	share_avl = &smb_export.e_share_avl;
629 
630 	if (!STYPE_ISDSK(shr->shr_type)) {
631 		if ((rc = smb_avl_add(share_avl, shr)) != 0) {
632 			cmn_err(CE_WARN, "export[%s]: failed caching (%d)",
633 			    shr->shr_name, rc);
634 		}
635 
636 		return (rc);
637 	}
638 
639 	if ((auto_shr = smb_avl_lookup(share_avl, shr)) != NULL) {
640 		if ((auto_shr->shr_flags & SMB_SHRF_AUTOHOME) == 0) {
641 			smb_avl_release(share_avl, auto_shr);
642 			return (EEXIST);
643 		}
644 
645 		mutex_enter(&auto_shr->shr_mutex);
646 		auto_shr->shr_autocnt++;
647 		mutex_exit(&auto_shr->shr_mutex);
648 		smb_avl_release(share_avl, auto_shr);
649 		return (0);
650 	}
651 
652 	if ((rc = smb_server_sharevp(shr->shr_path, &vp)) != 0) {
653 		cmn_err(CE_WARN, "export[%s(%s)]: failed obtaining vnode (%d)",
654 		    shr->shr_name, shr->shr_path, rc);
655 		return (rc);
656 	}
657 
658 	if ((rc = smb_vfs_hold(&smb_export, vp->v_vfsp)) == 0) {
659 		if ((rc = smb_avl_add(share_avl, shr)) != 0) {
660 			cmn_err(CE_WARN, "export[%s]: failed caching (%d)",
661 			    shr->shr_name, rc);
662 			smb_vfs_rele(&smb_export, vp->v_vfsp);
663 		}
664 	} else {
665 		cmn_err(CE_WARN, "export[%s(%s)]: failed holding VFS (%d)",
666 		    shr->shr_name, shr->shr_path, rc);
667 	}
668 
669 	VN_RELE(vp);
670 	return (rc);
671 }
672 
673 /*
674  * Removes the share specified by 'shrname' from the AVL
675  * tree of the given server if it's there.
676  *
677  * If the share is an Autohome share, the autohome count
678  * is decremented and the share is only removed if the
679  * count goes to zero.
680  *
681  * If the share is a disk share, the hold on the corresponding
682  * file system is released before removing the share from
683  * the AVL tree.
684  */
685 static int
686 smb_kshare_unexport(const char *shrname)
687 {
688 	smb_avl_t	*share_avl;
689 	smb_kshare_t	key;
690 	smb_kshare_t	*shr;
691 	vnode_t		*vp;
692 	int		rc;
693 	boolean_t	auto_unexport;
694 
695 	share_avl = &smb_export.e_share_avl;
696 
697 	key.shr_name = (char *)shrname;
698 	if ((shr = smb_avl_lookup(share_avl, &key)) == NULL)
699 		return (ENOENT);
700 
701 	if ((shr->shr_flags & SMB_SHRF_AUTOHOME) != 0) {
702 		mutex_enter(&shr->shr_mutex);
703 		shr->shr_autocnt--;
704 		auto_unexport = (shr->shr_autocnt == 0);
705 		mutex_exit(&shr->shr_mutex);
706 		if (!auto_unexport) {
707 			smb_avl_release(share_avl, shr);
708 			return (0);
709 		}
710 	}
711 
712 	if (STYPE_ISDSK(shr->shr_type)) {
713 		if ((rc = smb_server_sharevp(shr->shr_path, &vp)) != 0) {
714 			smb_avl_release(share_avl, shr);
715 			cmn_err(CE_WARN, "unexport[%s]: failed obtaining vnode"
716 			    " (%d)", shrname, rc);
717 			return (rc);
718 		}
719 
720 		smb_vfs_rele(&smb_export, vp->v_vfsp);
721 		VN_RELE(vp);
722 	}
723 
724 	smb_avl_remove(share_avl, shr);
725 	smb_avl_release(share_avl, shr);
726 
727 	return (0);
728 }
729 
730 /*
731  * Exports IPC$ or Admin shares
732  */
733 static int
734 smb_kshare_export_trans(char *name, char *path, char *cmnt)
735 {
736 	smb_kshare_t *shr;
737 
738 	ASSERT(name);
739 	ASSERT(path);
740 
741 	shr = kmem_cache_alloc(smb_export.e_cache_share, KM_SLEEP);
742 	bzero(shr, sizeof (smb_kshare_t));
743 
744 	shr->shr_magic = SMB_SHARE_MAGIC;
745 	shr->shr_cache = smb_export.e_cache_share;
746 	shr->shr_refcnt = 1;
747 	shr->shr_flags = SMB_SHRF_TRANS | smb_kshare_is_admin(shr->shr_name);
748 	if (strcasecmp(name, "IPC$") == 0)
749 		shr->shr_type = STYPE_IPC;
750 	else
751 		shr->shr_type = STYPE_DISKTREE;
752 
753 	shr->shr_type |= smb_kshare_is_special(shr->shr_name);
754 
755 	shr->shr_name = smb_mem_strdup(name);
756 	if (path)
757 		shr->shr_path = smb_mem_strdup(path);
758 	if (cmnt)
759 		shr->shr_cmnt = smb_mem_strdup(cmnt);
760 	shr->shr_oemname = smb_kshare_oemname(name);
761 
762 	return (smb_kshare_export(shr));
763 }
764 
765 /*
766  * Decodes share information in an nvlist format into a smb_kshare_t
767  * structure.
768  *
769  * This is a temporary function and will be replaced by functions
770  * provided by libsharev2 code after it's available.
771  */
772 static smb_kshare_t *
773 smb_kshare_decode(nvlist_t *share)
774 {
775 	smb_kshare_t tmp;
776 	smb_kshare_t *shr;
777 	nvlist_t *smb;
778 	char *csc_name = NULL;
779 	int rc;
780 
781 	ASSERT(share);
782 
783 	bzero(&tmp, sizeof (smb_kshare_t));
784 
785 	rc = nvlist_lookup_string(share, "name", &tmp.shr_name);
786 	rc |= nvlist_lookup_string(share, "path", &tmp.shr_path);
787 	(void) nvlist_lookup_string(share, "desc", &tmp.shr_cmnt);
788 
789 	ASSERT(tmp.shr_name && tmp.shr_path);
790 
791 	rc |= nvlist_lookup_nvlist(share, "smb", &smb);
792 	if (rc != 0) {
793 		cmn_err(CE_WARN, "kshare: failed looking up SMB properties"
794 		    " (%d)", rc);
795 		return (NULL);
796 	}
797 
798 	(void) nvlist_lookup_string(smb, SHOPT_AD_CONTAINER,
799 	    &tmp.shr_container);
800 	(void) nvlist_lookup_string(smb, SHOPT_NONE, &tmp.shr_access_none);
801 	(void) nvlist_lookup_string(smb, SHOPT_RO, &tmp.shr_access_ro);
802 	(void) nvlist_lookup_string(smb, SHOPT_RW, &tmp.shr_access_rw);
803 
804 	tmp.shr_flags |= smb_kshare_decode_bool(smb, SHOPT_ABE, SMB_SHRF_ABE);
805 	tmp.shr_flags |= smb_kshare_decode_bool(smb, SHOPT_CATIA,
806 	    SMB_SHRF_CATIA);
807 	tmp.shr_flags |= smb_kshare_decode_bool(smb, SHOPT_GUEST,
808 	    SMB_SHRF_GUEST_OK);
809 	tmp.shr_flags |= smb_kshare_decode_bool(smb, SHOPT_DFSROOT,
810 	    SMB_SHRF_DFSROOT);
811 	tmp.shr_flags |= smb_kshare_decode_bool(smb, "Autohome",
812 	    SMB_SHRF_AUTOHOME);
813 
814 	if ((tmp.shr_flags & SMB_SHRF_AUTOHOME) == SMB_SHRF_AUTOHOME) {
815 		rc = nvlist_lookup_uint32(smb, "uid", &tmp.shr_uid);
816 		rc |= nvlist_lookup_uint32(smb, "gid", &tmp.shr_gid);
817 		if (rc != 0) {
818 			cmn_err(CE_WARN, "kshare: failed looking up UID/GID"
819 			    " (%d)", rc);
820 			return (NULL);
821 		}
822 	}
823 
824 	(void) nvlist_lookup_string(smb, SHOPT_CSC, &csc_name);
825 	smb_kshare_csc_flags(&tmp, csc_name);
826 
827 	shr = kmem_cache_alloc(smb_export.e_cache_share, KM_SLEEP);
828 	bzero(shr, sizeof (smb_kshare_t));
829 
830 	shr->shr_magic = SMB_SHARE_MAGIC;
831 	shr->shr_cache = smb_export.e_cache_share;
832 	shr->shr_refcnt = 1;
833 
834 	shr->shr_name = smb_mem_strdup(tmp.shr_name);
835 	shr->shr_path = smb_mem_strdup(tmp.shr_path);
836 	if (tmp.shr_cmnt)
837 		shr->shr_cmnt = smb_mem_strdup(tmp.shr_cmnt);
838 	if (tmp.shr_container)
839 		shr->shr_container = smb_mem_strdup(tmp.shr_container);
840 	if (tmp.shr_access_none)
841 		shr->shr_access_none = smb_mem_strdup(tmp.shr_access_none);
842 	if (tmp.shr_access_ro)
843 		shr->shr_access_ro = smb_mem_strdup(tmp.shr_access_ro);
844 	if (tmp.shr_access_rw)
845 		shr->shr_access_rw = smb_mem_strdup(tmp.shr_access_rw);
846 
847 	shr->shr_oemname = smb_kshare_oemname(shr->shr_name);
848 	shr->shr_flags = tmp.shr_flags | smb_kshare_is_admin(shr->shr_name);
849 	shr->shr_type = STYPE_DISKTREE | smb_kshare_is_special(shr->shr_name);
850 
851 	shr->shr_uid = tmp.shr_uid;
852 	shr->shr_gid = tmp.shr_gid;
853 
854 	if ((shr->shr_flags & SMB_SHRF_AUTOHOME) == SMB_SHRF_AUTOHOME)
855 		shr->shr_autocnt = 1;
856 
857 	return (shr);
858 }
859 
860 #if 0
861 static void
862 smb_kshare_log(smb_kshare_t *shr)
863 {
864 	cmn_err(CE_NOTE, "Share info:");
865 	cmn_err(CE_NOTE, "\tname: %s", (shr->shr_name) ? shr->shr_name : "");
866 	cmn_err(CE_NOTE, "\tpath: %s", (shr->shr_path) ? shr->shr_path : "");
867 	cmn_err(CE_NOTE, "\tcmnt: (%s)",
868 	    (shr->shr_cmnt) ? shr->shr_cmnt : "NULL");
869 	cmn_err(CE_NOTE, "\toemname: (%s)",
870 	    (shr->shr_oemname) ? shr->shr_oemname : "NULL");
871 	cmn_err(CE_NOTE, "\tflags: %X", shr->shr_flags);
872 	cmn_err(CE_NOTE, "\ttype: %d", shr->shr_type);
873 }
874 #endif
875 
876 /*
877  * Compare function used by shares AVL
878  */
879 static int
880 smb_kshare_cmp(const void *p1, const void *p2)
881 {
882 	smb_kshare_t *shr1 = (smb_kshare_t *)p1;
883 	smb_kshare_t *shr2 = (smb_kshare_t *)p2;
884 	int rc;
885 
886 	ASSERT(shr1);
887 	ASSERT(shr1->shr_name);
888 
889 	ASSERT(shr2);
890 	ASSERT(shr2->shr_name);
891 
892 	rc = smb_strcasecmp(shr1->shr_name, shr2->shr_name, 0);
893 
894 	if (rc < 0)
895 		return (-1);
896 
897 	if (rc > 0)
898 		return (1);
899 
900 	return (0);
901 }
902 
903 /*
904  * This function is called by smb_avl routines whenever
905  * there is a need to take a hold on a share structure
906  * inside AVL
907  */
908 static void
909 smb_kshare_hold(const void *p)
910 {
911 	smb_kshare_t *shr = (smb_kshare_t *)p;
912 
913 	ASSERT(shr);
914 	ASSERT(shr->shr_magic == SMB_SHARE_MAGIC);
915 
916 	mutex_enter(&shr->shr_mutex);
917 	shr->shr_refcnt++;
918 	mutex_exit(&shr->shr_mutex);
919 }
920 
921 /*
922  * This function must be called by smb_avl routines whenever
923  * smb_kshare_hold is called and the hold needs to be released.
924  */
925 static boolean_t
926 smb_kshare_rele(const void *p)
927 {
928 	smb_kshare_t *shr = (smb_kshare_t *)p;
929 	boolean_t destroy;
930 
931 	ASSERT(shr);
932 	ASSERT(shr->shr_magic == SMB_SHARE_MAGIC);
933 
934 	mutex_enter(&shr->shr_mutex);
935 	ASSERT(shr->shr_refcnt > 0);
936 	shr->shr_refcnt--;
937 	destroy = (shr->shr_refcnt == 0);
938 	mutex_exit(&shr->shr_mutex);
939 
940 	return (destroy);
941 }
942 
943 /*
944  * Frees all the memory allocated for the given
945  * share structure. It also removes the structure
946  * from the share cache.
947  */
948 static void
949 smb_kshare_destroy(void *p)
950 {
951 	smb_kshare_t *shr = (smb_kshare_t *)p;
952 
953 	ASSERT(shr);
954 	ASSERT(shr->shr_magic == SMB_SHARE_MAGIC);
955 
956 	smb_mem_free(shr->shr_name);
957 	smb_mem_free(shr->shr_path);
958 	smb_mem_free(shr->shr_cmnt);
959 	smb_mem_free(shr->shr_container);
960 	smb_mem_free(shr->shr_oemname);
961 	smb_mem_free(shr->shr_access_none);
962 	smb_mem_free(shr->shr_access_ro);
963 	smb_mem_free(shr->shr_access_rw);
964 
965 	kmem_cache_free(shr->shr_cache, shr);
966 }
967 
968 
969 /*
970  * Generate an OEM name for the given share name.  If the name is
971  * shorter than 13 bytes the oemname will be returned; otherwise NULL
972  * is returned.
973  */
974 static char *
975 smb_kshare_oemname(const char *shrname)
976 {
977 	smb_wchar_t *unibuf;
978 	char *oem_name;
979 	int length;
980 
981 	length = strlen(shrname) + 1;
982 
983 	oem_name = smb_mem_alloc(length);
984 	unibuf = smb_mem_alloc(length * sizeof (smb_wchar_t));
985 
986 	(void) smb_mbstowcs(unibuf, shrname, length);
987 
988 	if (ucstooem(oem_name, unibuf, length, OEM_CPG_850) == 0)
989 		(void) strcpy(oem_name, shrname);
990 
991 	smb_mem_free(unibuf);
992 
993 	if (strlen(oem_name) + 1 > SMB_SHARE_OEMNAME_MAX) {
994 		smb_mem_free(oem_name);
995 		return (NULL);
996 	}
997 
998 	return (oem_name);
999 }
1000 
1001 /*
1002  * Special share reserved for interprocess communication (IPC$) or
1003  * remote administration of the server (ADMIN$). Can also refer to
1004  * administrative shares such as C$, D$, E$, and so forth.
1005  */
1006 static int
1007 smb_kshare_is_special(const char *sharename)
1008 {
1009 	int len;
1010 
1011 	if (sharename == NULL)
1012 		return (0);
1013 
1014 	if ((len = strlen(sharename)) == 0)
1015 		return (0);
1016 
1017 	if (sharename[len - 1] == '$')
1018 		return (STYPE_SPECIAL);
1019 
1020 	return (0);
1021 }
1022 
1023 /*
1024  * Check whether or not this is a default admin share: C$, D$ etc.
1025  */
1026 static boolean_t
1027 smb_kshare_is_admin(const char *sharename)
1028 {
1029 	if (sharename == NULL)
1030 		return (B_FALSE);
1031 
1032 	if (strlen(sharename) == 2 &&
1033 	    smb_isalpha(sharename[0]) && sharename[1] == '$') {
1034 		return (B_TRUE);
1035 	}
1036 
1037 	return (B_FALSE);
1038 }
1039 
1040 /*
1041  * Decodes the given boolean share option.
1042  * If the option is present in the nvlist and it's value is true
1043  * returns the corresponding flag value, otherwise returns 0.
1044  */
1045 static uint32_t
1046 smb_kshare_decode_bool(nvlist_t *nvl, const char *propname, uint32_t flag)
1047 {
1048 	char *boolp;
1049 
1050 	if (nvlist_lookup_string(nvl, propname, &boolp) == 0)
1051 		if (strcasecmp(boolp, "true") == 0)
1052 			return (flag);
1053 
1054 	return (0);
1055 }
1056 
1057 /*
1058  * Map a client-side caching (CSC) option to the appropriate share
1059  * flag.  Only one option is allowed; an error will be logged if
1060  * multiple options have been specified.  We don't need to do anything
1061  * about multiple values here because the SRVSVC will not recognize
1062  * a value containing multiple flags and will return the default value.
1063  *
1064  * If the option value is not recognized, it will be ignored: invalid
1065  * values will typically be caught and rejected by sharemgr.
1066  */
1067 static void
1068 smb_kshare_csc_flags(smb_kshare_t *shr, const char *value)
1069 {
1070 	int i;
1071 	static struct {
1072 		char *value;
1073 		uint32_t flag;
1074 	} cscopt[] = {
1075 		{ "disabled",	SMB_SHRF_CSC_DISABLED },
1076 		{ "manual",	SMB_SHRF_CSC_MANUAL },
1077 		{ "auto",	SMB_SHRF_CSC_AUTO },
1078 		{ "vdo",	SMB_SHRF_CSC_VDO }
1079 	};
1080 
1081 	if (value == NULL)
1082 		return;
1083 
1084 	for (i = 0; i < (sizeof (cscopt) / sizeof (cscopt[0])); ++i) {
1085 		if (strcasecmp(value, cscopt[i].value) == 0) {
1086 			shr->shr_flags |= cscopt[i].flag;
1087 			break;
1088 		}
1089 	}
1090 
1091 	switch (shr->shr_flags & SMB_SHRF_CSC_MASK) {
1092 	case 0:
1093 	case SMB_SHRF_CSC_DISABLED:
1094 	case SMB_SHRF_CSC_MANUAL:
1095 	case SMB_SHRF_CSC_AUTO:
1096 	case SMB_SHRF_CSC_VDO:
1097 		break;
1098 
1099 	default:
1100 		cmn_err(CE_NOTE, "csc option conflict: 0x%08x",
1101 		    shr->shr_flags & SMB_SHRF_CSC_MASK);
1102 		break;
1103 	}
1104 }
1105 
1106 /*
1107  * This function processes the unexport event list and disconnects shares
1108  * asynchronously.  The function executes as a zone-specific thread.
1109  *
1110  * The server arg passed in is safe to use without a reference count, because
1111  * the server cannot be deleted until smb_thread_stop()/destroy() return,
1112  * which is also when the thread exits.
1113  */
1114 /*ARGSUSED*/
1115 static void
1116 smb_kshare_unexport_thread(smb_thread_t *thread, void *arg)
1117 {
1118 	smb_unshare_t	*ux;
1119 
1120 	while (smb_thread_continue(thread)) {
1121 		while ((ux = list_head(&smb_export.e_unexport_list.sl_list))
1122 		    != NULL) {
1123 			smb_slist_remove(&smb_export.e_unexport_list, ux);
1124 			(void) smb_server_unshare(ux->us_sharename);
1125 			kmem_cache_free(smb_export.e_cache_unexport, ux);
1126 		}
1127 	}
1128 }
1129 
1130 static boolean_t
1131 smb_export_isready(void)
1132 {
1133 	boolean_t ready;
1134 
1135 	mutex_enter(&smb_export.e_mutex);
1136 	ready = smb_export.e_ready;
1137 	mutex_exit(&smb_export.e_mutex);
1138 
1139 	return (ready);
1140 }
1141 
1142 /*
1143  * Return 0 upon success. Otherwise > 0
1144  */
1145 static int
1146 smb_kshare_chk_dsrv_status(int opcode, smb_dr_ctx_t *dec_ctx)
1147 {
1148 	int status = smb_dr_get_int32(dec_ctx);
1149 	int err;
1150 
1151 	switch (status) {
1152 	case SMB_SHARE_DSUCCESS:
1153 		return (0);
1154 
1155 	case SMB_SHARE_DERROR:
1156 		err = smb_dr_get_uint32(dec_ctx);
1157 		cmn_err(CE_WARN, "%d: Encountered door server error %d",
1158 		    opcode, err);
1159 		(void) smb_dr_decode_finish(dec_ctx);
1160 		return (err);
1161 	}
1162 
1163 	ASSERT(0);
1164 	return (EINVAL);
1165 }
1166