xref: /titanic_50/usr/src/uts/common/fs/zfs/dsl_synctask.c (revision 683b29499b14fddf042df3e4ecb71a1d5bebe3a8)
1 /*
2  * CDDL HEADER START
3  *
4  * The contents of this file are subject to the terms of the
5  * Common Development and Distribution License (the "License").
6  * You may not use this file except in compliance with the License.
7  *
8  * You can obtain a copy of the license at usr/src/OPENSOLARIS.LICENSE
9  * or http://www.opensolaris.org/os/licensing.
10  * See the License for the specific language governing permissions
11  * and limitations under the License.
12  *
13  * When distributing Covered Code, include this CDDL HEADER in each
14  * file and include the License file at usr/src/OPENSOLARIS.LICENSE.
15  * If applicable, add the following below this CDDL HEADER, with the
16  * fields enclosed by brackets "[]" replaced with your own identifying
17  * information: Portions Copyright [yyyy] [name of copyright owner]
18  *
19  * CDDL HEADER END
20  */
21 /*
22  * Copyright (c) 2005, 2010, Oracle and/or its affiliates. All rights reserved.
23  */
24 
25 #include <sys/dmu.h>
26 #include <sys/dmu_tx.h>
27 #include <sys/dsl_pool.h>
28 #include <sys/dsl_dir.h>
29 #include <sys/dsl_synctask.h>
30 #include <sys/metaslab.h>
31 
32 #define	DST_AVG_BLKSHIFT 14
33 
34 /* ARGSUSED */
35 static int
36 dsl_null_checkfunc(void *arg1, void *arg2, dmu_tx_t *tx)
37 {
38 	return (0);
39 }
40 
41 dsl_sync_task_group_t *
42 dsl_sync_task_group_create(dsl_pool_t *dp)
43 {
44 	dsl_sync_task_group_t *dstg;
45 
46 	dstg = kmem_zalloc(sizeof (dsl_sync_task_group_t), KM_SLEEP);
47 	list_create(&dstg->dstg_tasks, sizeof (dsl_sync_task_t),
48 	    offsetof(dsl_sync_task_t, dst_node));
49 	dstg->dstg_pool = dp;
50 
51 	return (dstg);
52 }
53 
54 void
55 dsl_sync_task_create(dsl_sync_task_group_t *dstg,
56     dsl_checkfunc_t *checkfunc, dsl_syncfunc_t *syncfunc,
57     void *arg1, void *arg2, int blocks_modified)
58 {
59 	dsl_sync_task_t *dst;
60 
61 	if (checkfunc == NULL)
62 		checkfunc = dsl_null_checkfunc;
63 	dst = kmem_zalloc(sizeof (dsl_sync_task_t), KM_SLEEP);
64 	dst->dst_checkfunc = checkfunc;
65 	dst->dst_syncfunc = syncfunc;
66 	dst->dst_arg1 = arg1;
67 	dst->dst_arg2 = arg2;
68 	list_insert_tail(&dstg->dstg_tasks, dst);
69 
70 	dstg->dstg_space += blocks_modified << DST_AVG_BLKSHIFT;
71 }
72 
73 int
74 dsl_sync_task_group_wait(dsl_sync_task_group_t *dstg)
75 {
76 	dmu_tx_t *tx;
77 	uint64_t txg;
78 	dsl_sync_task_t *dst;
79 
80 top:
81 	tx = dmu_tx_create_dd(dstg->dstg_pool->dp_mos_dir);
82 	VERIFY(0 == dmu_tx_assign(tx, TXG_WAIT));
83 
84 	txg = dmu_tx_get_txg(tx);
85 
86 	/* Do a preliminary error check. */
87 	dstg->dstg_err = 0;
88 	rw_enter(&dstg->dstg_pool->dp_config_rwlock, RW_READER);
89 	for (dst = list_head(&dstg->dstg_tasks); dst;
90 	    dst = list_next(&dstg->dstg_tasks, dst)) {
91 #ifdef ZFS_DEBUG
92 		/*
93 		 * Only check half the time, otherwise, the sync-context
94 		 * check will almost never fail.
95 		 */
96 		if (spa_get_random(2) == 0)
97 			continue;
98 #endif
99 		dst->dst_err =
100 		    dst->dst_checkfunc(dst->dst_arg1, dst->dst_arg2, tx);
101 		if (dst->dst_err)
102 			dstg->dstg_err = dst->dst_err;
103 	}
104 	rw_exit(&dstg->dstg_pool->dp_config_rwlock);
105 
106 	if (dstg->dstg_err) {
107 		dmu_tx_commit(tx);
108 		return (dstg->dstg_err);
109 	}
110 
111 	/*
112 	 * We don't generally have many sync tasks, so pay the price of
113 	 * add_tail to get the tasks executed in the right order.
114 	 */
115 	VERIFY(0 == txg_list_add_tail(&dstg->dstg_pool->dp_sync_tasks,
116 	    dstg, txg));
117 
118 	dmu_tx_commit(tx);
119 
120 	txg_wait_synced(dstg->dstg_pool, txg);
121 
122 	if (dstg->dstg_err == EAGAIN) {
123 		txg_wait_synced(dstg->dstg_pool, txg + TXG_DEFER_SIZE);
124 		goto top;
125 	}
126 
127 	return (dstg->dstg_err);
128 }
129 
130 void
131 dsl_sync_task_group_nowait(dsl_sync_task_group_t *dstg, dmu_tx_t *tx)
132 {
133 	uint64_t txg;
134 
135 	dstg->dstg_nowaiter = B_TRUE;
136 	txg = dmu_tx_get_txg(tx);
137 	/*
138 	 * We don't generally have many sync tasks, so pay the price of
139 	 * add_tail to get the tasks executed in the right order.
140 	 */
141 	VERIFY(0 == txg_list_add_tail(&dstg->dstg_pool->dp_sync_tasks,
142 	    dstg, txg));
143 }
144 
145 void
146 dsl_sync_task_group_destroy(dsl_sync_task_group_t *dstg)
147 {
148 	dsl_sync_task_t *dst;
149 
150 	while (dst = list_head(&dstg->dstg_tasks)) {
151 		list_remove(&dstg->dstg_tasks, dst);
152 		kmem_free(dst, sizeof (dsl_sync_task_t));
153 	}
154 	kmem_free(dstg, sizeof (dsl_sync_task_group_t));
155 }
156 
157 void
158 dsl_sync_task_group_sync(dsl_sync_task_group_t *dstg, dmu_tx_t *tx)
159 {
160 	dsl_sync_task_t *dst;
161 	dsl_pool_t *dp = dstg->dstg_pool;
162 	uint64_t quota, used;
163 
164 	ASSERT3U(dstg->dstg_err, ==, 0);
165 
166 	/*
167 	 * Check for sufficient space.  We just check against what's
168 	 * on-disk; we don't want any in-flight accounting to get in our
169 	 * way, because open context may have already used up various
170 	 * in-core limits (arc_tempreserve, dsl_pool_tempreserve).
171 	 */
172 	quota = dsl_pool_adjustedsize(dp, B_FALSE) -
173 	    metaslab_class_get_deferred(spa_normal_class(dp->dp_spa));
174 	used = dp->dp_root_dir->dd_phys->dd_used_bytes;
175 	/* MOS space is triple-dittoed, so we multiply by 3. */
176 	if (dstg->dstg_space > 0 && used + dstg->dstg_space * 3 > quota) {
177 		dstg->dstg_err = ENOSPC;
178 		return;
179 	}
180 
181 	/*
182 	 * Check for errors by calling checkfuncs.
183 	 */
184 	rw_enter(&dp->dp_config_rwlock, RW_WRITER);
185 	for (dst = list_head(&dstg->dstg_tasks); dst;
186 	    dst = list_next(&dstg->dstg_tasks, dst)) {
187 		dst->dst_err =
188 		    dst->dst_checkfunc(dst->dst_arg1, dst->dst_arg2, tx);
189 		if (dst->dst_err)
190 			dstg->dstg_err = dst->dst_err;
191 	}
192 
193 	if (dstg->dstg_err == 0) {
194 		/*
195 		 * Execute sync tasks.
196 		 */
197 		for (dst = list_head(&dstg->dstg_tasks); dst;
198 		    dst = list_next(&dstg->dstg_tasks, dst)) {
199 			dst->dst_syncfunc(dst->dst_arg1, dst->dst_arg2, tx);
200 		}
201 	}
202 	rw_exit(&dp->dp_config_rwlock);
203 
204 	if (dstg->dstg_nowaiter)
205 		dsl_sync_task_group_destroy(dstg);
206 }
207 
208 int
209 dsl_sync_task_do(dsl_pool_t *dp,
210     dsl_checkfunc_t *checkfunc, dsl_syncfunc_t *syncfunc,
211     void *arg1, void *arg2, int blocks_modified)
212 {
213 	dsl_sync_task_group_t *dstg;
214 	int err;
215 
216 	ASSERT(spa_writeable(dp->dp_spa));
217 
218 	dstg = dsl_sync_task_group_create(dp);
219 	dsl_sync_task_create(dstg, checkfunc, syncfunc,
220 	    arg1, arg2, blocks_modified);
221 	err = dsl_sync_task_group_wait(dstg);
222 	dsl_sync_task_group_destroy(dstg);
223 	return (err);
224 }
225 
226 void
227 dsl_sync_task_do_nowait(dsl_pool_t *dp,
228     dsl_checkfunc_t *checkfunc, dsl_syncfunc_t *syncfunc,
229     void *arg1, void *arg2, int blocks_modified, dmu_tx_t *tx)
230 {
231 	dsl_sync_task_group_t *dstg;
232 
233 	if (!spa_writeable(dp->dp_spa))
234 		return;
235 
236 	dstg = dsl_sync_task_group_create(dp);
237 	dsl_sync_task_create(dstg, checkfunc, syncfunc,
238 	    arg1, arg2, blocks_modified);
239 	dsl_sync_task_group_nowait(dstg, tx);
240 }
241