xref: /src/sys/contrib/openzfs/module/zfs/sa.c (revision 8a62a2a5659d1839d8799b4274c04469d7f17c78)
1 // SPDX-License-Identifier: CDDL-1.0
2 /*
3  * CDDL HEADER START
4  *
5  * The contents of this file are subject to the terms of the
6  * Common Development and Distribution License (the "License").
7  * You may not use this file except in compliance with the License.
8  *
9  * You can obtain a copy of the license at usr/src/OPENSOLARIS.LICENSE
10  * or https://opensource.org/licenses/CDDL-1.0.
11  * See the License for the specific language governing permissions
12  * and limitations under the License.
13  *
14  * When distributing Covered Code, include this CDDL HEADER in each
15  * file and include the License file at usr/src/OPENSOLARIS.LICENSE.
16  * If applicable, add the following below this CDDL HEADER, with the
17  * fields enclosed by brackets "[]" replaced with your own identifying
18  * information: Portions Copyright [yyyy] [name of copyright owner]
19  *
20  * CDDL HEADER END
21  */
22 
23 /*
24  * Copyright (c) 2010, Oracle and/or its affiliates. All rights reserved.
25  * Copyright (c) 2013, 2017 by Delphix. All rights reserved.
26  * Copyright (c) 2014 Spectra Logic Corporation, All rights reserved.
27  * Copyright 2023 RackTop Systems, Inc.
28  */
29 
30 #include <sys/zfs_context.h>
31 #include <sys/types.h>
32 #include <sys/param.h>
33 #include <sys/sysmacros.h>
34 #include <sys/dmu.h>
35 #include <sys/dmu_impl.h>
36 #include <sys/dmu_objset.h>
37 #include <sys/dmu_tx.h>
38 #include <sys/dbuf.h>
39 #include <sys/dnode.h>
40 #include <sys/zap.h>
41 #include <sys/sa.h>
42 #include <sys/sunddi.h>
43 #include <sys/sa_impl.h>
44 #include <sys/errno.h>
45 #include <sys/zfs_context.h>
46 
47 #ifdef _KERNEL
48 #include <sys/zfs_znode.h>
49 #endif
50 
51 /*
52  * ZFS System attributes:
53  *
54  * A generic mechanism to allow for arbitrary attributes
55  * to be stored in a dnode.  The data will be stored in the bonus buffer of
56  * the dnode and if necessary a special "spill" block will be used to handle
57  * overflow situations.  The spill block will be sized to fit the data
58  * from 512 - 128K.  When a spill block is used the BP (blkptr_t) for the
59  * spill block is stored at the end of the current bonus buffer.  Any
60  * attributes that would be in the way of the blkptr_t will be relocated
61  * into the spill block.
62  *
63  * Attribute registration:
64  *
65  * Stored persistently on a per dataset basis
66  * a mapping between attribute "string" names and their actual attribute
67  * numeric values, length, and byteswap function.  The names are only used
68  * during registration.  All  attributes are known by their unique attribute
69  * id value.  If an attribute can have a variable size then the value
70  * 0 will be used to indicate this.
71  *
72  * Attribute Layout:
73  *
74  * Attribute layouts are a way to compactly store multiple attributes, but
75  * without taking the overhead associated with managing each attribute
76  * individually.  Since you will typically have the same set of attributes
77  * stored in the same order a single table will be used to represent that
78  * layout.  The ZPL for example will usually have only about 10 different
79  * layouts (regular files, device files, symlinks,
80  * regular files + scanstamp, files/dir with extended attributes, and then
81  * you have the possibility of all of those minus ACL, because it would
82  * be kicked out into the spill block)
83  *
84  * Layouts are simply an array of the attributes and their
85  * ordering i.e. [0, 1, 4, 5, 2]
86  *
87  * Each distinct layout is given a unique layout number and that is what's
88  * stored in the header at the beginning of the SA data buffer.
89  *
90  * A layout only covers a single dbuf (bonus or spill).  If a set of
91  * attributes is split up between the bonus buffer and a spill buffer then
92  * two different layouts will be used.  This allows us to byteswap the
93  * spill without looking at the bonus buffer and keeps the on disk format of
94  * the bonus and spill buffer the same.
95  *
96  * Adding a single attribute will cause the entire set of attributes to
97  * be rewritten and could result in a new layout number being constructed
98  * as part of the rewrite if no such layout exists for the new set of
99  * attributes.  The new attribute will be appended to the end of the already
100  * existing attributes.
101  *
102  * Both the attribute registration and attribute layout information are
103  * stored in normal ZAP attributes.  Their should be a small number of
104  * known layouts and the set of attributes is assumed to typically be quite
105  * small.
106  *
107  * The registered attributes and layout "table" information is maintained
108  * in core and a special "sa_os_t" is attached to the objset_t.
109  *
110  * A special interface is provided to allow for quickly applying
111  * a large set of attributes at once.  sa_replace_all_by_template() is
112  * used to set an array of attributes.  This is used by the ZPL when
113  * creating a brand new file.  The template that is passed into the function
114  * specifies the attribute, size for variable length attributes, location of
115  * data and special "data locator" function if the data isn't in a contiguous
116  * location.
117  *
118  * Byteswap implications:
119  *
120  * Since the SA attributes are not entirely self describing we can't do
121  * the normal byteswap processing.  The special ZAP layout attribute and
122  * attribute registration attributes define the byteswap function and the
123  * size of the attributes, unless it is variable sized.
124  * The normal ZFS byteswapping infrastructure assumes you don't need
125  * to read any objects in order to do the necessary byteswapping.  Whereas
126  * SA attributes can only be properly byteswapped if the dataset is opened
127  * and the layout/attribute ZAP attributes are available.  Because of this
128  * the SA attributes will be byteswapped when they are first accessed by
129  * the SA code that will read the SA data.
130  */
131 
132 typedef void (sa_iterfunc_t)(void *hdr, void *addr, sa_attr_type_t,
133     uint16_t length, int length_idx, boolean_t, void *userp);
134 
135 static int sa_build_index(sa_handle_t *hdl, sa_buf_type_t buftype);
136 static void sa_idx_tab_hold(objset_t *os, sa_idx_tab_t *idx_tab);
137 static sa_idx_tab_t *sa_find_idx_tab(objset_t *os, dmu_object_type_t bonustype,
138     sa_hdr_phys_t *hdr);
139 static void sa_idx_tab_rele(objset_t *os, void *arg);
140 static void sa_copy_data(sa_data_locator_t *func, void *start, void *target,
141     int buflen);
142 static int sa_modify_attrs(sa_handle_t *hdl, sa_attr_type_t newattr,
143     sa_data_op_t action, sa_data_locator_t *locator, void *datastart,
144     uint16_t buflen, dmu_tx_t *tx);
145 
146 static arc_byteswap_func_t sa_bswap_table[] = {
147 	byteswap_uint64_array,
148 	byteswap_uint32_array,
149 	byteswap_uint16_array,
150 	byteswap_uint8_array,
151 	zfs_acl_byteswap,
152 };
153 
154 #ifdef HAVE_EFFICIENT_UNALIGNED_ACCESS
155 #define	SA_COPY_DATA(f, s, t, l)				\
156 do {								\
157 	if (f == NULL) {					\
158 		if (l == 8) {					\
159 			*(uint64_t *)t = *(uint64_t *)s;	\
160 		} else if (l == 16) {				\
161 			*(uint64_t *)t = *(uint64_t *)s;	\
162 			*(uint64_t *)((uintptr_t)t + 8) =	\
163 			    *(uint64_t *)((uintptr_t)s + 8);	\
164 		} else {					\
165 			memcpy(t, s, l);				\
166 		}						\
167 	} else {						\
168 		sa_copy_data(f, s, t, l);			\
169 	}							\
170 } while (0)
171 #else
172 #define	SA_COPY_DATA(f, s, t, l)	sa_copy_data(f, s, t, l)
173 #endif
174 
175 /*
176  * This table is fixed and cannot be changed.  Its purpose is to
177  * allow the SA code to work with both old/new ZPL file systems.
178  * It contains the list of legacy attributes.  These attributes aren't
179  * stored in the "attribute" registry zap objects, since older ZPL file systems
180  * won't have the registry.  Only objsets of type ZFS_TYPE_FILESYSTEM will
181  * use this static table.
182  */
183 static const sa_attr_reg_t sa_legacy_attrs[] = {
184 	{"ZPL_ATIME", sizeof (uint64_t) * 2, SA_UINT64_ARRAY, 0},
185 	{"ZPL_MTIME", sizeof (uint64_t) * 2, SA_UINT64_ARRAY, 1},
186 	{"ZPL_CTIME", sizeof (uint64_t) * 2, SA_UINT64_ARRAY, 2},
187 	{"ZPL_CRTIME", sizeof (uint64_t) * 2, SA_UINT64_ARRAY, 3},
188 	{"ZPL_GEN", sizeof (uint64_t), SA_UINT64_ARRAY, 4},
189 	{"ZPL_MODE", sizeof (uint64_t), SA_UINT64_ARRAY, 5},
190 	{"ZPL_SIZE", sizeof (uint64_t), SA_UINT64_ARRAY, 6},
191 	{"ZPL_PARENT", sizeof (uint64_t), SA_UINT64_ARRAY, 7},
192 	{"ZPL_LINKS", sizeof (uint64_t), SA_UINT64_ARRAY, 8},
193 	{"ZPL_XATTR", sizeof (uint64_t), SA_UINT64_ARRAY, 9},
194 	{"ZPL_RDEV", sizeof (uint64_t), SA_UINT64_ARRAY, 10},
195 	{"ZPL_FLAGS", sizeof (uint64_t), SA_UINT64_ARRAY, 11},
196 	{"ZPL_UID", sizeof (uint64_t), SA_UINT64_ARRAY, 12},
197 	{"ZPL_GID", sizeof (uint64_t), SA_UINT64_ARRAY, 13},
198 	{"ZPL_PAD", sizeof (uint64_t) * 4, SA_UINT64_ARRAY, 14},
199 	{"ZPL_ZNODE_ACL", 88, SA_UINT8_ARRAY, 15},
200 };
201 
202 /*
203  * This is only used for objects of type DMU_OT_ZNODE
204  */
205 static const sa_attr_type_t sa_legacy_zpl_layout[] = {
206     0, 1, 2, 3, 4, 5, 6, 7, 8, 9, 10, 11, 12, 13, 14, 15
207 };
208 
209 /*
210  * Special dummy layout used for buffers with no attributes.
211  */
212 static const sa_attr_type_t sa_dummy_zpl_layout[] = { 0 };
213 
214 static const size_t sa_legacy_attr_count = ARRAY_SIZE(sa_legacy_attrs);
215 static kmem_cache_t *sa_cache = NULL;
216 
217 static int
sa_cache_constructor(void * buf,void * unused,int kmflag)218 sa_cache_constructor(void *buf, void *unused, int kmflag)
219 {
220 	(void) unused, (void) kmflag;
221 	sa_handle_t *hdl = buf;
222 
223 	mutex_init(&hdl->sa_lock, NULL, MUTEX_DEFAULT, NULL);
224 	return (0);
225 }
226 
227 static void
sa_cache_destructor(void * buf,void * unused)228 sa_cache_destructor(void *buf, void *unused)
229 {
230 	(void) unused;
231 	sa_handle_t *hdl = buf;
232 	mutex_destroy(&hdl->sa_lock);
233 }
234 
235 void
sa_cache_init(void)236 sa_cache_init(void)
237 {
238 	sa_cache = kmem_cache_create("sa_cache",
239 	    sizeof (sa_handle_t), 0, sa_cache_constructor,
240 	    sa_cache_destructor, NULL, NULL, NULL, KMC_RECLAIMABLE);
241 }
242 
243 void
sa_cache_fini(void)244 sa_cache_fini(void)
245 {
246 	if (sa_cache)
247 		kmem_cache_destroy(sa_cache);
248 }
249 
250 static int
layout_num_compare(const void * arg1,const void * arg2)251 layout_num_compare(const void *arg1, const void *arg2)
252 {
253 	const sa_lot_t *node1 = (const sa_lot_t *)arg1;
254 	const sa_lot_t *node2 = (const sa_lot_t *)arg2;
255 
256 	return (TREE_CMP(node1->lot_num, node2->lot_num));
257 }
258 
259 static int
layout_hash_compare(const void * arg1,const void * arg2)260 layout_hash_compare(const void *arg1, const void *arg2)
261 {
262 	const sa_lot_t *node1 = (const sa_lot_t *)arg1;
263 	const sa_lot_t *node2 = (const sa_lot_t *)arg2;
264 
265 	int cmp = TREE_CMP(node1->lot_hash, node2->lot_hash);
266 	if (likely(cmp))
267 		return (cmp);
268 
269 	return (TREE_CMP(node1->lot_instance, node2->lot_instance));
270 }
271 
272 static boolean_t
sa_layout_equal(sa_lot_t * tbf,sa_attr_type_t * attrs,int count)273 sa_layout_equal(sa_lot_t *tbf, sa_attr_type_t *attrs, int count)
274 {
275 	int i;
276 
277 	if (count != tbf->lot_attr_count)
278 		return (1);
279 
280 	for (i = 0; i != count; i++) {
281 		if (attrs[i] != tbf->lot_attrs[i])
282 			return (1);
283 	}
284 	return (0);
285 }
286 
287 #define	SA_ATTR_HASH(attr) (zfs_crc64_table[(-1ULL ^ attr) & 0xFF])
288 
289 static uint64_t
sa_layout_info_hash(const sa_attr_type_t * attrs,int attr_count)290 sa_layout_info_hash(const sa_attr_type_t *attrs, int attr_count)
291 {
292 	uint64_t crc = -1ULL;
293 
294 	for (int i = 0; i != attr_count; i++)
295 		crc ^= SA_ATTR_HASH(attrs[i]);
296 
297 	return (crc);
298 }
299 
300 static int
sa_get_spill(sa_handle_t * hdl)301 sa_get_spill(sa_handle_t *hdl)
302 {
303 	int rc;
304 	if (hdl->sa_spill == NULL) {
305 		if ((rc = dmu_spill_hold_existing(hdl->sa_bonus, NULL,
306 		    &hdl->sa_spill)) == 0)
307 			VERIFY0(sa_build_index(hdl, SA_SPILL));
308 	} else {
309 		rc = 0;
310 	}
311 
312 	return (rc);
313 }
314 
315 /*
316  * Main attribute lookup/update function
317  * returns 0 for success or non zero for failures
318  *
319  * Operates on bulk array, first failure will abort further processing
320  */
321 static int
sa_attr_op(sa_handle_t * hdl,sa_bulk_attr_t * bulk,int count,sa_data_op_t data_op,dmu_tx_t * tx)322 sa_attr_op(sa_handle_t *hdl, sa_bulk_attr_t *bulk, int count,
323     sa_data_op_t data_op, dmu_tx_t *tx)
324 {
325 	sa_os_t *sa = hdl->sa_os->os_sa;
326 	int i;
327 	int error = 0;
328 	sa_buf_type_t buftypes;
329 
330 	buftypes = 0;
331 
332 	ASSERT(count > 0);
333 	for (i = 0; i != count; i++) {
334 		ASSERT(bulk[i].sa_attr <= hdl->sa_os->os_sa->sa_num_attrs);
335 
336 		bulk[i].sa_addr = NULL;
337 		/* First check the bonus buffer */
338 
339 		if (hdl->sa_bonus_tab && TOC_ATTR_PRESENT(
340 		    hdl->sa_bonus_tab->sa_idx_tab[bulk[i].sa_attr])) {
341 			SA_ATTR_INFO(sa, hdl->sa_bonus_tab,
342 			    SA_GET_HDR(hdl, SA_BONUS),
343 			    bulk[i].sa_attr, bulk[i], SA_BONUS, hdl);
344 			if (tx && !(buftypes & SA_BONUS)) {
345 				dmu_buf_will_dirty(hdl->sa_bonus, tx);
346 				buftypes |= SA_BONUS;
347 			}
348 		}
349 		if (bulk[i].sa_addr == NULL &&
350 		    ((error = sa_get_spill(hdl)) == 0)) {
351 			if (TOC_ATTR_PRESENT(
352 			    hdl->sa_spill_tab->sa_idx_tab[bulk[i].sa_attr])) {
353 				SA_ATTR_INFO(sa, hdl->sa_spill_tab,
354 				    SA_GET_HDR(hdl, SA_SPILL),
355 				    bulk[i].sa_attr, bulk[i], SA_SPILL, hdl);
356 				if (tx && !(buftypes & SA_SPILL) &&
357 				    bulk[i].sa_size == bulk[i].sa_length) {
358 					dmu_buf_will_dirty(hdl->sa_spill, tx);
359 					buftypes |= SA_SPILL;
360 				}
361 			}
362 		}
363 		if (error && error != ENOENT) {
364 			return ((error == ECKSUM) ? EIO : error);
365 		}
366 
367 		switch (data_op) {
368 		case SA_LOOKUP:
369 			if (bulk[i].sa_addr == NULL)
370 				return (SET_ERROR(ENOENT));
371 			if (bulk[i].sa_data) {
372 				SA_COPY_DATA(bulk[i].sa_data_func,
373 				    bulk[i].sa_addr, bulk[i].sa_data,
374 				    MIN(bulk[i].sa_size, bulk[i].sa_length));
375 			}
376 			continue;
377 
378 		case SA_UPDATE:
379 			/* existing rewrite of attr */
380 			if (bulk[i].sa_addr &&
381 			    bulk[i].sa_size == bulk[i].sa_length) {
382 				SA_COPY_DATA(bulk[i].sa_data_func,
383 				    bulk[i].sa_data, bulk[i].sa_addr,
384 				    bulk[i].sa_length);
385 				continue;
386 			} else if (bulk[i].sa_addr) { /* attr size change */
387 				error = sa_modify_attrs(hdl, bulk[i].sa_attr,
388 				    SA_REPLACE, bulk[i].sa_data_func,
389 				    bulk[i].sa_data, bulk[i].sa_length, tx);
390 			} else { /* adding new attribute */
391 				error = sa_modify_attrs(hdl, bulk[i].sa_attr,
392 				    SA_ADD, bulk[i].sa_data_func,
393 				    bulk[i].sa_data, bulk[i].sa_length, tx);
394 			}
395 			if (error)
396 				return (error);
397 			break;
398 		default:
399 			break;
400 		}
401 	}
402 	return (error);
403 }
404 
405 static sa_lot_t *
sa_add_layout_entry(objset_t * os,const sa_attr_type_t * attrs,int attr_count,uint64_t lot_num,uint64_t hash,boolean_t zapadd,dmu_tx_t * tx)406 sa_add_layout_entry(objset_t *os, const sa_attr_type_t *attrs, int attr_count,
407     uint64_t lot_num, uint64_t hash, boolean_t zapadd, dmu_tx_t *tx)
408 {
409 	sa_os_t *sa = os->os_sa;
410 	sa_lot_t *tb, *findtb;
411 	int i;
412 	avl_index_t loc;
413 
414 	ASSERT(MUTEX_HELD(&sa->sa_lock));
415 	tb = kmem_zalloc(sizeof (sa_lot_t), KM_SLEEP);
416 	tb->lot_attr_count = attr_count;
417 	tb->lot_attrs = kmem_alloc(sizeof (sa_attr_type_t) * attr_count,
418 	    KM_SLEEP);
419 	memcpy(tb->lot_attrs, attrs, sizeof (sa_attr_type_t) * attr_count);
420 	tb->lot_num = lot_num;
421 	tb->lot_hash = hash;
422 	tb->lot_instance = 0;
423 
424 	if (zapadd) {
425 		char attr_name[8];
426 
427 		if (sa->sa_layout_attr_obj == 0) {
428 			sa->sa_layout_attr_obj = zap_create_link(os,
429 			    DMU_OT_SA_ATTR_LAYOUTS,
430 			    sa->sa_master_obj, SA_LAYOUTS, tx);
431 		}
432 
433 		(void) snprintf(attr_name, sizeof (attr_name),
434 		    "%d", (int)lot_num);
435 		VERIFY0(zap_update(os, os->os_sa->sa_layout_attr_obj,
436 		    attr_name, 2, attr_count, attrs, tx));
437 	}
438 
439 	list_create(&tb->lot_idx_tab, sizeof (sa_idx_tab_t),
440 	    offsetof(sa_idx_tab_t, sa_next));
441 
442 	for (i = 0; i != attr_count; i++) {
443 		if (sa->sa_attr_table[tb->lot_attrs[i]].sa_length == 0)
444 			tb->lot_var_sizes++;
445 	}
446 
447 	avl_add(&sa->sa_layout_num_tree, tb);
448 
449 	/* verify we don't have a hash collision */
450 	if ((findtb = avl_find(&sa->sa_layout_hash_tree, tb, &loc)) != NULL) {
451 		for (; findtb && findtb->lot_hash == hash;
452 		    findtb = AVL_NEXT(&sa->sa_layout_hash_tree, findtb)) {
453 			if (findtb->lot_instance != tb->lot_instance)
454 				break;
455 			tb->lot_instance++;
456 		}
457 	}
458 	avl_add(&sa->sa_layout_hash_tree, tb);
459 	return (tb);
460 }
461 
462 static void
sa_find_layout(objset_t * os,uint64_t hash,sa_attr_type_t * attrs,int count,dmu_tx_t * tx,sa_lot_t ** lot)463 sa_find_layout(objset_t *os, uint64_t hash, sa_attr_type_t *attrs,
464     int count, dmu_tx_t *tx, sa_lot_t **lot)
465 {
466 	sa_lot_t *tb, tbsearch;
467 	avl_index_t loc;
468 	sa_os_t *sa = os->os_sa;
469 	boolean_t found = B_FALSE;
470 
471 	mutex_enter(&sa->sa_lock);
472 	tbsearch.lot_hash = hash;
473 	tbsearch.lot_instance = 0;
474 	tb = avl_find(&sa->sa_layout_hash_tree, &tbsearch, &loc);
475 	if (tb) {
476 		for (; tb && tb->lot_hash == hash;
477 		    tb = AVL_NEXT(&sa->sa_layout_hash_tree, tb)) {
478 			if (sa_layout_equal(tb, attrs, count) == 0) {
479 				found = B_TRUE;
480 				break;
481 			}
482 		}
483 	}
484 	if (!found) {
485 		tb = sa_add_layout_entry(os, attrs, count,
486 		    avl_numnodes(&sa->sa_layout_num_tree), hash, B_TRUE, tx);
487 	}
488 	mutex_exit(&sa->sa_lock);
489 	*lot = tb;
490 }
491 
492 static int
sa_resize_spill(sa_handle_t * hdl,uint32_t size,dmu_tx_t * tx)493 sa_resize_spill(sa_handle_t *hdl, uint32_t size, dmu_tx_t *tx)
494 {
495 	int error;
496 	uint32_t blocksize;
497 
498 	if (size == 0) {
499 		blocksize = SPA_MINBLOCKSIZE;
500 	} else if (size > SPA_OLD_MAXBLOCKSIZE) {
501 		ASSERT(0);
502 		return (SET_ERROR(EFBIG));
503 	} else {
504 		blocksize = P2ROUNDUP_TYPED(size, SPA_MINBLOCKSIZE, uint32_t);
505 	}
506 
507 	error = dbuf_spill_set_blksz(hdl->sa_spill, blocksize, tx);
508 	ASSERT0(error);
509 	return (error);
510 }
511 
512 static void
sa_copy_data(sa_data_locator_t * func,void * datastart,void * target,int buflen)513 sa_copy_data(sa_data_locator_t *func, void *datastart, void *target, int buflen)
514 {
515 	if (func == NULL) {
516 		memcpy(target, datastart, buflen);
517 	} else {
518 		boolean_t start;
519 		int bytes;
520 		void *dataptr;
521 		void *saptr = target;
522 		uint32_t length;
523 
524 		start = B_TRUE;
525 		bytes = 0;
526 		while (bytes < buflen) {
527 			func(&dataptr, &length, buflen, start, datastart);
528 			memcpy(saptr, dataptr, length);
529 			saptr = (void *)((caddr_t)saptr + length);
530 			bytes += length;
531 			start = B_FALSE;
532 		}
533 	}
534 }
535 
536 /*
537  * Determine several different values pertaining to system attribute
538  * buffers.
539  *
540  * Return the size of the sa_hdr_phys_t header for the buffer. Each
541  * variable length attribute except the first contributes two bytes to
542  * the header size, which is then rounded up to an 8-byte boundary.
543  *
544  * The following output parameters are also computed.
545  *
546  *  index - The index of the first attribute in attr_desc that will
547  *  spill over. Only valid if will_spill is set.
548  *
549  *  total - The total number of bytes of all system attributes described
550  *  in attr_desc.
551  *
552  *  will_spill - Set when spilling is necessary. It is only set when
553  *  the buftype is SA_BONUS.
554  */
555 static int
sa_find_sizes(sa_os_t * sa,sa_bulk_attr_t * attr_desc,int attr_count,dmu_buf_t * db,sa_buf_type_t buftype,int full_space,int * index,int * total,boolean_t * will_spill)556 sa_find_sizes(sa_os_t *sa, sa_bulk_attr_t *attr_desc, int attr_count,
557     dmu_buf_t *db, sa_buf_type_t buftype, int full_space, int *index,
558     int *total, boolean_t *will_spill)
559 {
560 	int var_size_count = 0;
561 	int i;
562 	int hdrsize;
563 	int extra_hdrsize;
564 
565 	if (buftype == SA_BONUS && sa->sa_force_spill) {
566 		*total = 0;
567 		*index = 0;
568 		*will_spill = B_TRUE;
569 		return (0);
570 	}
571 
572 	*index = -1;
573 	*total = 0;
574 	*will_spill = B_FALSE;
575 
576 	extra_hdrsize = 0;
577 	hdrsize = (SA_BONUSTYPE_FROM_DB(db) == DMU_OT_ZNODE) ? 0 :
578 	    sizeof (sa_hdr_phys_t);
579 
580 	ASSERT(IS_P2ALIGNED(full_space, 8));
581 
582 	for (i = 0; i != attr_count; i++) {
583 		boolean_t is_var_sz, might_spill_here;
584 		int tmp_hdrsize;
585 
586 		*total = P2ROUNDUP(*total, 8);
587 		*total += attr_desc[i].sa_length;
588 		if (*will_spill)
589 			continue;
590 
591 		is_var_sz = (SA_REGISTERED_LEN(sa, attr_desc[i].sa_attr) == 0);
592 		if (is_var_sz)
593 			var_size_count++;
594 
595 		/*
596 		 * Calculate what the SA header size would be if this
597 		 * attribute doesn't spill.
598 		 */
599 		tmp_hdrsize = hdrsize + ((is_var_sz && var_size_count > 1) ?
600 		    sizeof (uint16_t) : 0);
601 
602 		/*
603 		 * Check whether this attribute spans into the space
604 		 * that would be used by the spill block pointer should
605 		 * a spill block be needed.
606 		 */
607 		might_spill_here =
608 		    buftype == SA_BONUS && *index == -1 &&
609 		    (*total + P2ROUNDUP(tmp_hdrsize, 8)) >
610 		    (full_space - sizeof (blkptr_t));
611 
612 		if (is_var_sz && var_size_count > 1) {
613 			if (buftype == SA_SPILL ||
614 			    tmp_hdrsize + *total < full_space) {
615 				/*
616 				 * Record the extra header size in case this
617 				 * increase needs to be reversed due to
618 				 * spill-over.
619 				 */
620 				hdrsize = tmp_hdrsize;
621 				if (*index != -1 || might_spill_here)
622 					extra_hdrsize += sizeof (uint16_t);
623 			} else {
624 				ASSERT(buftype == SA_BONUS);
625 				if (*index == -1)
626 					*index = i;
627 				*will_spill = B_TRUE;
628 				continue;
629 			}
630 		}
631 
632 		/*
633 		 * Store index of where spill *could* occur. Then
634 		 * continue to count the remaining attribute sizes. The
635 		 * sum is used later for sizing bonus and spill buffer.
636 		 */
637 		if (might_spill_here)
638 			*index = i;
639 
640 		if ((*total + P2ROUNDUP(hdrsize, 8)) > full_space &&
641 		    buftype == SA_BONUS)
642 			*will_spill = B_TRUE;
643 	}
644 
645 	if (*will_spill)
646 		hdrsize -= extra_hdrsize;
647 
648 	hdrsize = P2ROUNDUP(hdrsize, 8);
649 	return (hdrsize);
650 }
651 
652 #define	BUF_SPACE_NEEDED(total, header) (total + header)
653 
654 /*
655  * Find layout that corresponds to ordering of attributes
656  * If not found a new layout number is created and added to
657  * persistent layout tables.
658  */
659 static int
sa_build_layouts(sa_handle_t * hdl,sa_bulk_attr_t * attr_desc,int attr_count,dmu_tx_t * tx)660 sa_build_layouts(sa_handle_t *hdl, sa_bulk_attr_t *attr_desc, int attr_count,
661     dmu_tx_t *tx)
662 {
663 	sa_os_t *sa = hdl->sa_os->os_sa;
664 	uint64_t hash;
665 	sa_buf_type_t buftype;
666 	sa_hdr_phys_t *sahdr;
667 	void *data_start;
668 	sa_attr_type_t *attrs, *attrs_start;
669 	int i, lot_count;
670 	int dnodesize;
671 	int spill_idx;
672 	int hdrsize;
673 	int spillhdrsize = 0;
674 	int used;
675 	dmu_object_type_t bonustype;
676 	sa_lot_t *lot;
677 	int len_idx;
678 	int spill_used;
679 	int bonuslen;
680 	boolean_t spilling;
681 
682 	dmu_buf_will_dirty(hdl->sa_bonus, tx);
683 	bonustype = SA_BONUSTYPE_FROM_DB(hdl->sa_bonus);
684 	dmu_object_dnsize_from_db(hdl->sa_bonus, &dnodesize);
685 	bonuslen = DN_BONUS_SIZE(dnodesize);
686 
687 	/* first determine bonus header size and sum of all attributes */
688 	hdrsize = sa_find_sizes(sa, attr_desc, attr_count, hdl->sa_bonus,
689 	    SA_BONUS, bonuslen, &spill_idx, &used, &spilling);
690 
691 	if (used > SPA_OLD_MAXBLOCKSIZE)
692 		return (SET_ERROR(EFBIG));
693 
694 	VERIFY0(dmu_set_bonus(hdl->sa_bonus, spilling ?
695 	    MIN(bonuslen - sizeof (blkptr_t), used + hdrsize) :
696 	    used + hdrsize, tx));
697 
698 	ASSERT((bonustype == DMU_OT_ZNODE && spilling == 0) ||
699 	    bonustype == DMU_OT_SA);
700 
701 	/* setup and size spill buffer when needed */
702 	if (spilling) {
703 		boolean_t dummy;
704 
705 		if (hdl->sa_spill == NULL) {
706 			VERIFY0(dmu_spill_hold_by_bonus(hdl->sa_bonus,
707 			    DB_RF_MUST_SUCCEED, NULL, &hdl->sa_spill));
708 		}
709 		dmu_buf_will_dirty(hdl->sa_spill, tx);
710 
711 		spillhdrsize = sa_find_sizes(sa, &attr_desc[spill_idx],
712 		    attr_count - spill_idx, hdl->sa_spill, SA_SPILL,
713 		    hdl->sa_spill->db_size, &i, &spill_used, &dummy);
714 
715 		if (spill_used > SPA_OLD_MAXBLOCKSIZE)
716 			return (SET_ERROR(EFBIG));
717 
718 		if (BUF_SPACE_NEEDED(spill_used, spillhdrsize) >
719 		    hdl->sa_spill->db_size)
720 			VERIFY0(sa_resize_spill(hdl,
721 			    BUF_SPACE_NEEDED(spill_used, spillhdrsize), tx));
722 	}
723 
724 	/* setup starting pointers to lay down data */
725 	data_start = (void *)((uintptr_t)hdl->sa_bonus->db_data + hdrsize);
726 	sahdr = (sa_hdr_phys_t *)hdl->sa_bonus->db_data;
727 	buftype = SA_BONUS;
728 
729 	attrs_start = attrs = kmem_alloc(sizeof (sa_attr_type_t) * attr_count,
730 	    KM_SLEEP);
731 	lot_count = 0;
732 
733 	for (i = 0, len_idx = 0, hash = -1ULL; i != attr_count; i++) {
734 		uint16_t length;
735 
736 		ASSERT(IS_P2ALIGNED(data_start, 8));
737 		attrs[i] = attr_desc[i].sa_attr;
738 		length = SA_REGISTERED_LEN(sa, attrs[i]);
739 		if (length == 0)
740 			length = attr_desc[i].sa_length;
741 
742 		if (spilling && i == spill_idx) { /* switch to spill buffer */
743 			VERIFY(bonustype == DMU_OT_SA);
744 			if (buftype == SA_BONUS && !sa->sa_force_spill) {
745 				sa_find_layout(hdl->sa_os, hash, attrs_start,
746 				    lot_count, tx, &lot);
747 				SA_SET_HDR(sahdr, lot->lot_num, hdrsize);
748 			}
749 
750 			buftype = SA_SPILL;
751 			hash = -1ULL;
752 			len_idx = 0;
753 
754 			sahdr = (sa_hdr_phys_t *)hdl->sa_spill->db_data;
755 			sahdr->sa_magic = SA_MAGIC;
756 			data_start = (void *)((uintptr_t)sahdr +
757 			    spillhdrsize);
758 			attrs_start = &attrs[i];
759 			lot_count = 0;
760 		}
761 		hash ^= SA_ATTR_HASH(attrs[i]);
762 		attr_desc[i].sa_addr = data_start;
763 		attr_desc[i].sa_size = length;
764 		SA_COPY_DATA(attr_desc[i].sa_data_func, attr_desc[i].sa_data,
765 		    data_start, length);
766 		if (sa->sa_attr_table[attrs[i]].sa_length == 0) {
767 			sahdr->sa_lengths[len_idx++] = length;
768 		}
769 		data_start = (void *)P2ROUNDUP(((uintptr_t)data_start +
770 		    length), 8);
771 		lot_count++;
772 	}
773 
774 	sa_find_layout(hdl->sa_os, hash, attrs_start, lot_count, tx, &lot);
775 
776 	/*
777 	 * Verify that old znodes always have layout number 0.
778 	 * Must be DMU_OT_SA for arbitrary layouts
779 	 */
780 	VERIFY((bonustype == DMU_OT_ZNODE && lot->lot_num == 0) ||
781 	    (bonustype == DMU_OT_SA && lot->lot_num > 1));
782 
783 	if (bonustype == DMU_OT_SA) {
784 		SA_SET_HDR(sahdr, lot->lot_num,
785 		    buftype == SA_BONUS ? hdrsize : spillhdrsize);
786 	}
787 
788 	kmem_free(attrs, sizeof (sa_attr_type_t) * attr_count);
789 	if (hdl->sa_bonus_tab) {
790 		sa_idx_tab_rele(hdl->sa_os, hdl->sa_bonus_tab);
791 		hdl->sa_bonus_tab = NULL;
792 	}
793 	if (!sa->sa_force_spill)
794 		VERIFY0(sa_build_index(hdl, SA_BONUS));
795 	if (hdl->sa_spill) {
796 		sa_idx_tab_rele(hdl->sa_os, hdl->sa_spill_tab);
797 		if (!spilling) {
798 			/*
799 			 * remove spill block that is no longer needed.
800 			 */
801 			dmu_buf_rele(hdl->sa_spill, NULL);
802 			hdl->sa_spill = NULL;
803 			hdl->sa_spill_tab = NULL;
804 			VERIFY0(dmu_rm_spill(hdl->sa_os,
805 			    sa_handle_object(hdl), tx));
806 		} else {
807 			VERIFY0(sa_build_index(hdl, SA_SPILL));
808 		}
809 	}
810 
811 	return (0);
812 }
813 
814 static void
sa_free_attr_table(sa_os_t * sa)815 sa_free_attr_table(sa_os_t *sa)
816 {
817 	int i;
818 
819 	if (sa->sa_attr_table == NULL)
820 		return;
821 
822 	for (i = 0; i != sa->sa_num_attrs; i++) {
823 		if (sa->sa_attr_table[i].sa_name)
824 			kmem_free(sa->sa_attr_table[i].sa_name,
825 			    strlen(sa->sa_attr_table[i].sa_name) + 1);
826 	}
827 
828 	kmem_free(sa->sa_attr_table,
829 	    sizeof (sa_attr_table_t) * sa->sa_num_attrs);
830 
831 	sa->sa_attr_table = NULL;
832 }
833 
834 static int
sa_attr_table_setup(objset_t * os,const sa_attr_reg_t * reg_attrs,int count)835 sa_attr_table_setup(objset_t *os, const sa_attr_reg_t *reg_attrs, int count)
836 {
837 	sa_os_t *sa = os->os_sa;
838 	uint64_t sa_attr_count = 0;
839 	uint64_t sa_reg_count = 0;
840 	int error = 0;
841 	uint64_t attr_value;
842 	sa_attr_table_t *tb;
843 	zap_cursor_t zc;
844 	zap_attribute_t *za;
845 	int registered_count = 0;
846 	int i;
847 	dmu_objset_type_t ostype = dmu_objset_type(os);
848 
849 	sa->sa_user_table =
850 	    kmem_zalloc(count * sizeof (sa_attr_type_t), KM_SLEEP);
851 	sa->sa_user_table_sz = count * sizeof (sa_attr_type_t);
852 
853 	if (sa->sa_reg_attr_obj != 0) {
854 		error = zap_count(os, sa->sa_reg_attr_obj,
855 		    &sa_attr_count);
856 
857 		/*
858 		 * Make sure we retrieved a count and that it isn't zero
859 		 */
860 		if (error || (error == 0 && sa_attr_count == 0)) {
861 			if (error == 0)
862 				error = SET_ERROR(EINVAL);
863 			goto bail;
864 		}
865 		sa_reg_count = sa_attr_count;
866 	}
867 
868 	if (ostype == DMU_OST_ZFS && sa_attr_count == 0)
869 		sa_attr_count += sa_legacy_attr_count;
870 
871 	/* Allocate attribute numbers for attributes that aren't registered */
872 	for (i = 0; i != count; i++) {
873 		boolean_t found = B_FALSE;
874 		int j;
875 
876 		if (ostype == DMU_OST_ZFS) {
877 			for (j = 0; j != sa_legacy_attr_count; j++) {
878 				if (strcmp(reg_attrs[i].sa_name,
879 				    sa_legacy_attrs[j].sa_name) == 0) {
880 					sa->sa_user_table[i] =
881 					    sa_legacy_attrs[j].sa_attr;
882 					found = B_TRUE;
883 				}
884 			}
885 		}
886 		if (found)
887 			continue;
888 
889 		if (sa->sa_reg_attr_obj)
890 			error = zap_lookup(os, sa->sa_reg_attr_obj,
891 			    reg_attrs[i].sa_name, 8, 1, &attr_value);
892 		else
893 			error = SET_ERROR(ENOENT);
894 		switch (error) {
895 		case ENOENT:
896 			sa->sa_user_table[i] = (sa_attr_type_t)sa_attr_count;
897 			sa_attr_count++;
898 			break;
899 		case 0:
900 			sa->sa_user_table[i] = ATTR_NUM(attr_value);
901 			break;
902 		default:
903 			goto bail;
904 		}
905 	}
906 
907 	sa->sa_num_attrs = sa_attr_count;
908 	tb = sa->sa_attr_table =
909 	    kmem_zalloc(sizeof (sa_attr_table_t) * sa_attr_count, KM_SLEEP);
910 
911 	/*
912 	 * Attribute table is constructed from requested attribute list,
913 	 * previously foreign registered attributes, and also the legacy
914 	 * ZPL set of attributes.
915 	 */
916 
917 	if (sa->sa_reg_attr_obj) {
918 		za = zap_attribute_alloc();
919 		for (zap_cursor_init(&zc, os, sa->sa_reg_attr_obj);
920 		    (error = zap_cursor_retrieve(&zc, za)) == 0;
921 		    zap_cursor_advance(&zc)) {
922 			uint64_t value;
923 			value  = za->za_first_integer;
924 
925 			registered_count++;
926 			tb[ATTR_NUM(value)].sa_attr = ATTR_NUM(value);
927 			tb[ATTR_NUM(value)].sa_length = ATTR_LENGTH(value);
928 			tb[ATTR_NUM(value)].sa_byteswap = ATTR_BSWAP(value);
929 			tb[ATTR_NUM(value)].sa_registered = B_TRUE;
930 
931 			if (tb[ATTR_NUM(value)].sa_name) {
932 				continue;
933 			}
934 			tb[ATTR_NUM(value)].sa_name =
935 			    kmem_zalloc(strlen(za->za_name) +1, KM_SLEEP);
936 			(void) strlcpy(tb[ATTR_NUM(value)].sa_name, za->za_name,
937 			    strlen(za->za_name) +1);
938 		}
939 		zap_cursor_fini(&zc);
940 		zap_attribute_free(za);
941 		/*
942 		 * Make sure we processed the correct number of registered
943 		 * attributes
944 		 */
945 		if (registered_count != sa_reg_count) {
946 			ASSERT(error != 0);
947 			goto bail;
948 		}
949 
950 	}
951 
952 	if (ostype == DMU_OST_ZFS) {
953 		for (i = 0; i != sa_legacy_attr_count; i++) {
954 			if (tb[i].sa_name)
955 				continue;
956 			tb[i].sa_attr = sa_legacy_attrs[i].sa_attr;
957 			tb[i].sa_length = sa_legacy_attrs[i].sa_length;
958 			tb[i].sa_byteswap = sa_legacy_attrs[i].sa_byteswap;
959 			tb[i].sa_registered = B_FALSE;
960 			tb[i].sa_name =
961 			    kmem_zalloc(strlen(sa_legacy_attrs[i].sa_name) +1,
962 			    KM_SLEEP);
963 			(void) strlcpy(tb[i].sa_name,
964 			    sa_legacy_attrs[i].sa_name,
965 			    strlen(sa_legacy_attrs[i].sa_name) + 1);
966 		}
967 	}
968 
969 	for (i = 0; i != count; i++) {
970 		sa_attr_type_t attr_id;
971 
972 		attr_id = sa->sa_user_table[i];
973 		if (tb[attr_id].sa_name)
974 			continue;
975 
976 		tb[attr_id].sa_length = reg_attrs[i].sa_length;
977 		tb[attr_id].sa_byteswap = reg_attrs[i].sa_byteswap;
978 		tb[attr_id].sa_attr = attr_id;
979 		tb[attr_id].sa_name =
980 		    kmem_zalloc(strlen(reg_attrs[i].sa_name) + 1, KM_SLEEP);
981 		(void) strlcpy(tb[attr_id].sa_name, reg_attrs[i].sa_name,
982 		    strlen(reg_attrs[i].sa_name) + 1);
983 	}
984 
985 	sa->sa_need_attr_registration =
986 	    (sa_attr_count != registered_count);
987 
988 	return (0);
989 bail:
990 	kmem_free(sa->sa_user_table, count * sizeof (sa_attr_type_t));
991 	sa->sa_user_table = NULL;
992 	sa_free_attr_table(sa);
993 	ASSERT(error != 0);
994 	return (error);
995 }
996 
997 int
sa_setup(objset_t * os,uint64_t sa_obj,const sa_attr_reg_t * reg_attrs,int count,sa_attr_type_t ** user_table)998 sa_setup(objset_t *os, uint64_t sa_obj, const sa_attr_reg_t *reg_attrs,
999     int count, sa_attr_type_t **user_table)
1000 {
1001 	zap_cursor_t zc;
1002 	zap_attribute_t *za;
1003 	sa_os_t *sa;
1004 	dmu_objset_type_t ostype = dmu_objset_type(os);
1005 	sa_attr_type_t *tb;
1006 	int error;
1007 
1008 	mutex_enter(&os->os_user_ptr_lock);
1009 	if (os->os_sa) {
1010 		mutex_enter(&os->os_sa->sa_lock);
1011 		mutex_exit(&os->os_user_ptr_lock);
1012 		tb = os->os_sa->sa_user_table;
1013 		mutex_exit(&os->os_sa->sa_lock);
1014 		*user_table = tb;
1015 		return (0);
1016 	}
1017 
1018 	sa = kmem_zalloc(sizeof (sa_os_t), KM_SLEEP);
1019 	mutex_init(&sa->sa_lock, NULL, MUTEX_NOLOCKDEP, NULL);
1020 	sa->sa_master_obj = sa_obj;
1021 
1022 	os->os_sa = sa;
1023 	mutex_enter(&sa->sa_lock);
1024 	mutex_exit(&os->os_user_ptr_lock);
1025 	avl_create(&sa->sa_layout_num_tree, layout_num_compare,
1026 	    sizeof (sa_lot_t), offsetof(sa_lot_t, lot_num_node));
1027 	avl_create(&sa->sa_layout_hash_tree, layout_hash_compare,
1028 	    sizeof (sa_lot_t), offsetof(sa_lot_t, lot_hash_node));
1029 
1030 	if (sa_obj) {
1031 		error = zap_lookup(os, sa_obj, SA_LAYOUTS,
1032 		    8, 1, &sa->sa_layout_attr_obj);
1033 		if (error != 0 && error != ENOENT)
1034 			goto fail;
1035 		error = zap_lookup(os, sa_obj, SA_REGISTRY,
1036 		    8, 1, &sa->sa_reg_attr_obj);
1037 		if (error != 0 && error != ENOENT)
1038 			goto fail;
1039 	}
1040 
1041 	if ((error = sa_attr_table_setup(os, reg_attrs, count)) != 0)
1042 		goto fail;
1043 
1044 	if (sa->sa_layout_attr_obj != 0) {
1045 		uint64_t layout_count;
1046 
1047 		error = zap_count(os, sa->sa_layout_attr_obj,
1048 		    &layout_count);
1049 
1050 		/*
1051 		 * Layout number count should be > 0
1052 		 */
1053 		if (error || (error == 0 && layout_count == 0)) {
1054 			if (error == 0)
1055 				error = SET_ERROR(EINVAL);
1056 			goto fail;
1057 		}
1058 
1059 		za = zap_attribute_alloc();
1060 		for (zap_cursor_init(&zc, os, sa->sa_layout_attr_obj);
1061 		    (error = zap_cursor_retrieve(&zc, za)) == 0;
1062 		    zap_cursor_advance(&zc)) {
1063 			sa_attr_type_t *lot_attrs;
1064 			uint64_t lot_num;
1065 
1066 			lot_attrs = kmem_zalloc(sizeof (sa_attr_type_t) *
1067 			    za->za_num_integers, KM_SLEEP);
1068 
1069 			if ((error = (zap_lookup(os, sa->sa_layout_attr_obj,
1070 			    za->za_name, 2, za->za_num_integers,
1071 			    lot_attrs))) != 0) {
1072 				kmem_free(lot_attrs, sizeof (sa_attr_type_t) *
1073 				    za->za_num_integers);
1074 				break;
1075 			}
1076 			VERIFY0(ddi_strtoull(za->za_name, NULL, 10,
1077 			    (unsigned long long *)&lot_num));
1078 
1079 			(void) sa_add_layout_entry(os, lot_attrs,
1080 			    za->za_num_integers, lot_num,
1081 			    sa_layout_info_hash(lot_attrs,
1082 			    za->za_num_integers), B_FALSE, NULL);
1083 			kmem_free(lot_attrs, sizeof (sa_attr_type_t) *
1084 			    za->za_num_integers);
1085 		}
1086 		zap_cursor_fini(&zc);
1087 		zap_attribute_free(za);
1088 
1089 		/*
1090 		 * Make sure layout count matches number of entries added
1091 		 * to AVL tree
1092 		 */
1093 		if (avl_numnodes(&sa->sa_layout_num_tree) != layout_count) {
1094 			ASSERT(error != 0);
1095 			goto fail;
1096 		}
1097 	}
1098 
1099 	/* Add special layout number for old ZNODES */
1100 	if (ostype == DMU_OST_ZFS) {
1101 		(void) sa_add_layout_entry(os, sa_legacy_zpl_layout,
1102 		    sa_legacy_attr_count, 0,
1103 		    sa_layout_info_hash(sa_legacy_zpl_layout,
1104 		    sa_legacy_attr_count), B_FALSE, NULL);
1105 
1106 		(void) sa_add_layout_entry(os, sa_dummy_zpl_layout, 0, 1,
1107 		    0, B_FALSE, NULL);
1108 	}
1109 	*user_table = os->os_sa->sa_user_table;
1110 	mutex_exit(&sa->sa_lock);
1111 	return (0);
1112 fail:
1113 	os->os_sa = NULL;
1114 	sa_free_attr_table(sa);
1115 	if (sa->sa_user_table)
1116 		kmem_free(sa->sa_user_table, sa->sa_user_table_sz);
1117 	mutex_exit(&sa->sa_lock);
1118 	avl_destroy(&sa->sa_layout_hash_tree);
1119 	avl_destroy(&sa->sa_layout_num_tree);
1120 	mutex_destroy(&sa->sa_lock);
1121 	kmem_free(sa, sizeof (sa_os_t));
1122 	return ((error == ECKSUM) ? EIO : error);
1123 }
1124 
1125 void
sa_tear_down(objset_t * os)1126 sa_tear_down(objset_t *os)
1127 {
1128 	sa_os_t *sa = os->os_sa;
1129 	sa_lot_t *layout;
1130 	void *cookie;
1131 
1132 	kmem_free(sa->sa_user_table, sa->sa_user_table_sz);
1133 
1134 	/* Free up attr table */
1135 
1136 	sa_free_attr_table(sa);
1137 
1138 	cookie = NULL;
1139 	while ((layout =
1140 	    avl_destroy_nodes(&sa->sa_layout_hash_tree, &cookie))) {
1141 		sa_idx_tab_t *tab;
1142 		while ((tab = list_head(&layout->lot_idx_tab))) {
1143 			ASSERT(zfs_refcount_count(&tab->sa_refcount));
1144 			sa_idx_tab_rele(os, tab);
1145 		}
1146 	}
1147 
1148 	cookie = NULL;
1149 	while ((layout = avl_destroy_nodes(&sa->sa_layout_num_tree, &cookie))) {
1150 		kmem_free(layout->lot_attrs,
1151 		    sizeof (sa_attr_type_t) * layout->lot_attr_count);
1152 		kmem_free(layout, sizeof (sa_lot_t));
1153 	}
1154 
1155 	avl_destroy(&sa->sa_layout_hash_tree);
1156 	avl_destroy(&sa->sa_layout_num_tree);
1157 	mutex_destroy(&sa->sa_lock);
1158 
1159 	kmem_free(sa, sizeof (sa_os_t));
1160 	os->os_sa = NULL;
1161 }
1162 
1163 static void
sa_build_idx_tab(void * hdr,void * attr_addr,sa_attr_type_t attr,uint16_t length,int length_idx,boolean_t var_length,void * userp)1164 sa_build_idx_tab(void *hdr, void *attr_addr, sa_attr_type_t attr,
1165     uint16_t length, int length_idx, boolean_t var_length, void *userp)
1166 {
1167 	sa_idx_tab_t *idx_tab = userp;
1168 
1169 	if (var_length) {
1170 		ASSERT(idx_tab->sa_variable_lengths);
1171 		idx_tab->sa_variable_lengths[length_idx] = length;
1172 	}
1173 	TOC_ATTR_ENCODE(idx_tab->sa_idx_tab[attr], length_idx,
1174 	    (uint32_t)((uintptr_t)attr_addr - (uintptr_t)hdr));
1175 }
1176 
1177 static void
sa_attr_iter(objset_t * os,sa_hdr_phys_t * hdr,dmu_object_type_t type,sa_iterfunc_t func,sa_lot_t * tab,void * userp)1178 sa_attr_iter(objset_t *os, sa_hdr_phys_t *hdr, dmu_object_type_t type,
1179     sa_iterfunc_t func, sa_lot_t *tab, void *userp)
1180 {
1181 	void *data_start;
1182 	sa_lot_t *tb = tab;
1183 	sa_lot_t search;
1184 	avl_index_t loc;
1185 	sa_os_t *sa = os->os_sa;
1186 	int i;
1187 	uint16_t *length_start = NULL;
1188 	uint8_t length_idx = 0;
1189 
1190 	if (tab == NULL) {
1191 		search.lot_num = SA_LAYOUT_NUM(hdr, type);
1192 		tb = avl_find(&sa->sa_layout_num_tree, &search, &loc);
1193 		ASSERT(tb);
1194 	}
1195 
1196 	if (IS_SA_BONUSTYPE(type)) {
1197 		data_start = (void *)P2ROUNDUP(((uintptr_t)hdr +
1198 		    offsetof(sa_hdr_phys_t, sa_lengths) +
1199 		    (sizeof (uint16_t) * tb->lot_var_sizes)), 8);
1200 		length_start = hdr->sa_lengths;
1201 	} else {
1202 		data_start = hdr;
1203 	}
1204 
1205 	for (i = 0; i != tb->lot_attr_count; i++) {
1206 		int attr_length, reg_length;
1207 		uint8_t idx_len;
1208 
1209 		reg_length = sa->sa_attr_table[tb->lot_attrs[i]].sa_length;
1210 		IMPLY(reg_length == 0, IS_SA_BONUSTYPE(type));
1211 		if (reg_length) {
1212 			attr_length = reg_length;
1213 			idx_len = 0;
1214 		} else {
1215 			attr_length = length_start[length_idx];
1216 			idx_len = length_idx++;
1217 		}
1218 
1219 		func(hdr, data_start, tb->lot_attrs[i], attr_length,
1220 		    idx_len, reg_length == 0 ? B_TRUE : B_FALSE, userp);
1221 
1222 		data_start = (void *)P2ROUNDUP(((uintptr_t)data_start +
1223 		    attr_length), 8);
1224 	}
1225 }
1226 
1227 static void
sa_byteswap_cb(void * hdr,void * attr_addr,sa_attr_type_t attr,uint16_t length,int length_idx,boolean_t variable_length,void * userp)1228 sa_byteswap_cb(void *hdr, void *attr_addr, sa_attr_type_t attr,
1229     uint16_t length, int length_idx, boolean_t variable_length, void *userp)
1230 {
1231 	(void) hdr, (void) length_idx, (void) variable_length;
1232 	sa_handle_t *hdl = userp;
1233 	sa_os_t *sa = hdl->sa_os->os_sa;
1234 
1235 	sa_bswap_table[sa->sa_attr_table[attr].sa_byteswap](attr_addr, length);
1236 }
1237 
1238 static void
sa_byteswap(sa_handle_t * hdl,sa_buf_type_t buftype)1239 sa_byteswap(sa_handle_t *hdl, sa_buf_type_t buftype)
1240 {
1241 	sa_hdr_phys_t *sa_hdr_phys = SA_GET_HDR(hdl, buftype);
1242 	dmu_buf_impl_t *db;
1243 	int num_lengths = 1;
1244 	int i;
1245 	sa_os_t *sa __maybe_unused = hdl->sa_os->os_sa;
1246 
1247 	ASSERT(MUTEX_HELD(&sa->sa_lock));
1248 	if (sa_hdr_phys->sa_magic == SA_MAGIC)
1249 		return;
1250 
1251 	db = SA_GET_DB(hdl, buftype);
1252 
1253 	/*
1254 	 * Acquire db_mtx to protect against concurrent access to the buffer
1255 	 * by dmu_objset_userquota_get_ids() while we perform byte-swapping.
1256 	 * We also need it for doing arc_release()/arc_buf_freeze().
1257 	 */
1258 	mutex_enter(&db->db_mtx);
1259 
1260 	if (buftype == SA_SPILL)
1261 		arc_release(db->db_buf, NULL);
1262 
1263 	sa_hdr_phys->sa_magic = BSWAP_32(sa_hdr_phys->sa_magic);
1264 	sa_hdr_phys->sa_layout_info = BSWAP_16(sa_hdr_phys->sa_layout_info);
1265 
1266 	/*
1267 	 * Determine number of variable lengths in header
1268 	 * The standard 8 byte header has one for free and a
1269 	 * 16 byte header would have 4 + 1;
1270 	 */
1271 	if (SA_HDR_SIZE(sa_hdr_phys) > 8)
1272 		num_lengths += (SA_HDR_SIZE(sa_hdr_phys) - 8) >> 1;
1273 	for (i = 0; i != num_lengths; i++)
1274 		sa_hdr_phys->sa_lengths[i] =
1275 		    BSWAP_16(sa_hdr_phys->sa_lengths[i]);
1276 
1277 	sa_attr_iter(hdl->sa_os, sa_hdr_phys, DMU_OT_SA,
1278 	    sa_byteswap_cb, NULL, hdl);
1279 
1280 	if (buftype == SA_SPILL)
1281 		arc_buf_freeze(db->db_buf);
1282 
1283 	mutex_exit(&db->db_mtx);
1284 }
1285 
1286 static int
sa_build_index(sa_handle_t * hdl,sa_buf_type_t buftype)1287 sa_build_index(sa_handle_t *hdl, sa_buf_type_t buftype)
1288 {
1289 	sa_hdr_phys_t *sa_hdr_phys;
1290 	dmu_buf_impl_t *db = SA_GET_DB(hdl, buftype);
1291 	dmu_object_type_t bonustype = SA_BONUSTYPE_FROM_DB(db);
1292 	sa_os_t *sa = hdl->sa_os->os_sa;
1293 	sa_idx_tab_t *idx_tab;
1294 
1295 	sa_hdr_phys = SA_GET_HDR(hdl, buftype);
1296 
1297 	mutex_enter(&sa->sa_lock);
1298 
1299 	/* Do we need to byteswap? */
1300 
1301 	/* only check if not old znode */
1302 	if (IS_SA_BONUSTYPE(bonustype) && sa_hdr_phys->sa_magic != SA_MAGIC &&
1303 	    sa_hdr_phys->sa_magic != 0) {
1304 		if (BSWAP_32(sa_hdr_phys->sa_magic) != SA_MAGIC) {
1305 			mutex_exit(&sa->sa_lock);
1306 			zfs_dbgmsg("Buffer Header: %x != SA_MAGIC:%x "
1307 			    "object=%#llx\n", sa_hdr_phys->sa_magic, SA_MAGIC,
1308 			    (u_longlong_t)db->db.db_object);
1309 			return (SET_ERROR(EIO));
1310 		}
1311 		sa_byteswap(hdl, buftype);
1312 	}
1313 
1314 	idx_tab = sa_find_idx_tab(hdl->sa_os, bonustype, sa_hdr_phys);
1315 
1316 	if (buftype == SA_BONUS)
1317 		hdl->sa_bonus_tab = idx_tab;
1318 	else
1319 		hdl->sa_spill_tab = idx_tab;
1320 
1321 	mutex_exit(&sa->sa_lock);
1322 	return (0);
1323 }
1324 
1325 static void
sa_evict_sync(void * dbu)1326 sa_evict_sync(void *dbu)
1327 {
1328 	(void) dbu;
1329 	panic("evicting sa dbuf\n");
1330 }
1331 
1332 static void
sa_idx_tab_rele(objset_t * os,void * arg)1333 sa_idx_tab_rele(objset_t *os, void *arg)
1334 {
1335 	sa_os_t *sa = os->os_sa;
1336 	sa_idx_tab_t *idx_tab = arg;
1337 
1338 	if (idx_tab == NULL)
1339 		return;
1340 
1341 	mutex_enter(&sa->sa_lock);
1342 	if (zfs_refcount_remove(&idx_tab->sa_refcount, NULL) == 0) {
1343 		list_remove(&idx_tab->sa_layout->lot_idx_tab, idx_tab);
1344 		if (idx_tab->sa_variable_lengths)
1345 			kmem_free(idx_tab->sa_variable_lengths,
1346 			    sizeof (uint16_t) *
1347 			    idx_tab->sa_layout->lot_var_sizes);
1348 		zfs_refcount_destroy(&idx_tab->sa_refcount);
1349 		kmem_free(idx_tab->sa_idx_tab,
1350 		    sizeof (uint32_t) * sa->sa_num_attrs);
1351 		kmem_free(idx_tab, sizeof (sa_idx_tab_t));
1352 	}
1353 	mutex_exit(&sa->sa_lock);
1354 }
1355 
1356 static void
sa_idx_tab_hold(objset_t * os,sa_idx_tab_t * idx_tab)1357 sa_idx_tab_hold(objset_t *os, sa_idx_tab_t *idx_tab)
1358 {
1359 	sa_os_t *sa __maybe_unused = os->os_sa;
1360 
1361 	ASSERT(MUTEX_HELD(&sa->sa_lock));
1362 	(void) zfs_refcount_add(&idx_tab->sa_refcount, NULL);
1363 }
1364 
1365 void
sa_spill_rele(sa_handle_t * hdl)1366 sa_spill_rele(sa_handle_t *hdl)
1367 {
1368 	mutex_enter(&hdl->sa_lock);
1369 	if (hdl->sa_spill) {
1370 		sa_idx_tab_rele(hdl->sa_os, hdl->sa_spill_tab);
1371 		dmu_buf_rele(hdl->sa_spill, NULL);
1372 		hdl->sa_spill = NULL;
1373 		hdl->sa_spill_tab = NULL;
1374 	}
1375 	mutex_exit(&hdl->sa_lock);
1376 }
1377 
1378 void
sa_handle_destroy(sa_handle_t * hdl)1379 sa_handle_destroy(sa_handle_t *hdl)
1380 {
1381 	dmu_buf_t *db = hdl->sa_bonus;
1382 
1383 	mutex_enter(&hdl->sa_lock);
1384 	(void) dmu_buf_remove_user(db, &hdl->sa_dbu);
1385 
1386 	if (hdl->sa_bonus_tab)
1387 		sa_idx_tab_rele(hdl->sa_os, hdl->sa_bonus_tab);
1388 
1389 	if (hdl->sa_spill_tab)
1390 		sa_idx_tab_rele(hdl->sa_os, hdl->sa_spill_tab);
1391 
1392 	dmu_buf_rele(hdl->sa_bonus, NULL);
1393 
1394 	if (hdl->sa_spill)
1395 		dmu_buf_rele(hdl->sa_spill, NULL);
1396 	mutex_exit(&hdl->sa_lock);
1397 
1398 	kmem_cache_free(sa_cache, hdl);
1399 }
1400 
1401 int
sa_handle_get_from_db(objset_t * os,dmu_buf_t * db,void * userp,sa_handle_type_t hdl_type,sa_handle_t ** handlepp)1402 sa_handle_get_from_db(objset_t *os, dmu_buf_t *db, void *userp,
1403     sa_handle_type_t hdl_type, sa_handle_t **handlepp)
1404 {
1405 	int error = 0;
1406 	sa_handle_t *handle = NULL;
1407 #ifdef ZFS_DEBUG
1408 	dmu_object_info_t doi;
1409 
1410 	dmu_object_info_from_db(db, &doi);
1411 	ASSERT(doi.doi_bonus_type == DMU_OT_SA ||
1412 	    doi.doi_bonus_type == DMU_OT_ZNODE);
1413 #endif
1414 	/* find handle, if it exists */
1415 	/* if one doesn't exist then create a new one, and initialize it */
1416 
1417 	if (hdl_type == SA_HDL_SHARED)
1418 		handle = dmu_buf_get_user(db);
1419 
1420 	if (handle == NULL) {
1421 		sa_handle_t *winner = NULL;
1422 
1423 		handle = kmem_cache_alloc(sa_cache, KM_SLEEP);
1424 		handle->sa_dbu.dbu_evict_func_sync = NULL;
1425 		handle->sa_dbu.dbu_evict_func_async = NULL;
1426 		handle->sa_userp = userp;
1427 		handle->sa_bonus = db;
1428 		handle->sa_os = os;
1429 		handle->sa_spill = NULL;
1430 		handle->sa_bonus_tab = NULL;
1431 		handle->sa_spill_tab = NULL;
1432 
1433 		error = sa_build_index(handle, SA_BONUS);
1434 
1435 		if (hdl_type == SA_HDL_SHARED) {
1436 			dmu_buf_init_user(&handle->sa_dbu, sa_evict_sync, NULL,
1437 			    NULL);
1438 			winner = dmu_buf_set_user_ie(db, &handle->sa_dbu);
1439 		}
1440 
1441 		if (winner != NULL) {
1442 			kmem_cache_free(sa_cache, handle);
1443 			handle = winner;
1444 		}
1445 	}
1446 	*handlepp = handle;
1447 
1448 	return (error);
1449 }
1450 
1451 int
sa_handle_get(objset_t * objset,uint64_t objid,void * userp,sa_handle_type_t hdl_type,sa_handle_t ** handlepp)1452 sa_handle_get(objset_t *objset, uint64_t objid, void *userp,
1453     sa_handle_type_t hdl_type, sa_handle_t **handlepp)
1454 {
1455 	dmu_buf_t *db;
1456 	int error;
1457 
1458 	if ((error = dmu_bonus_hold(objset, objid, NULL, &db)))
1459 		return (error);
1460 
1461 	return (sa_handle_get_from_db(objset, db, userp, hdl_type,
1462 	    handlepp));
1463 }
1464 
1465 int
sa_buf_hold(objset_t * objset,uint64_t obj_num,const void * tag,dmu_buf_t ** db)1466 sa_buf_hold(objset_t *objset, uint64_t obj_num, const void *tag, dmu_buf_t **db)
1467 {
1468 	return (dmu_bonus_hold(objset, obj_num, tag, db));
1469 }
1470 
1471 void
sa_buf_rele(dmu_buf_t * db,const void * tag)1472 sa_buf_rele(dmu_buf_t *db, const void *tag)
1473 {
1474 	dmu_buf_rele(db, tag);
1475 }
1476 
1477 static int
sa_lookup_impl(sa_handle_t * hdl,sa_bulk_attr_t * bulk,int count)1478 sa_lookup_impl(sa_handle_t *hdl, sa_bulk_attr_t *bulk, int count)
1479 {
1480 	ASSERT(hdl);
1481 	ASSERT(MUTEX_HELD(&hdl->sa_lock));
1482 	return (sa_attr_op(hdl, bulk, count, SA_LOOKUP, NULL));
1483 }
1484 
1485 static int
sa_lookup_locked(sa_handle_t * hdl,sa_attr_type_t attr,void * buf,uint32_t buflen)1486 sa_lookup_locked(sa_handle_t *hdl, sa_attr_type_t attr, void *buf,
1487     uint32_t buflen)
1488 {
1489 	int error;
1490 	sa_bulk_attr_t bulk;
1491 
1492 	VERIFY3U(buflen, <=, SA_ATTR_MAX_LEN);
1493 
1494 	bulk.sa_attr = attr;
1495 	bulk.sa_data = buf;
1496 	bulk.sa_length = buflen;
1497 	bulk.sa_data_func = NULL;
1498 
1499 	ASSERT(hdl);
1500 	error = sa_lookup_impl(hdl, &bulk, 1);
1501 	return (error);
1502 }
1503 
1504 int
sa_lookup(sa_handle_t * hdl,sa_attr_type_t attr,void * buf,uint32_t buflen)1505 sa_lookup(sa_handle_t *hdl, sa_attr_type_t attr, void *buf, uint32_t buflen)
1506 {
1507 	int error;
1508 
1509 	mutex_enter(&hdl->sa_lock);
1510 	error = sa_lookup_locked(hdl, attr, buf, buflen);
1511 	mutex_exit(&hdl->sa_lock);
1512 
1513 	return (error);
1514 }
1515 
1516 /*
1517  * Return size of an attribute
1518  */
1519 
1520 static int
sa_size_locked(sa_handle_t * hdl,sa_attr_type_t attr,int * size)1521 sa_size_locked(sa_handle_t *hdl, sa_attr_type_t attr, int *size)
1522 {
1523 	sa_bulk_attr_t bulk;
1524 	int error;
1525 
1526 	bulk.sa_data = NULL;
1527 	bulk.sa_attr = attr;
1528 	bulk.sa_data_func = NULL;
1529 
1530 	ASSERT(hdl);
1531 	ASSERT(MUTEX_HELD(&hdl->sa_lock));
1532 	if ((error = sa_attr_op(hdl, &bulk, 1, SA_LOOKUP, NULL)) != 0) {
1533 		return (error);
1534 	}
1535 	*size = bulk.sa_size;
1536 
1537 	return (0);
1538 }
1539 
1540 int
sa_size(sa_handle_t * hdl,sa_attr_type_t attr,int * size)1541 sa_size(sa_handle_t *hdl, sa_attr_type_t attr, int *size)
1542 {
1543 	int error;
1544 
1545 	mutex_enter(&hdl->sa_lock);
1546 	error = sa_size_locked(hdl, attr, size);
1547 	mutex_exit(&hdl->sa_lock);
1548 
1549 	return (error);
1550 }
1551 
1552 #ifdef _KERNEL
1553 int
sa_lookup_uio(sa_handle_t * hdl,sa_attr_type_t attr,zfs_uio_t * uio)1554 sa_lookup_uio(sa_handle_t *hdl, sa_attr_type_t attr, zfs_uio_t *uio)
1555 {
1556 	int error;
1557 	sa_bulk_attr_t bulk;
1558 
1559 	bulk.sa_data = NULL;
1560 	bulk.sa_attr = attr;
1561 	bulk.sa_data_func = NULL;
1562 
1563 	ASSERT(hdl);
1564 
1565 	mutex_enter(&hdl->sa_lock);
1566 	if ((error = sa_attr_op(hdl, &bulk, 1, SA_LOOKUP, NULL)) == 0) {
1567 		error = zfs_uiomove((void *)bulk.sa_addr, MIN(bulk.sa_size,
1568 		    zfs_uio_resid(uio)), UIO_READ, uio);
1569 	}
1570 	mutex_exit(&hdl->sa_lock);
1571 	return (error);
1572 }
1573 
1574 /*
1575  * For the existed object that is upgraded from old system, its ondisk layout
1576  * has no slot for the project ID attribute. But quota accounting logic needs
1577  * to access related slots by offset directly. So we need to adjust these old
1578  * objects' layout to make the project ID to some unified and fixed offset.
1579  */
1580 int
sa_add_projid(sa_handle_t * hdl,dmu_tx_t * tx,uint64_t projid)1581 sa_add_projid(sa_handle_t *hdl, dmu_tx_t *tx, uint64_t projid)
1582 {
1583 	znode_t *zp = sa_get_userdata(hdl);
1584 	dmu_buf_t *db = sa_get_db(hdl);
1585 	zfsvfs_t *zfsvfs = ZTOZSB(zp);
1586 	int count = 0, err = 0;
1587 	sa_bulk_attr_t *bulk, *attrs;
1588 	zfs_acl_locator_cb_t locate = { 0 };
1589 	uint64_t uid, gid, mode, rdev, xattr = 0, parent, gen, links;
1590 	uint64_t crtime[2], mtime[2], ctime[2], atime[2];
1591 	zfs_acl_phys_t znode_acl = { 0 };
1592 	char scanstamp[AV_SCANSTAMP_SZ];
1593 	char *dxattr_obj = NULL;
1594 	int dxattr_size = 0;
1595 
1596 	if (zp->z_acl_cached == NULL) {
1597 		zfs_acl_t *aclp;
1598 
1599 		mutex_enter(&zp->z_acl_lock);
1600 		err = zfs_acl_node_read(zp, B_FALSE, &aclp, B_FALSE);
1601 		mutex_exit(&zp->z_acl_lock);
1602 		if (err != 0 && err != ENOENT)
1603 			return (err);
1604 	}
1605 
1606 	bulk = kmem_zalloc(sizeof (sa_bulk_attr_t) * ZPL_END, KM_SLEEP);
1607 	attrs = kmem_zalloc(sizeof (sa_bulk_attr_t) * ZPL_END, KM_SLEEP);
1608 	mutex_enter(&hdl->sa_lock);
1609 	mutex_enter(&zp->z_lock);
1610 
1611 	err = sa_lookup_locked(hdl, SA_ZPL_PROJID(zfsvfs), &projid,
1612 	    sizeof (uint64_t));
1613 	if (unlikely(err == 0))
1614 		/* Someone has added project ID attr by race. */
1615 		err = EEXIST;
1616 	if (err != ENOENT)
1617 		goto out;
1618 
1619 	/* First do a bulk query of the attributes that aren't cached */
1620 	if (zp->z_is_sa) {
1621 		SA_ADD_BULK_ATTR(bulk, count, SA_ZPL_MODE(zfsvfs), NULL,
1622 		    &mode, 8);
1623 		SA_ADD_BULK_ATTR(bulk, count, SA_ZPL_GEN(zfsvfs), NULL,
1624 		    &gen, 8);
1625 		SA_ADD_BULK_ATTR(bulk, count, SA_ZPL_UID(zfsvfs), NULL,
1626 		    &uid, 8);
1627 		SA_ADD_BULK_ATTR(bulk, count, SA_ZPL_GID(zfsvfs), NULL,
1628 		    &gid, 8);
1629 		SA_ADD_BULK_ATTR(bulk, count, SA_ZPL_PARENT(zfsvfs), NULL,
1630 		    &parent, 8);
1631 		SA_ADD_BULK_ATTR(bulk, count, SA_ZPL_ATIME(zfsvfs), NULL,
1632 		    &atime, 16);
1633 		SA_ADD_BULK_ATTR(bulk, count, SA_ZPL_MTIME(zfsvfs), NULL,
1634 		    &mtime, 16);
1635 		SA_ADD_BULK_ATTR(bulk, count, SA_ZPL_CTIME(zfsvfs), NULL,
1636 		    &ctime, 16);
1637 		SA_ADD_BULK_ATTR(bulk, count, SA_ZPL_CRTIME(zfsvfs), NULL,
1638 		    &crtime, 16);
1639 		if (Z_ISBLK(ZTOTYPE(zp)) || Z_ISCHR(ZTOTYPE(zp)))
1640 			SA_ADD_BULK_ATTR(bulk, count, SA_ZPL_RDEV(zfsvfs), NULL,
1641 			    &rdev, 8);
1642 	} else {
1643 		SA_ADD_BULK_ATTR(bulk, count, SA_ZPL_ATIME(zfsvfs), NULL,
1644 		    &atime, 16);
1645 		SA_ADD_BULK_ATTR(bulk, count, SA_ZPL_MTIME(zfsvfs), NULL,
1646 		    &mtime, 16);
1647 		SA_ADD_BULK_ATTR(bulk, count, SA_ZPL_CTIME(zfsvfs), NULL,
1648 		    &ctime, 16);
1649 		SA_ADD_BULK_ATTR(bulk, count, SA_ZPL_CRTIME(zfsvfs), NULL,
1650 		    &crtime, 16);
1651 		SA_ADD_BULK_ATTR(bulk, count, SA_ZPL_GEN(zfsvfs), NULL,
1652 		    &gen, 8);
1653 		SA_ADD_BULK_ATTR(bulk, count, SA_ZPL_MODE(zfsvfs), NULL,
1654 		    &mode, 8);
1655 		SA_ADD_BULK_ATTR(bulk, count, SA_ZPL_PARENT(zfsvfs), NULL,
1656 		    &parent, 8);
1657 		SA_ADD_BULK_ATTR(bulk, count, SA_ZPL_XATTR(zfsvfs), NULL,
1658 		    &xattr, 8);
1659 		SA_ADD_BULK_ATTR(bulk, count, SA_ZPL_RDEV(zfsvfs), NULL,
1660 		    &rdev, 8);
1661 		SA_ADD_BULK_ATTR(bulk, count, SA_ZPL_UID(zfsvfs), NULL,
1662 		    &uid, 8);
1663 		SA_ADD_BULK_ATTR(bulk, count, SA_ZPL_GID(zfsvfs), NULL,
1664 		    &gid, 8);
1665 		SA_ADD_BULK_ATTR(bulk, count, SA_ZPL_ZNODE_ACL(zfsvfs), NULL,
1666 		    &znode_acl, 88);
1667 	}
1668 	err = sa_bulk_lookup_locked(hdl, bulk, count);
1669 	if (err != 0)
1670 		goto out;
1671 
1672 	err = sa_lookup_locked(hdl, SA_ZPL_XATTR(zfsvfs), &xattr, 8);
1673 	if (err != 0 && err != ENOENT)
1674 		goto out;
1675 
1676 	err = sa_size_locked(hdl, SA_ZPL_DXATTR(zfsvfs), &dxattr_size);
1677 	if (err != 0 && err != ENOENT)
1678 		goto out;
1679 	if (dxattr_size != 0) {
1680 		dxattr_obj = vmem_alloc(dxattr_size, KM_SLEEP);
1681 		err = sa_lookup_locked(hdl, SA_ZPL_DXATTR(zfsvfs), dxattr_obj,
1682 		    dxattr_size);
1683 		if (err != 0 && err != ENOENT)
1684 			goto out;
1685 	}
1686 
1687 	zp->z_projid = projid;
1688 	zp->z_pflags |= ZFS_PROJID;
1689 	links = ZTONLNK(zp);
1690 	count = 0;
1691 	err = 0;
1692 
1693 	SA_ADD_BULK_ATTR(attrs, count, SA_ZPL_MODE(zfsvfs), NULL, &mode, 8);
1694 	SA_ADD_BULK_ATTR(attrs, count, SA_ZPL_SIZE(zfsvfs), NULL,
1695 	    &zp->z_size, 8);
1696 	SA_ADD_BULK_ATTR(attrs, count, SA_ZPL_GEN(zfsvfs), NULL, &gen, 8);
1697 	SA_ADD_BULK_ATTR(attrs, count, SA_ZPL_UID(zfsvfs), NULL, &uid, 8);
1698 	SA_ADD_BULK_ATTR(attrs, count, SA_ZPL_GID(zfsvfs), NULL, &gid, 8);
1699 	SA_ADD_BULK_ATTR(attrs, count, SA_ZPL_PARENT(zfsvfs), NULL, &parent, 8);
1700 	SA_ADD_BULK_ATTR(attrs, count, SA_ZPL_FLAGS(zfsvfs), NULL,
1701 	    &zp->z_pflags, 8);
1702 	SA_ADD_BULK_ATTR(attrs, count, SA_ZPL_ATIME(zfsvfs), NULL, &atime, 16);
1703 	SA_ADD_BULK_ATTR(attrs, count, SA_ZPL_MTIME(zfsvfs), NULL, &mtime, 16);
1704 	SA_ADD_BULK_ATTR(attrs, count, SA_ZPL_CTIME(zfsvfs), NULL, &ctime, 16);
1705 	SA_ADD_BULK_ATTR(attrs, count, SA_ZPL_CRTIME(zfsvfs), NULL,
1706 	    &crtime, 16);
1707 	SA_ADD_BULK_ATTR(attrs, count, SA_ZPL_LINKS(zfsvfs), NULL, &links, 8);
1708 	SA_ADD_BULK_ATTR(attrs, count, SA_ZPL_PROJID(zfsvfs), NULL, &projid, 8);
1709 
1710 	if (Z_ISBLK(ZTOTYPE(zp)) || Z_ISCHR(ZTOTYPE(zp)))
1711 		SA_ADD_BULK_ATTR(attrs, count, SA_ZPL_RDEV(zfsvfs), NULL,
1712 		    &rdev, 8);
1713 
1714 	if (zp->z_acl_cached != NULL) {
1715 		SA_ADD_BULK_ATTR(attrs, count, SA_ZPL_DACL_COUNT(zfsvfs), NULL,
1716 		    &zp->z_acl_cached->z_acl_count, 8);
1717 		if (zp->z_acl_cached->z_version < ZFS_ACL_VERSION_FUID)
1718 			zfs_acl_xform(zp, zp->z_acl_cached, CRED());
1719 		locate.cb_aclp = zp->z_acl_cached;
1720 		SA_ADD_BULK_ATTR(attrs, count, SA_ZPL_DACL_ACES(zfsvfs),
1721 		    zfs_acl_data_locator, &locate,
1722 		    zp->z_acl_cached->z_acl_bytes);
1723 	}
1724 
1725 	if (xattr)
1726 		SA_ADD_BULK_ATTR(attrs, count, SA_ZPL_XATTR(zfsvfs), NULL,
1727 		    &xattr, 8);
1728 
1729 	if (zp->z_pflags & ZFS_BONUS_SCANSTAMP) {
1730 		memcpy(scanstamp,
1731 		    (caddr_t)db->db_data + ZFS_OLD_ZNODE_PHYS_SIZE,
1732 		    AV_SCANSTAMP_SZ);
1733 		SA_ADD_BULK_ATTR(attrs, count, SA_ZPL_SCANSTAMP(zfsvfs), NULL,
1734 		    scanstamp, AV_SCANSTAMP_SZ);
1735 		zp->z_pflags &= ~ZFS_BONUS_SCANSTAMP;
1736 	}
1737 
1738 	if (dxattr_obj) {
1739 		SA_ADD_BULK_ATTR(attrs, count, SA_ZPL_DXATTR(zfsvfs),
1740 		    NULL, dxattr_obj, dxattr_size);
1741 	}
1742 
1743 	VERIFY0(dmu_set_bonustype(db, DMU_OT_SA, tx));
1744 	VERIFY0(sa_replace_all_by_template_locked(hdl, attrs, count, tx));
1745 	if (znode_acl.z_acl_extern_obj) {
1746 		VERIFY0(dmu_object_free(zfsvfs->z_os,
1747 		    znode_acl.z_acl_extern_obj, tx));
1748 	}
1749 
1750 	zp->z_is_sa = B_TRUE;
1751 
1752 out:
1753 	mutex_exit(&zp->z_lock);
1754 	mutex_exit(&hdl->sa_lock);
1755 	kmem_free(attrs, sizeof (sa_bulk_attr_t) * ZPL_END);
1756 	kmem_free(bulk, sizeof (sa_bulk_attr_t) * ZPL_END);
1757 	if (dxattr_obj)
1758 		vmem_free(dxattr_obj, dxattr_size);
1759 	return (err);
1760 }
1761 #endif
1762 
1763 static sa_idx_tab_t *
sa_find_idx_tab(objset_t * os,dmu_object_type_t bonustype,sa_hdr_phys_t * hdr)1764 sa_find_idx_tab(objset_t *os, dmu_object_type_t bonustype, sa_hdr_phys_t *hdr)
1765 {
1766 	sa_idx_tab_t *idx_tab;
1767 	sa_os_t *sa = os->os_sa;
1768 	sa_lot_t *tb, search;
1769 	avl_index_t loc;
1770 
1771 	/*
1772 	 * Deterimine layout number.  If SA node and header == 0 then
1773 	 * force the index table to the dummy "1" empty layout.
1774 	 *
1775 	 * The layout number would only be zero for a newly created file
1776 	 * that has not added any attributes yet, or with crypto enabled which
1777 	 * doesn't write any attributes to the bonus buffer.
1778 	 */
1779 
1780 	search.lot_num = SA_LAYOUT_NUM(hdr, bonustype);
1781 
1782 	tb = avl_find(&sa->sa_layout_num_tree, &search, &loc);
1783 
1784 	/* Verify header size is consistent with layout information */
1785 	ASSERT(tb);
1786 	ASSERT((IS_SA_BONUSTYPE(bonustype) &&
1787 	    SA_HDR_SIZE_MATCH_LAYOUT(hdr, tb)) || !IS_SA_BONUSTYPE(bonustype) ||
1788 	    (IS_SA_BONUSTYPE(bonustype) && hdr->sa_layout_info == 0));
1789 
1790 	/*
1791 	 * See if any of the already existing TOC entries can be reused?
1792 	 */
1793 
1794 	for (idx_tab = list_head(&tb->lot_idx_tab); idx_tab;
1795 	    idx_tab = list_next(&tb->lot_idx_tab, idx_tab)) {
1796 		boolean_t valid_idx = B_TRUE;
1797 		int i;
1798 
1799 		if (tb->lot_var_sizes != 0 &&
1800 		    idx_tab->sa_variable_lengths != NULL) {
1801 			for (i = 0; i != tb->lot_var_sizes; i++) {
1802 				if (hdr->sa_lengths[i] !=
1803 				    idx_tab->sa_variable_lengths[i]) {
1804 					valid_idx = B_FALSE;
1805 					break;
1806 				}
1807 			}
1808 		}
1809 		if (valid_idx) {
1810 			sa_idx_tab_hold(os, idx_tab);
1811 			return (idx_tab);
1812 		}
1813 	}
1814 
1815 	/* No such luck, create a new entry */
1816 	idx_tab = kmem_zalloc(sizeof (sa_idx_tab_t), KM_SLEEP);
1817 	idx_tab->sa_idx_tab =
1818 	    kmem_zalloc(sizeof (uint32_t) * sa->sa_num_attrs, KM_SLEEP);
1819 	idx_tab->sa_layout = tb;
1820 	zfs_refcount_create(&idx_tab->sa_refcount);
1821 	if (tb->lot_var_sizes)
1822 		idx_tab->sa_variable_lengths = kmem_alloc(sizeof (uint16_t) *
1823 		    tb->lot_var_sizes, KM_SLEEP);
1824 
1825 	sa_attr_iter(os, hdr, bonustype, sa_build_idx_tab,
1826 	    tb, idx_tab);
1827 	sa_idx_tab_hold(os, idx_tab);   /* one hold for consumer */
1828 	sa_idx_tab_hold(os, idx_tab);	/* one for layout */
1829 	list_insert_tail(&tb->lot_idx_tab, idx_tab);
1830 	return (idx_tab);
1831 }
1832 
1833 void
sa_default_locator(void ** dataptr,uint32_t * len,uint32_t total_len,boolean_t start,void * userdata)1834 sa_default_locator(void **dataptr, uint32_t *len, uint32_t total_len,
1835     boolean_t start, void *userdata)
1836 {
1837 	ASSERT(start);
1838 
1839 	*dataptr = userdata;
1840 	*len = total_len;
1841 }
1842 
1843 static void
sa_attr_register_sync(sa_handle_t * hdl,dmu_tx_t * tx)1844 sa_attr_register_sync(sa_handle_t *hdl, dmu_tx_t *tx)
1845 {
1846 	uint64_t attr_value = 0;
1847 	sa_os_t *sa = hdl->sa_os->os_sa;
1848 	sa_attr_table_t *tb = sa->sa_attr_table;
1849 	int i;
1850 
1851 	mutex_enter(&sa->sa_lock);
1852 
1853 	if (!sa->sa_need_attr_registration || sa->sa_master_obj == 0) {
1854 		mutex_exit(&sa->sa_lock);
1855 		return;
1856 	}
1857 
1858 	if (sa->sa_reg_attr_obj == 0) {
1859 		sa->sa_reg_attr_obj = zap_create_link(hdl->sa_os,
1860 		    DMU_OT_SA_ATTR_REGISTRATION,
1861 		    sa->sa_master_obj, SA_REGISTRY, tx);
1862 	}
1863 	for (i = 0; i != sa->sa_num_attrs; i++) {
1864 		if (sa->sa_attr_table[i].sa_registered)
1865 			continue;
1866 		ATTR_ENCODE(attr_value, tb[i].sa_attr, tb[i].sa_length,
1867 		    tb[i].sa_byteswap);
1868 		VERIFY0(zap_update(hdl->sa_os, sa->sa_reg_attr_obj,
1869 		    tb[i].sa_name, 8, 1, &attr_value, tx));
1870 		tb[i].sa_registered = B_TRUE;
1871 	}
1872 	sa->sa_need_attr_registration = B_FALSE;
1873 	mutex_exit(&sa->sa_lock);
1874 }
1875 
1876 /*
1877  * Replace all attributes with attributes specified in template.
1878  * If dnode had a spill buffer then those attributes will be
1879  * also be replaced, possibly with just an empty spill block
1880  *
1881  * This interface is intended to only be used for bulk adding of
1882  * attributes for a new file.  It will also be used by the ZPL
1883  * when converting and old formatted znode to native SA support.
1884  */
1885 int
sa_replace_all_by_template_locked(sa_handle_t * hdl,sa_bulk_attr_t * attr_desc,int attr_count,dmu_tx_t * tx)1886 sa_replace_all_by_template_locked(sa_handle_t *hdl, sa_bulk_attr_t *attr_desc,
1887     int attr_count, dmu_tx_t *tx)
1888 {
1889 	sa_os_t *sa = hdl->sa_os->os_sa;
1890 
1891 	if (sa->sa_need_attr_registration)
1892 		sa_attr_register_sync(hdl, tx);
1893 	return (sa_build_layouts(hdl, attr_desc, attr_count, tx));
1894 }
1895 
1896 int
sa_replace_all_by_template(sa_handle_t * hdl,sa_bulk_attr_t * attr_desc,int attr_count,dmu_tx_t * tx)1897 sa_replace_all_by_template(sa_handle_t *hdl, sa_bulk_attr_t *attr_desc,
1898     int attr_count, dmu_tx_t *tx)
1899 {
1900 	int error;
1901 
1902 	mutex_enter(&hdl->sa_lock);
1903 	error = sa_replace_all_by_template_locked(hdl, attr_desc,
1904 	    attr_count, tx);
1905 	mutex_exit(&hdl->sa_lock);
1906 	return (error);
1907 }
1908 
1909 /*
1910  * Add/remove a single attribute or replace a variable-sized attribute value
1911  * with a value of a different size, and then rewrite the entire set
1912  * of attributes.
1913  * Same-length attribute value replacement (including fixed-length attributes)
1914  * is handled more efficiently by the upper layers.
1915  */
1916 static int
sa_modify_attrs(sa_handle_t * hdl,sa_attr_type_t newattr,sa_data_op_t action,sa_data_locator_t * locator,void * datastart,uint16_t buflen,dmu_tx_t * tx)1917 sa_modify_attrs(sa_handle_t *hdl, sa_attr_type_t newattr,
1918     sa_data_op_t action, sa_data_locator_t *locator, void *datastart,
1919     uint16_t buflen, dmu_tx_t *tx)
1920 {
1921 	sa_os_t *sa = hdl->sa_os->os_sa;
1922 	dmu_buf_impl_t *db = (dmu_buf_impl_t *)hdl->sa_bonus;
1923 	sa_bulk_attr_t *attr_desc;
1924 	void *old_data[2];
1925 	int bonus_attr_count = 0;
1926 	int bonus_data_size = 0;
1927 	int spill_data_size = 0;
1928 	int spill_attr_count = 0;
1929 	int error;
1930 	uint16_t length, reg_length;
1931 	int i, j, k, length_idx;
1932 	sa_hdr_phys_t *hdr;
1933 	sa_idx_tab_t *idx_tab;
1934 	int attr_count;
1935 	int count;
1936 
1937 	ASSERT(MUTEX_HELD(&hdl->sa_lock));
1938 
1939 	/* First make of copy of the old data */
1940 
1941 	DB_DNODE_ENTER(db);
1942 	if (DB_DNODE(db)->dn_bonuslen != 0) {
1943 		bonus_data_size = hdl->sa_bonus->db_size;
1944 		old_data[0] = kmem_alloc(bonus_data_size, KM_SLEEP);
1945 		memcpy(old_data[0], hdl->sa_bonus->db_data,
1946 		    hdl->sa_bonus->db_size);
1947 		bonus_attr_count = hdl->sa_bonus_tab->sa_layout->lot_attr_count;
1948 	} else {
1949 		old_data[0] = NULL;
1950 	}
1951 	DB_DNODE_EXIT(db);
1952 
1953 	/* Bring spill buffer online if it isn't currently */
1954 
1955 	if ((error = sa_get_spill(hdl)) == 0) {
1956 		spill_data_size = hdl->sa_spill->db_size;
1957 		old_data[1] = vmem_alloc(spill_data_size, KM_SLEEP);
1958 		memcpy(old_data[1], hdl->sa_spill->db_data,
1959 		    hdl->sa_spill->db_size);
1960 		spill_attr_count =
1961 		    hdl->sa_spill_tab->sa_layout->lot_attr_count;
1962 	} else if (error && error != ENOENT) {
1963 		if (old_data[0])
1964 			kmem_free(old_data[0], bonus_data_size);
1965 		return (error);
1966 	} else {
1967 		old_data[1] = NULL;
1968 	}
1969 
1970 	/* build descriptor of all attributes */
1971 
1972 	attr_count = bonus_attr_count + spill_attr_count;
1973 	if (action == SA_ADD)
1974 		attr_count++;
1975 	else if (action == SA_REMOVE)
1976 		attr_count--;
1977 
1978 	attr_desc = kmem_zalloc(sizeof (sa_bulk_attr_t) * attr_count, KM_SLEEP);
1979 
1980 	/*
1981 	 * loop through bonus and spill buffer if it exists, and
1982 	 * build up new attr_descriptor to reset the attributes
1983 	 */
1984 	k = j = 0;
1985 	count = bonus_attr_count;
1986 	hdr = SA_GET_HDR(hdl, SA_BONUS);
1987 	idx_tab = SA_IDX_TAB_GET(hdl, SA_BONUS);
1988 	for (; ; k++) {
1989 		/*
1990 		 * Iterate over each attribute in layout.  Fetch the
1991 		 * size of variable-length attributes needing rewrite
1992 		 * from sa_lengths[].
1993 		 */
1994 		for (i = 0, length_idx = 0; i != count; i++) {
1995 			sa_attr_type_t attr;
1996 
1997 			attr = idx_tab->sa_layout->lot_attrs[i];
1998 			reg_length = SA_REGISTERED_LEN(sa, attr);
1999 			if (reg_length == 0) {
2000 				length = hdr->sa_lengths[length_idx];
2001 				length_idx++;
2002 			} else {
2003 				length = reg_length;
2004 			}
2005 			if (attr == newattr) {
2006 				/*
2007 				 * There is nothing to do for SA_REMOVE,
2008 				 * so it is just skipped.
2009 				 */
2010 				if (action == SA_REMOVE)
2011 					continue;
2012 
2013 				/*
2014 				 * Duplicate attributes are not allowed, so the
2015 				 * action can not be SA_ADD here.
2016 				 */
2017 				ASSERT3S(action, ==, SA_REPLACE);
2018 
2019 				/*
2020 				 * Only a variable-sized attribute can be
2021 				 * replaced here, and its size must be changing.
2022 				 */
2023 				ASSERT0(reg_length);
2024 				ASSERT3U(length, !=, buflen);
2025 				SA_ADD_BULK_ATTR(attr_desc, j, attr,
2026 				    locator, datastart, buflen);
2027 			} else {
2028 				SA_ADD_BULK_ATTR(attr_desc, j, attr,
2029 				    NULL, (void *)
2030 				    (TOC_OFF(idx_tab->sa_idx_tab[attr]) +
2031 				    (uintptr_t)old_data[k]), length);
2032 			}
2033 		}
2034 		if (k == 0 && hdl->sa_spill) {
2035 			hdr = SA_GET_HDR(hdl, SA_SPILL);
2036 			idx_tab = SA_IDX_TAB_GET(hdl, SA_SPILL);
2037 			count = spill_attr_count;
2038 		} else {
2039 			break;
2040 		}
2041 	}
2042 	if (action == SA_ADD) {
2043 		reg_length = SA_REGISTERED_LEN(sa, newattr);
2044 		IMPLY(reg_length != 0, reg_length == buflen);
2045 		SA_ADD_BULK_ATTR(attr_desc, j, newattr, locator,
2046 		    datastart, buflen);
2047 	}
2048 	ASSERT3U(j, ==, attr_count);
2049 
2050 	error = sa_build_layouts(hdl, attr_desc, attr_count, tx);
2051 
2052 	if (old_data[0])
2053 		kmem_free(old_data[0], bonus_data_size);
2054 	if (old_data[1])
2055 		vmem_free(old_data[1], spill_data_size);
2056 	kmem_free(attr_desc, sizeof (sa_bulk_attr_t) * attr_count);
2057 
2058 	return (error);
2059 }
2060 
2061 static int
sa_bulk_update_impl(sa_handle_t * hdl,sa_bulk_attr_t * bulk,int count,dmu_tx_t * tx)2062 sa_bulk_update_impl(sa_handle_t *hdl, sa_bulk_attr_t *bulk, int count,
2063     dmu_tx_t *tx)
2064 {
2065 	int error;
2066 	sa_os_t *sa = hdl->sa_os->os_sa;
2067 	dmu_object_type_t bonustype;
2068 	dmu_buf_t *saved_spill;
2069 
2070 	ASSERT(hdl);
2071 	ASSERT(MUTEX_HELD(&hdl->sa_lock));
2072 
2073 	bonustype = SA_BONUSTYPE_FROM_DB(SA_GET_DB(hdl, SA_BONUS));
2074 	saved_spill = hdl->sa_spill;
2075 
2076 	/* sync out registration table if necessary */
2077 	if (sa->sa_need_attr_registration)
2078 		sa_attr_register_sync(hdl, tx);
2079 
2080 	error = sa_attr_op(hdl, bulk, count, SA_UPDATE, tx);
2081 	if (error == 0 && !IS_SA_BONUSTYPE(bonustype) && sa->sa_update_cb)
2082 		sa->sa_update_cb(hdl, tx);
2083 
2084 	/*
2085 	 * If saved_spill is NULL and current sa_spill is not NULL that
2086 	 * means we increased the refcount of the spill buffer through
2087 	 * sa_get_spill() or dmu_spill_hold_by_dnode().  Therefore we
2088 	 * must release the hold before calling dmu_tx_commit() to avoid
2089 	 * making a copy of this buffer in dbuf_sync_leaf() due to the
2090 	 * reference count now being greater than 1.
2091 	 */
2092 	if (!saved_spill && hdl->sa_spill) {
2093 		if (hdl->sa_spill_tab) {
2094 			sa_idx_tab_rele(hdl->sa_os, hdl->sa_spill_tab);
2095 			hdl->sa_spill_tab = NULL;
2096 		}
2097 
2098 		dmu_buf_rele(hdl->sa_spill, NULL);
2099 		hdl->sa_spill = NULL;
2100 	}
2101 
2102 	return (error);
2103 }
2104 
2105 /*
2106  * update or add new attribute
2107  */
2108 int
sa_update(sa_handle_t * hdl,sa_attr_type_t type,void * buf,uint32_t buflen,dmu_tx_t * tx)2109 sa_update(sa_handle_t *hdl, sa_attr_type_t type,
2110     void *buf, uint32_t buflen, dmu_tx_t *tx)
2111 {
2112 	int error;
2113 	sa_bulk_attr_t bulk;
2114 
2115 	VERIFY3U(buflen, <=, SA_ATTR_MAX_LEN);
2116 
2117 	bulk.sa_attr = type;
2118 	bulk.sa_data_func = NULL;
2119 	bulk.sa_length = buflen;
2120 	bulk.sa_data = buf;
2121 
2122 	mutex_enter(&hdl->sa_lock);
2123 	error = sa_bulk_update_impl(hdl, &bulk, 1, tx);
2124 	mutex_exit(&hdl->sa_lock);
2125 	return (error);
2126 }
2127 
2128 int
sa_bulk_lookup_locked(sa_handle_t * hdl,sa_bulk_attr_t * attrs,int count)2129 sa_bulk_lookup_locked(sa_handle_t *hdl, sa_bulk_attr_t *attrs, int count)
2130 {
2131 	ASSERT(hdl);
2132 	ASSERT(MUTEX_HELD(&hdl->sa_lock));
2133 	return (sa_lookup_impl(hdl, attrs, count));
2134 }
2135 
2136 int
sa_bulk_lookup(sa_handle_t * hdl,sa_bulk_attr_t * attrs,int count)2137 sa_bulk_lookup(sa_handle_t *hdl, sa_bulk_attr_t *attrs, int count)
2138 {
2139 	int error;
2140 
2141 	ASSERT(hdl);
2142 	mutex_enter(&hdl->sa_lock);
2143 	error = sa_bulk_lookup_locked(hdl, attrs, count);
2144 	mutex_exit(&hdl->sa_lock);
2145 	return (error);
2146 }
2147 
2148 int
sa_bulk_update(sa_handle_t * hdl,sa_bulk_attr_t * attrs,int count,dmu_tx_t * tx)2149 sa_bulk_update(sa_handle_t *hdl, sa_bulk_attr_t *attrs, int count, dmu_tx_t *tx)
2150 {
2151 	int error;
2152 
2153 	ASSERT(hdl);
2154 	mutex_enter(&hdl->sa_lock);
2155 	error = sa_bulk_update_impl(hdl, attrs, count, tx);
2156 	mutex_exit(&hdl->sa_lock);
2157 	return (error);
2158 }
2159 
2160 int
sa_remove(sa_handle_t * hdl,sa_attr_type_t attr,dmu_tx_t * tx)2161 sa_remove(sa_handle_t *hdl, sa_attr_type_t attr, dmu_tx_t *tx)
2162 {
2163 	int error;
2164 
2165 	mutex_enter(&hdl->sa_lock);
2166 	error = sa_modify_attrs(hdl, attr, SA_REMOVE, NULL,
2167 	    NULL, 0, tx);
2168 	mutex_exit(&hdl->sa_lock);
2169 	return (error);
2170 }
2171 
2172 void
sa_object_info(sa_handle_t * hdl,dmu_object_info_t * doi)2173 sa_object_info(sa_handle_t *hdl, dmu_object_info_t *doi)
2174 {
2175 	dmu_object_info_from_db(hdl->sa_bonus, doi);
2176 }
2177 
2178 void
sa_object_size(sa_handle_t * hdl,uint32_t * blksize,u_longlong_t * nblocks)2179 sa_object_size(sa_handle_t *hdl, uint32_t *blksize, u_longlong_t *nblocks)
2180 {
2181 	dmu_object_size_from_db(hdl->sa_bonus,
2182 	    blksize, nblocks);
2183 }
2184 
2185 void
sa_set_userp(sa_handle_t * hdl,void * ptr)2186 sa_set_userp(sa_handle_t *hdl, void *ptr)
2187 {
2188 	hdl->sa_userp = ptr;
2189 }
2190 
2191 dmu_buf_t *
sa_get_db(sa_handle_t * hdl)2192 sa_get_db(sa_handle_t *hdl)
2193 {
2194 	return (hdl->sa_bonus);
2195 }
2196 
2197 void *
sa_get_userdata(sa_handle_t * hdl)2198 sa_get_userdata(sa_handle_t *hdl)
2199 {
2200 	return (hdl->sa_userp);
2201 }
2202 
2203 void
sa_register_update_callback_locked(objset_t * os,sa_update_cb_t * func)2204 sa_register_update_callback_locked(objset_t *os, sa_update_cb_t *func)
2205 {
2206 	ASSERT(MUTEX_HELD(&os->os_sa->sa_lock));
2207 	os->os_sa->sa_update_cb = func;
2208 }
2209 
2210 void
sa_register_update_callback(objset_t * os,sa_update_cb_t * func)2211 sa_register_update_callback(objset_t *os, sa_update_cb_t *func)
2212 {
2213 
2214 	mutex_enter(&os->os_sa->sa_lock);
2215 	sa_register_update_callback_locked(os, func);
2216 	mutex_exit(&os->os_sa->sa_lock);
2217 }
2218 
2219 uint64_t
sa_handle_object(sa_handle_t * hdl)2220 sa_handle_object(sa_handle_t *hdl)
2221 {
2222 	return (hdl->sa_bonus->db_object);
2223 }
2224 
2225 boolean_t
sa_enabled(objset_t * os)2226 sa_enabled(objset_t *os)
2227 {
2228 	return (os->os_sa == NULL);
2229 }
2230 
2231 int
sa_set_sa_object(objset_t * os,uint64_t sa_object)2232 sa_set_sa_object(objset_t *os, uint64_t sa_object)
2233 {
2234 	sa_os_t *sa = os->os_sa;
2235 
2236 	if (sa->sa_master_obj)
2237 		return (1);
2238 
2239 	sa->sa_master_obj = sa_object;
2240 
2241 	return (0);
2242 }
2243 
2244 int
sa_hdrsize(void * arg)2245 sa_hdrsize(void *arg)
2246 {
2247 	sa_hdr_phys_t *hdr = arg;
2248 
2249 	return (SA_HDR_SIZE(hdr));
2250 }
2251 
2252 void
sa_handle_lock(sa_handle_t * hdl)2253 sa_handle_lock(sa_handle_t *hdl)
2254 {
2255 	ASSERT(hdl);
2256 	mutex_enter(&hdl->sa_lock);
2257 }
2258 
2259 void
sa_handle_unlock(sa_handle_t * hdl)2260 sa_handle_unlock(sa_handle_t *hdl)
2261 {
2262 	ASSERT(hdl);
2263 	mutex_exit(&hdl->sa_lock);
2264 }
2265 
2266 #ifdef _KERNEL
2267 EXPORT_SYMBOL(sa_handle_get);
2268 EXPORT_SYMBOL(sa_handle_get_from_db);
2269 EXPORT_SYMBOL(sa_handle_destroy);
2270 EXPORT_SYMBOL(sa_buf_hold);
2271 EXPORT_SYMBOL(sa_buf_rele);
2272 EXPORT_SYMBOL(sa_spill_rele);
2273 EXPORT_SYMBOL(sa_lookup);
2274 EXPORT_SYMBOL(sa_update);
2275 EXPORT_SYMBOL(sa_remove);
2276 EXPORT_SYMBOL(sa_bulk_lookup);
2277 EXPORT_SYMBOL(sa_bulk_lookup_locked);
2278 EXPORT_SYMBOL(sa_bulk_update);
2279 EXPORT_SYMBOL(sa_size);
2280 EXPORT_SYMBOL(sa_object_info);
2281 EXPORT_SYMBOL(sa_object_size);
2282 EXPORT_SYMBOL(sa_get_userdata);
2283 EXPORT_SYMBOL(sa_set_userp);
2284 EXPORT_SYMBOL(sa_get_db);
2285 EXPORT_SYMBOL(sa_handle_object);
2286 EXPORT_SYMBOL(sa_register_update_callback);
2287 EXPORT_SYMBOL(sa_setup);
2288 EXPORT_SYMBOL(sa_replace_all_by_template);
2289 EXPORT_SYMBOL(sa_replace_all_by_template_locked);
2290 EXPORT_SYMBOL(sa_enabled);
2291 EXPORT_SYMBOL(sa_cache_init);
2292 EXPORT_SYMBOL(sa_cache_fini);
2293 EXPORT_SYMBOL(sa_set_sa_object);
2294 EXPORT_SYMBOL(sa_hdrsize);
2295 EXPORT_SYMBOL(sa_handle_lock);
2296 EXPORT_SYMBOL(sa_handle_unlock);
2297 EXPORT_SYMBOL(sa_lookup_uio);
2298 EXPORT_SYMBOL(sa_add_projid);
2299 #endif /* _KERNEL */
2300