From mboxrd@z Thu Jan 1 00:00:00 1970 From: hsiangkao@aol.com (Gao Xiang) Date: Mon, 12 Aug 2019 01:10:49 +0800 Subject: [PATCH v2] erofs-utils: introduce preliminary xattr support In-Reply-To: <20190805172850.GA13290@hsiangkao-HP-ZHAN-66-Pro-G1> References: <20190805172850.GA13290@hsiangkao-HP-ZHAN-66-Pro-G1> Message-ID: <20190811171049.26111-1-hsiangkao@aol.com> From: "htyuxe+dhbrei4sq0df8@grr.la" Load xattrs from source files and pack them into target image. Signed-off-by: htyuxe+dhbrei4sq0df8 at grr.la Signed-off-by: Li Guifu Signed-off-by: Gao Xiang --- This is a modified version of xattr implementation from anonymous person. Guifu, please also take some time to test the functionality of this patch, therefore I can merge this to dev branch ASAP. configure.ac | 1 + include/erofs/defs.h | 6 + include/erofs/hashtable.h | 451 ++++++++++++++++++++++++++++++++++++++ include/erofs/internal.h | 2 +- include/erofs/xattr.h | 44 ++++ lib/Makefile.am | 3 +- lib/inode.c | 23 ++ lib/xattr.c | 289 ++++++++++++++++++++++++ 8 files changed, 817 insertions(+), 2 deletions(-) create mode 100644 include/erofs/hashtable.h create mode 100644 include/erofs/xattr.h create mode 100644 lib/xattr.c diff --git a/configure.ac b/configure.ac index fcdf30a..9174711 100644 --- a/configure.ac +++ b/configure.ac @@ -75,6 +75,7 @@ AC_CHECK_HEADERS(m4_flatten([ linux/falloc.h linux/fs.h linux/types.h + linux/xattr.h limits.h stddef.h stdint.h diff --git a/include/erofs/defs.h b/include/erofs/defs.h index 0d9910c..b325d01 100644 --- a/include/erofs/defs.h +++ b/include/erofs/defs.h @@ -147,6 +147,12 @@ typedef int64_t s64; #define BITS_PER_BYTE 8 #define BITS_TO_LONGS(nr) DIV_ROUND_UP(nr, BITS_PER_BYTE * sizeof(long)) +#ifdef __SIZEOF_LONG__ +#define BITS_PER_LONG (__CHAR_BIT__ * __SIZEOF_LONG__) +#else +#define BITS_PER_LONG __WORDSIZE +#endif + #define BUG_ON(cond) assert(!(cond)) #ifdef NDEBUG diff --git a/include/erofs/hashtable.h b/include/erofs/hashtable.h new file mode 100644 index 0000000..231ce99 --- /dev/null +++ b/include/erofs/hashtable.h @@ -0,0 +1,451 @@ +/* SPDX-License-Identifier: GPL-2.0+ */ +/* + * erofs_utils/include/erofs/hashtable.h + * (code originally taken from include/linux/hash{,table}.h) + */ +#ifndef __EROFS_HASHTABLE_H +#define __EROFS_HASHTABLE_H + +#include "defs.h" + +/* + * The "GOLDEN_RATIO_PRIME" is used in ifs/btrfs/brtfs_inode.h and + * fs/inode.c. It's not actually prime any more (the previous primes + * were actively bad for hashing), but the name remains. + */ +#if BITS_PER_LONG == 32 +#define GOLDEN_RATIO_PRIME GOLDEN_RATIO_32 +#define hash_long(val, bits) hash_32(val, bits) +#elif BITS_PER_LONG == 64 +#define hash_long(val, bits) hash_64(val, bits) +#define GOLDEN_RATIO_PRIME GOLDEN_RATIO_64 +#else +#error Wordsize not 32 or 64 +#endif + +/* + * This hash multiplies the input by a large odd number and takes the + * high bits. Since multiplication propagates changes to the most + * significant end only, it is essential that the high bits of the + * product be used for the hash value. + * + * Chuck Lever verified the effectiveness of this technique: + * http://www.citi.umich.edu/techreports/reports/citi-tr-00-1.pdf + * + * Although a random odd number will do, it turns out that the golden + * ratio phi = (sqrt(5)-1)/2, or its negative, has particularly nice + * properties. (See Knuth vol 3, section 6.4, exercise 9.) + * + * These are the negative, (1 - phi) = phi**2 = (3 - sqrt(5))/2, + * which is very slightly easier to multiply by and makes no + * difference to the hash distribution. + */ +#define GOLDEN_RATIO_32 0x61C88647 +#define GOLDEN_RATIO_64 0x61C8864680B583EBull + +struct hlist_head { + struct hlist_node *first; +}; + +struct hlist_node { + struct hlist_node *next, **pprev; +}; + +/* + * Architectures might want to move the poison pointer offset + * into some well-recognized area such as 0xdead000000000000, + * that is also not mappable by user-space exploits: + */ +#ifdef CONFIG_ILLEGAL_POINTER_VALUE +# define POISON_POINTER_DELTA _AC(CONFIG_ILLEGAL_POINTER_VALUE, UL) +#else +# define POISON_POINTER_DELTA 0 +#endif + +/* + * These are non-NULL pointers that will result in page faults + * under normal circumstances, used to verify that nobody uses + * non-initialized list entries. + */ +#define LIST_POISON1 ((void *) 0x00100100 + POISON_POINTER_DELTA) +#define LIST_POISON2 ((void *) 0x00200200 + POISON_POINTER_DELTA) + +/* + * Double linked lists with a single pointer list head. + * Mostly useful for hash tables where the two pointer list head is + * too wasteful. + * You lose the ability to access the tail in O(1). + */ + +#define HLIST_HEAD_INIT { .first = NULL } +#define HLIST_HEAD(name) struct hlist_head name = { .first = NULL } +#define INIT_HLIST_HEAD(ptr) ((ptr)->first = NULL) +static inline void INIT_HLIST_NODE(struct hlist_node *h) +{ + h->next = NULL; + h->pprev = NULL; +} + +static inline int hlist_unhashed(const struct hlist_node *h) +{ + return !h->pprev; +} + +static inline int hlist_empty(const struct hlist_head *h) +{ + return !h->first; +} + +static inline void __hlist_del(struct hlist_node *n) +{ + struct hlist_node *next = n->next; + struct hlist_node **pprev = n->pprev; + + *pprev = next; + if (next) + next->pprev = pprev; +} + +static inline void hlist_del(struct hlist_node *n) +{ + __hlist_del(n); + n->next = LIST_POISON1; + n->pprev = LIST_POISON2; +} + +static inline void hlist_del_init(struct hlist_node *n) +{ + if (!hlist_unhashed(n)) { + __hlist_del(n); + INIT_HLIST_NODE(n); + } +} + +static inline void hlist_add_head(struct hlist_node *n, struct hlist_head *h) +{ + struct hlist_node *first = h->first; + + n->next = first; + if (first) + first->pprev = &n->next; + h->first = n; + n->pprev = &h->first; +} + +/* next must be != NULL */ +static inline void hlist_add_before(struct hlist_node *n, + struct hlist_node *next) +{ + n->pprev = next->pprev; + n->next = next; + next->pprev = &n->next; + *(n->pprev) = n; +} + +static inline void hlist_add_behind(struct hlist_node *n, + struct hlist_node *prev) +{ + n->next = prev->next; + prev->next = n; + n->pprev = &prev->next; + + if (n->next) + n->next->pprev = &n->next; +} + +/* after that we'll appear to be on some hlist and hlist_del will work */ +static inline void hlist_add_fake(struct hlist_node *n) +{ + n->pprev = &n->next; +} + +/* + * Move a list from one list head to another. Fixup the pprev + * reference of the first entry if it exists. + */ +static inline void hlist_move_list(struct hlist_head *old, + struct hlist_head *new) +{ + new->first = old->first; + if (new->first) + new->first->pprev = &new->first; + old->first = NULL; +} + +#define hlist_entry(ptr, type, member) container_of(ptr, type, member) + +#define hlist_for_each(pos, head) \ + for (pos = (head)->first; pos; pos = pos->next) + +#define hlist_for_each_safe(pos, n, head) \ + for (pos = (head)->first; pos && ({ n = pos->next; 1; }); \ + pos = n) + +#define hlist_entry_safe(ptr, type, member) \ + ({ typeof(ptr) ____ptr = (ptr); \ + ____ptr ? hlist_entry(____ptr, type, member) : NULL; \ + }) + +/** + * hlist_for_each_entry - iterate over list of given type + * @pos:the type * to use as a loop cursor. + * @head:the head for your list. + * @member:the name of the hlist_node within the struct. + */ +#define hlist_for_each_entry(pos, head, member) \ + for (pos = hlist_entry_safe((head)->first, typeof(*(pos)), member);\ + pos; \ + pos = hlist_entry_safe((pos)->member.next, typeof(*(pos)), member)) + +/** + * hlist_for_each_entry_continue + * iterate over a hlist continuing after current point + * @pos:the type * to use as a loop cursor. + * @member:the name of the hlist_node within the struct. + */ +#define hlist_for_each_entry_continue(pos, member) \ + for (pos = hlist_entry_safe((pos)->member.next, typeof(*(pos)), member);\ + pos; \ + pos = hlist_entry_safe((pos)->member.next, typeof(*(pos)), member)) + +/** + * hlist_for_each_entry_from + * iterate over a hlist continuing from current point + * @pos: the type * to use as a loop cursor. + * @member: the name of the hlist_node within the struct. + */ +#define hlist_for_each_entry_from(pos, member) \ + for (; pos; \ + pos = hlist_entry_safe((pos)->member.next, typeof(*(pos)), member)) + +/** + * hlist_for_each_entry_safe + * iterate over list of given type safe against removal of list entry + * @pos:the type * to use as a loop cursor. + * @n:another &struct hlist_node to use as temporary storage + * @head:the head for your list. + * @member:the name of the hlist_node within the struct. + */ +#define hlist_for_each_entry_safe(pos, n, head, member) \ + for (pos = hlist_entry_safe((head)->first, typeof(*pos), member);\ + pos && ({ n = pos->member.next; 1; }); \ + pos = hlist_entry_safe(n, typeof(*pos), member)) + +static inline u32 __hash_32(u32 val) +{ + return val * GOLDEN_RATIO_32; +} + +static inline u32 hash_32(u32 val, unsigned int bits) +{ + /* High bits are more random, so use them. */ + return __hash_32(val) >> (32 - bits); +} + +static __always_inline u32 hash_64(u64 val, unsigned int bits) +{ +#if BITS_PER_LONG == 64 + /* 64x64-bit multiply is efficient on all 64-bit processors */ + return val * GOLDEN_RATIO_64 >> (64 - bits); +#else + /* Hash 64 bits using only 32x32-bit multiply. */ + return hash_32((u32)val ^ __hash_32(val >> 32), bits); +#endif +} + +/** + * ilog2 - log of base 2 of 32-bit or a 64-bit unsigned value + * @n - parameter + * + * constant-capable log of base 2 calculation + * - this can be used to initialise global variables from constant data, hence + * the massive ternary operator construction + * + * selects the appropriately-sized optimised version depending on sizeof(n) + */ +#define ilog2(n) \ +( \ + (n) & (1ULL << 63) ? 63 : \ + (n) & (1ULL << 62) ? 62 : \ + (n) & (1ULL << 61) ? 61 : \ + (n) & (1ULL << 60) ? 60 : \ + (n) & (1ULL << 59) ? 59 : \ + (n) & (1ULL << 58) ? 58 : \ + (n) & (1ULL << 57) ? 57 : \ + (n) & (1ULL << 56) ? 56 : \ + (n) & (1ULL << 55) ? 55 : \ + (n) & (1ULL << 54) ? 54 : \ + (n) & (1ULL << 53) ? 53 : \ + (n) & (1ULL << 52) ? 52 : \ + (n) & (1ULL << 51) ? 51 : \ + (n) & (1ULL << 50) ? 50 : \ + (n) & (1ULL << 49) ? 49 : \ + (n) & (1ULL << 48) ? 48 : \ + (n) & (1ULL << 47) ? 47 : \ + (n) & (1ULL << 46) ? 46 : \ + (n) & (1ULL << 45) ? 45 : \ + (n) & (1ULL << 44) ? 44 : \ + (n) & (1ULL << 43) ? 43 : \ + (n) & (1ULL << 42) ? 42 : \ + (n) & (1ULL << 41) ? 41 : \ + (n) & (1ULL << 40) ? 40 : \ + (n) & (1ULL << 39) ? 39 : \ + (n) & (1ULL << 38) ? 38 : \ + (n) & (1ULL << 37) ? 37 : \ + (n) & (1ULL << 36) ? 36 : \ + (n) & (1ULL << 35) ? 35 : \ + (n) & (1ULL << 34) ? 34 : \ + (n) & (1ULL << 33) ? 33 : \ + (n) & (1ULL << 32) ? 32 : \ + (n) & (1ULL << 31) ? 31 : \ + (n) & (1ULL << 30) ? 30 : \ + (n) & (1ULL << 29) ? 29 : \ + (n) & (1ULL << 28) ? 28 : \ + (n) & (1ULL << 27) ? 27 : \ + (n) & (1ULL << 26) ? 26 : \ + (n) & (1ULL << 25) ? 25 : \ + (n) & (1ULL << 24) ? 24 : \ + (n) & (1ULL << 23) ? 23 : \ + (n) & (1ULL << 22) ? 22 : \ + (n) & (1ULL << 21) ? 21 : \ + (n) & (1ULL << 20) ? 20 : \ + (n) & (1ULL << 19) ? 19 : \ + (n) & (1ULL << 18) ? 18 : \ + (n) & (1ULL << 17) ? 17 : \ + (n) & (1ULL << 16) ? 16 : \ + (n) & (1ULL << 15) ? 15 : \ + (n) & (1ULL << 14) ? 14 : \ + (n) & (1ULL << 13) ? 13 : \ + (n) & (1ULL << 12) ? 12 : \ + (n) & (1ULL << 11) ? 11 : \ + (n) & (1ULL << 10) ? 10 : \ + (n) & (1ULL << 9) ? 9 : \ + (n) & (1ULL << 8) ? 8 : \ + (n) & (1ULL << 7) ? 7 : \ + (n) & (1ULL << 6) ? 6 : \ + (n) & (1ULL << 5) ? 5 : \ + (n) & (1ULL << 4) ? 4 : \ + (n) & (1ULL << 3) ? 3 : \ + (n) & (1ULL << 2) ? 2 : \ + (n) & (1ULL << 1) ? 1 : 0 \ +) + +#define DEFINE_HASHTABLE(name, bits) \ + struct hlist_head name[1 << (bits)] = \ + { [0 ... ((1 << (bits)) - 1)] = HLIST_HEAD_INIT } + +#define DECLARE_HASHTABLE(name, bits) \ + struct hlist_head name[1 << (bits)] + +#define HASH_SIZE(name) (ARRAY_SIZE(name)) +#define HASH_BITS(name) ilog2(HASH_SIZE(name)) + +/* Use hash_32 when possible to allow for fast 32bit hashing in 64bit kernels*/ +#define hash_min(val, bits) \ + (sizeof(val) <= 4 ? hash_32(val, bits) : hash_long(val, bits)) + +static inline void __hash_init(struct hlist_head *ht, unsigned int sz) +{ + unsigned int i; + + for (i = 0; i < sz; i++) + INIT_HLIST_HEAD(&ht[i]); +} + +/** + * hash_init - initialize a hash table + * @hashtable: hashtable to be initialized + * + * Calculates the size of the hashtable from the given parameter, otherwise + * same as hash_init_size. + * + * This has to be a macro since HASH_BITS() will not work on pointers since + * it calculates the size during preprocessing. + */ +#define hash_init(hashtable) __hash_init(hashtable, HASH_SIZE(hashtable)) + +/** + * hash_add - add an object to a hashtable + * @hashtable: hashtable to add to + * @node: the &struct hlist_node of the object to be added + * @key: the key of the object to be added + */ +#define hash_add(hashtable, node, key) \ + hlist_add_head(node, &hashtable[hash_min(key, HASH_BITS(hashtable))]) + +/** + * hash_hashed - check whether an object is in any hashtable + * @node: the &struct hlist_node of the object to be checked + */ +static inline bool hash_hashed(struct hlist_node *node) +{ + return !hlist_unhashed(node); +} + +static inline bool __hash_empty(struct hlist_head *ht, unsigned int sz) +{ + unsigned int i; + + for (i = 0; i < sz; i++) + if (!hlist_empty(&ht[i])) + return false; + + return true; +} + +/** + * hash_empty - check whether a hashtable is empty + * @hashtable: hashtable to check + * + * This has to be a macro since HASH_BITS() will not work on pointers since + * it calculates the size during preprocessing. + */ +#define hash_empty(hashtable) __hash_empty(hashtable, HASH_SIZE(hashtable)) + +/** + * hash_del - remove an object from a hashtable + * @node: &struct hlist_node of the object to remove + */ +static inline void hash_del(struct hlist_node *node) +{ + hlist_del_init(node); +} + +/** + * hash_for_each - iterate over a hashtable + * @name: hashtable to iterate + * @bkt: integer to use as bucket loop cursor + * @obj: the type * to use as a loop cursor for each entry + * @member: the name of the hlist_node within the struct + */ +#define hash_for_each(name, bkt, obj, member) \ + for ((bkt) = 0, obj = NULL; obj == NULL && (bkt) < HASH_SIZE(name);\ + (bkt)++)\ + hlist_for_each_entry(obj, &name[bkt], member) + +/** + * hash_for_each_safe - iterate over a hashtable safe against removal of + * hash entry + * @name: hashtable to iterate + * @bkt: integer to use as bucket loop cursor + * @tmp: a &struct used for temporary storage + * @obj: the type * to use as a loop cursor for each entry + * @member: the name of the hlist_node within the struct + */ +#define hash_for_each_safe(name, bkt, tmp, obj, member) \ + for ((bkt) = 0, obj = NULL; obj == NULL && (bkt) < HASH_SIZE(name);\ + (bkt)++)\ + hlist_for_each_entry_safe(obj, tmp, &name[bkt], member) + +/** + * hash_for_each_possible - iterate over all possible objects hashing to the + * same bucket + * @name: hashtable to iterate + * @obj: the type * to use as a loop cursor for each entry + * @member: the name of the hlist_node within the struct + * @key: the key of the objects to iterate over + */ +#define hash_for_each_possible(name, obj, member, key) \ + hlist_for_each_entry(obj, &name[hash_min(key, HASH_BITS(name))], member) + +#endif diff --git a/include/erofs/internal.h b/include/erofs/internal.h index b7ce6f8..33a72b5 100644 --- a/include/erofs/internal.h +++ b/include/erofs/internal.h @@ -59,7 +59,7 @@ struct erofs_sb_info { extern struct erofs_sb_info sbi; struct erofs_inode { - struct list_head i_hash, i_subdirs; + struct list_head i_hash, i_subdirs, i_xattrs; unsigned int i_count; struct erofs_inode *i_parent; diff --git a/include/erofs/xattr.h b/include/erofs/xattr.h new file mode 100644 index 0000000..5fd6a59 --- /dev/null +++ b/include/erofs/xattr.h @@ -0,0 +1,44 @@ +// SPDX-License-Identifier: GPL-2.0+ +/* + * erofs_utils/include/erofs/xattr.h + */ +#ifndef __EROFS_XATTR_H +#define __EROFS_XATTR_H + +#include "internal.h" + +#define EROFS_INODE_XATTR_ICOUNT(_size) ({\ + u32 __size = le16_to_cpu(_size); \ + ((__size) == 0) ? 0 : \ + (_size - sizeof(struct erofs_xattr_ibody_header)) / \ + sizeof(struct erofs_xattr_entry) + 1; }) + +#ifndef XATTR_USER_PREFIX +#define XATTR_USER_PREFIX "user." +#endif +#ifndef XATTR_USER_PREFIX_LEN +#define XATTR_USER_PREFIX_LEN (sizeof(XATTR_USER_PREFIX) - 1) +#endif +#ifndef XATTR_SECURITY_PREFIX +#define XATTR_SECURITY_PREFIX "security." +#endif +#ifndef XATTR_SECURITY_PREFIX_LEN +#define XATTR_SECURITY_PREFIX_LEN (sizeof(XATTR_SECURITY_PREFIX) - 1) +#endif +#ifndef XATTR_TRUSTED_PREFIX +#define XATTR_TRUSTED_PREFIX "trusted." +#endif +#ifndef XATTR_TRUSTED_PREFIX_LEN +#define XATTR_TRUSTED_PREFIX_LEN (sizeof(XATTR_TRUSTED_PREFIX) - 1) +#endif +#ifndef XATTR_NAME_POSIX_ACL_ACCESS +#define XATTR_NAME_POSIX_ACL_ACCESS "system.posix_acl_access" +#endif +#ifndef XATTR_NAME_POSIX_ACL_DEFAULT +#define XATTR_NAME_POSIX_ACL_DEFAULT "system.posix_acl_default" +#endif + +int erofs_prepare_xattr_ibody(const char *path, struct list_head *ixattrs); +char *erofs_export_xattr_ibody(struct list_head *ixattrs, unsigned int size); + +#endif diff --git a/lib/Makefile.am b/lib/Makefile.am index dea82f7..1ff81f9 100644 --- a/lib/Makefile.am +++ b/lib/Makefile.am @@ -2,7 +2,8 @@ # Makefile.am noinst_LTLIBRARIES = liberofs.la -liberofs_la_SOURCES = config.c io.c cache.c inode.c compress.c compressor.c +liberofs_la_SOURCES = config.c io.c cache.c inode.c xattr.c \ + compress.c compressor.c liberofs_la_CFLAGS = -Wall -Werror -I$(top_srcdir)/include if ENABLE_LZ4 liberofs_la_CFLAGS += ${LZ4_CFLAGS} diff --git a/lib/inode.c b/lib/inode.c index 581f263..6521a28 100644 --- a/lib/inode.c +++ b/lib/inode.c @@ -18,6 +18,7 @@ #include "erofs/cache.h" #include "erofs/io.h" #include "erofs/compress.h" +#include "erofs/xattr.h" struct erofs_sb_info sbi; @@ -363,9 +364,11 @@ static bool erofs_bh_flush_write_inode(struct erofs_buffer_head *bh) /* let's support v1 currently */ struct erofs_inode_v1 v1 = {0}; + const u16 icount = EROFS_INODE_XATTR_ICOUNT(inode->xattr_isize); int ret; v1.i_advise = cpu_to_le16(0 | (inode->data_mapping_mode << 1)); + v1.i_xattr_icount = cpu_to_le16(icount); v1.i_mode = cpu_to_le16(inode->i_mode); v1.i_nlink = cpu_to_le16(inode->i_nlink); v1.i_size = cpu_to_le32((u32)inode->i_size); @@ -398,6 +401,20 @@ static bool erofs_bh_flush_write_inode(struct erofs_buffer_head *bh) return false; off += inode->inode_isize; + if (inode->xattr_isize) { + char *xattrs = erofs_export_xattr_ibody(&inode->i_xattrs, + inode->xattr_isize); + if (IS_ERR(xattrs)) + return false; + + ret = dev_write(xattrs, off, inode->xattr_isize); + free(xattrs); + if (ret) + return false; + + off += inode->xattr_isize; + } + if (inode->extent_isize) { /* write compression metadata */ off = Z_EROFS_VLE_EXTENT_ALIGN(off); @@ -612,6 +629,7 @@ struct erofs_inode *erofs_new_inode(void) inode->i_count = 1; init_list_head(&inode->i_subdirs); + init_list_head(&inode->i_xattrs); inode->xattr_isize = 0; inode->extent_isize = 0; @@ -699,6 +717,11 @@ struct erofs_inode *erofs_mkfs_build_tree(struct erofs_inode *dir) struct dirent *dp; struct erofs_dentry *d; + ret = erofs_prepare_xattr_ibody(dir->i_srcpath, &dir->i_xattrs); + if (ret < 0) + return ERR_PTR(ret); + dir->xattr_isize = ret; + if (!S_ISDIR(dir->i_mode)) { if (S_ISLNK(dir->i_mode)) { char *const symlink = malloc(dir->i_size); diff --git a/lib/xattr.c b/lib/xattr.c new file mode 100644 index 0000000..1ab959b --- /dev/null +++ b/lib/xattr.c @@ -0,0 +1,289 @@ +// SPDX-License-Identifier: GPL-2.0+ +/* + * erofs_utils/lib/xattr.c + */ +#include +#include +#ifdef HAVE_LINUX_XATTR_H +#include +#endif +#include "erofs/print.h" +#include "erofs/hashtable.h" +#include "erofs/xattr.h" + +#define EA_HASHTABLE_BITS 16 + +struct xattr_item { + const char *kvbuf; + unsigned int hash[2], len[2], count; + u8 prefix; + struct hlist_node node; +}; + +struct inode_xattr_node { + struct list_head list; + struct xattr_item *item; +}; + +static DECLARE_HASHTABLE(ea_hashtable, EA_HASHTABLE_BITS); + +static struct xattr_prefix { + const char *prefix; + u16 prefix_len; +} prefix[] = { + [EROFS_XATTR_INDEX_USER] = { + XATTR_USER_PREFIX, + XATTR_USER_PREFIX_LEN + }, [EROFS_XATTR_INDEX_POSIX_ACL_ACCESS] = { + XATTR_NAME_POSIX_ACL_ACCESS, + sizeof(XATTR_NAME_POSIX_ACL_ACCESS) - 1 + }, [EROFS_XATTR_INDEX_POSIX_ACL_DEFAULT] = { + XATTR_NAME_POSIX_ACL_DEFAULT, + sizeof(XATTR_NAME_POSIX_ACL_DEFAULT) - 1 + }, [EROFS_XATTR_INDEX_TRUSTED] = { + XATTR_TRUSTED_PREFIX, + XATTR_TRUSTED_PREFIX_LEN + }, [EROFS_XATTR_INDEX_SECURITY] = { + XATTR_SECURITY_PREFIX, + XATTR_SECURITY_PREFIX_LEN + } +}; + +static unsigned int BKDRHash(char *str, unsigned int len) +{ + const unsigned int seed = 131313; + unsigned int hash = 0; + + while (len) { + hash = hash * seed + (*str++); + --len; + } + return hash; +} + +static unsigned int xattr_item_hash(u8 prefix, char *buf, + unsigned int len[2], unsigned int hash[2]) +{ + hash[0] = BKDRHash(buf, len[0]); /* key */ + hash[1] = BKDRHash(buf + len[0], len[1]); /* value */ + + return prefix ^ hash[0] ^ hash[1]; +} + +static struct xattr_item *xattr_item_get(u8 prefix, char *kvbuf, + unsigned int len[2]) +{ + struct xattr_item *item; + unsigned int hash[2], hkey; + + hkey = xattr_item_hash(prefix, kvbuf, len, hash); + + hash_for_each_possible(ea_hashtable, item, node, hkey) { + if (prefix == item->prefix && + item->len[0] == len[0] && item->len[1] == len[1] && + item->hash[0] == hash[0] && item->hash[1] == hash[1] && + !memcmp(kvbuf, item->kvbuf, len[0] + len[1])) { + free(kvbuf); + ++item->count; + return item; + } + } + + item = malloc(sizeof(*item)); + if (!item) { + free(kvbuf); + return ERR_PTR(-ENOMEM); + } + INIT_HLIST_NODE(&item->node); + item->count = 1; + item->kvbuf = kvbuf; + item->len[0] = len[0]; + item->len[1] = len[1]; + item->hash[0] = hash[0]; + item->hash[1] = hash[1]; + item->prefix = prefix; + hash_add(ea_hashtable, &item->node, hkey); + return item; +} + +static bool match_prefix(const char *key, u8 *index, u16 *len) +{ + struct xattr_prefix *p; + + for (p = prefix; p < prefix + ARRAY_SIZE(prefix); ++p) { + if (p->prefix && !strncmp(p->prefix, key, p->prefix_len)) { + *len = p->prefix_len; + *index = p - prefix; + return true; + } + } + return false; +} + +static struct xattr_item *parse_one_xattr(const char *path, const char *key, + unsigned int keylen) +{ + ssize_t ret; + u8 prefix; + u16 prefixlen; + unsigned int len[2]; + char *kvbuf; + + erofs_dbg("parse xattr [%s] of %s", path, key); + + if (!match_prefix(key, &prefix, &prefixlen)) + return ERR_PTR(-ENODATA); + + DBG_BUGON(keylen < prefixlen); + + /* determine length of the value */ + ret = lgetxattr(path, key, NULL, 0); + if (ret < 0) + return ERR_PTR(-errno); + len[1] = ret; + + /* allocate key-value buffer */ + len[0] = keylen - prefixlen; + + kvbuf = malloc(len[0] + len[1]); + if (!kvbuf) + return ERR_PTR(-ENOMEM); + + memcpy(kvbuf, key + prefixlen, len[0]); + if (len[1]) { + /* copy value to buffer */ + ret = lgetxattr(path, key, kvbuf + len[0], len[1]); + if (ret < 0) { + free(kvbuf); + return ERR_PTR(-errno); + } + if (len[1] != ret) { + erofs_err("xattr value changed just now (%u-> %ld)", + len[1], ret); + len[1] = ret; + } + } + return xattr_item_get(prefix, kvbuf, len); +} + +static int inode_xattr_add(struct list_head *hlist, struct xattr_item *item) +{ + struct inode_xattr_node *node = malloc(sizeof(*node)); + + if (!node) + return -ENOMEM; + init_list_head(&node->list); + node->item = item; + list_add(&node->list, hlist); + return 0; +} + +static int read_xattrs_from_file(const char *path, struct list_head *ixattrs) +{ + int ret = 0; + char *keylst, *key; + ssize_t kllen = llistxattr(path, NULL, 0); + + if (kllen < 0 && errno != ENODATA) + return -errno; + if (kllen <= 1) + return 0; + + keylst = malloc(kllen); + if (!keylst) + return -errno; + + /* copy the list of attribute keys to the buffer.*/ + kllen = llistxattr(path, keylst, kllen); + if (kllen < 0) { + ret = -errno; + goto err; + } + + /* + * loop over the list of zero terminated strings with the + * attribute keys. Use the remaining buffer length to determine + * the end of the list. + */ + key = keylst; + while (kllen > 0) { + unsigned int keylen = strlen(key); + struct xattr_item *item = parse_one_xattr(path, key, keylen); + + if (IS_ERR(item)) { + ret = PTR_ERR(item); + goto err; + } + + if (ixattrs) { + ret = inode_xattr_add(ixattrs, item); + if (ret < 0) + goto err; + } + kllen -= keylen + 1; + key += keylen + 1; + } +err: + free(keylst); + return ret; + +} + +int erofs_prepare_xattr_ibody(const char *path, struct list_head *ixattrs) +{ + int ret; + struct inode_xattr_node *node; + + ret = read_xattrs_from_file(path, ixattrs); + if (ret < 0) + return ret; + + if (list_empty(ixattrs)) + return 0; + + /* get xattr ibody size */ + ret = sizeof(struct erofs_xattr_ibody_header); + list_for_each_entry(node, ixattrs, list) { + const struct xattr_item *item = node->item; + + ret += sizeof(struct erofs_xattr_entry); + ret = EROFS_XATTR_ALIGN(ret + item->len[0] + item->len[1]); + } + return ret; +} + +char *erofs_export_xattr_ibody(struct list_head *ixattrs, unsigned int size) +{ + char *buf; + struct inode_xattr_node *node; + struct erofs_xattr_ibody_header *header; + unsigned int p; + + buf = calloc(1, size); + if (!buf) + return ERR_PTR(-ENOMEM); + + header = (struct erofs_xattr_ibody_header *)buf; + header->h_shared_count = 0; + + p = sizeof(struct erofs_xattr_ibody_header); + list_for_each_entry(node, ixattrs, list) { + struct xattr_item *const item = node->item; + const struct erofs_xattr_entry entry = { + .e_name_index = item->prefix, + .e_name_len = item->len[0], + .e_value_size = cpu_to_le16(item->len[1]) + }; + + memcpy(buf + p, &entry, sizeof(entry)); + p += sizeof(struct erofs_xattr_entry); + memcpy(buf + p, item->kvbuf, item->len[0] + item->len[1]); + p = EROFS_XATTR_ALIGN(p + item->len[0] + item->len[1]); + + list_del(&node->list); + free(node); + } + DBG_BUGON(p > size); + return buf; +} + -- 2.17.1