| /linux/fs/smb/server/mgmt/ |
| H A D | user_config.c | 16 struct ksmbd_user *user = NULL; in ksmbd_login_user() local 28 user = ksmbd_alloc_user(resp, resp_ext); in ksmbd_login_user() 31 return user; in ksmbd_login_user() 37 struct ksmbd_user *user; in ksmbd_alloc_user() local 39 user = kmalloc(sizeof(struct ksmbd_user), KSMBD_DEFAULT_GFP); in ksmbd_alloc_user() 40 if (!user) in ksmbd_alloc_user() 43 user->name = kstrdup(resp->account, KSMBD_DEFAULT_GFP); in ksmbd_alloc_user() 44 user->flags = resp->status; in ksmbd_alloc_user() 45 user->gid = resp->gid; in ksmbd_alloc_user() 46 user->uid = resp->uid; in ksmbd_alloc_user() [all …]
|
| H A D | user_config.h | 25 static inline bool user_guest(struct ksmbd_user *user) in user_guest() argument 27 return user->flags & KSMBD_USER_FLAG_GUEST_ACCOUNT; in user_guest() 30 static inline void set_user_flag(struct ksmbd_user *user, int flag) in set_user_flag() argument 32 user->flags |= flag; in set_user_flag() 35 static inline int test_user_flag(struct ksmbd_user *user, int flag) in test_user_flag() argument 37 return user->flags & flag; in test_user_flag() 40 static inline void set_user_guest(struct ksmbd_user *user) in set_user_guest() argument 44 static inline char *user_passkey(struct ksmbd_user *user) in user_passkey() argument 46 return user->passkey; in user_passkey() 49 static inline char *user_name(struct ksmbd_user *user) in user_name() argument [all …]
|
| /linux/arch/arm64/kernel/ |
| H A D | signal.c | 129 static void init_user_layout(struct rt_sigframe_user_layout *user) in init_user_layout() argument 132 sizeof(user->sigframe->uc.uc_mcontext.__reserved); in init_user_layout() 134 memset(user, 0, sizeof(*user)); in init_user_layout() 135 user->size = offsetof(struct rt_sigframe, uc.uc_mcontext.__reserved); in init_user_layout() 137 user->limit = user->size + reserved_size; in init_user_layout() 139 user->limit -= TERMINATOR_SIZE; in init_user_layout() 140 user->limit -= EXTRA_CONTEXT_SIZE; in init_user_layout() 144 static size_t sigframe_size(struct rt_sigframe_user_layout const *user) in sigframe_size() argument 146 return round_up(max(user->size, sizeof(struct rt_sigframe)), 16); in sigframe_size() 157 static int __sigframe_alloc(struct rt_sigframe_user_layout *user, in __sigframe_alloc() argument [all …]
|
| /linux/Documentation/admin-guide/sysctl/ |
| H A D | user.rst | 2 Documentation for /proc/sys/user/ 12 /proc/sys/user. 16 per user per user namespace limits. 24 The creation of per user per user namespace objects are charged to 25 the user in the user namespace who created the object and 26 verified to be below the per user limit in that user namespace. 29 who created user namespaces the creation of the object happens 30 in (user namespaces can be nested) and verified to be below the per user 31 limits in the user namespaces of those users. 34 user namespace does not allow a user to escape their current limits. [all …]
|
| /linux/kernel/trace/ |
| H A D | trace_events_user.c | 201 typedef void (*user_event_func_t) (struct user_event *user, struct iov_iter *i, 209 static struct user_event_mm *user_event_mm_get_all(struct user_event *user); 211 static int destroy_user_event(struct user_event *user); 212 static bool user_fields_match(struct user_event *user, int argc, 231 static struct user_event *user_event_get(struct user_event *user) in user_event_get() argument 233 refcount_inc(&user->refcnt); in user_event_get() 235 return user; in user_event_get() 240 struct user_event *user = container_of( in delayed_destroy_user_event() local 245 if (!refcount_dec_and_test(&user->refcnt)) in delayed_destroy_user_event() 248 if (destroy_user_event(user)) { in delayed_destroy_user_event() [all …]
|
| /linux/arch/arm/mach-omap2/ |
| H A D | omap_hwmod_2xxx_interconnect_data.c | 27 .user = OCP_USER_MPU | OCP_USER_SDMA, 34 .user = OCP_USER_MPU, 47 .user = OCP_USER_MPU | OCP_USER_SDMA, 54 .user = OCP_USER_MPU | OCP_USER_SDMA, 62 .user = OCP_USER_MPU | OCP_USER_SDMA, 70 .user = OCP_USER_MPU | OCP_USER_SDMA, 78 .user = OCP_USER_MPU | OCP_USER_SDMA, 86 .user = OCP_USER_MPU | OCP_USER_SDMA, 94 .user = OCP_USER_MPU | OCP_USER_SDMA, 102 .user = OCP_USER_MPU | OCP_USER_SDMA, [all …]
|
| H A D | omap_hwmod_3xxx_data.c | 1459 .user = OCP_USER_MPU | OCP_USER_SDMA, 1466 .user = OCP_USER_MPU | OCP_USER_SDMA, 1474 .user = OCP_USER_MPU, 1482 .user = OCP_USER_MPU, 1489 .user = OCP_USER_MPU | OCP_USER_SDMA, 1501 .user = OCP_USER_MPU | OCP_USER_SDMA, 1509 .user = OCP_USER_MPU, 1516 .user = OCP_USER_MPU | OCP_USER_SDMA, 1524 .user = OCP_USER_MPU | OCP_USER_SDMA, 1532 .user = OCP_USER_MPU | OCP_USER_SDMA, [all …]
|
| H A D | omap_hwmod_2430_data.c | 380 .user = OCP_USER_MPU, 388 .user = OCP_USER_MPU | OCP_USER_SDMA, 396 .user = OCP_USER_MPU | OCP_USER_SDMA, 404 .user = OCP_USER_MPU, 412 .user = OCP_USER_MPU | OCP_USER_SDMA, 420 .user = OCP_USER_MPU | OCP_USER_SDMA, 428 .user = OCP_USER_MPU | OCP_USER_SDMA, 436 .user = OCP_USER_MPU | OCP_USER_SDMA, 444 .user = OCP_USER_MPU | OCP_USER_SDMA, 452 .user = OCP_USER_MPU | OCP_USER_SDMA, [all …]
|
| H A D | omap_hwmod_81xx_data.c | 150 .user = OCP_USER_MPU, 157 .user = OCP_USER_MPU, 178 .user = OCP_USER_MPU, 185 .user = OCP_USER_MPU, 205 .user = OCP_USER_MPU, 212 .user = OCP_USER_MPU, 248 .user = OCP_USER_MPU, 287 .user = OCP_USER_MPU, 308 .user = OCP_USER_MPU, 329 .user = OCP_USER_MPU, [all …]
|
| H A D | omap_hwmod_2420_data.c | 239 .user = OCP_USER_MPU | OCP_USER_SDMA, 247 .user = OCP_USER_MPU | OCP_USER_SDMA, 255 .user = OCP_USER_MPU | OCP_USER_SDMA, 263 .user = OCP_USER_MPU | OCP_USER_SDMA, 271 .user = OCP_USER_MPU | OCP_USER_SDMA, 279 .user = OCP_USER_MPU | OCP_USER_SDMA, 287 .user = OCP_USER_MPU | OCP_USER_SDMA, 295 .user = OCP_USER_MPU | OCP_USER_SDMA, 303 .user = OCP_USER_MPU | OCP_USER_SDMA, 310 .user = OCP_USER_MPU | OCP_USER_SDMA, [all …]
|
| /linux/drivers/mtd/nand/raw/atmel/ |
| H A D | pmecc.c | 347 struct atmel_pmecc_user *user; in atmel_pmecc_create_user() local 355 size = sizeof(*user); in atmel_pmecc_create_user() 366 user = devm_kzalloc(pmecc->dev, size, GFP_KERNEL); in atmel_pmecc_create_user() 367 if (!user) in atmel_pmecc_create_user() 370 user->pmecc = pmecc; in atmel_pmecc_create_user() 372 user->partial_syn = (s16 *)PTR_ALIGN(user + 1, sizeof(u16)); in atmel_pmecc_create_user() 373 user->si = user->partial_syn + ((2 * req->ecc.strength) + 1); in atmel_pmecc_create_user() 374 user->lmu = user->si + ((2 * req->ecc.strength) + 1); in atmel_pmecc_create_user() 375 user->smu = user->lmu + (req->ecc.strength + 1); in atmel_pmecc_create_user() 376 user->mu = (s32 *)PTR_ALIGN(user->smu + in atmel_pmecc_create_user() [all …]
|
| /linux/scripts/ |
| H A D | Makefile.userprogs | 7 user-csingle := $(foreach m, $(userprogs), $(if $($(m)-objs),,$(m))) 10 user-cmulti := $(foreach m, $(userprogs), $(if $($(m)-objs),$(m))) 13 user-cobjs := $(sort $(foreach m, $(userprogs), $($(m)-objs))) 15 user-csingle := $(addprefix $(obj)/, $(user-csingle)) 16 user-cmulti := $(addprefix $(obj)/, $(user-cmulti)) 17 user-cobjs := $(addprefix $(obj)/, $(user-cobjs)) 28 $(user-csingle): $(obj)/%: $(src)/%.c FORCE 35 $(user-cmulti): FORCE 37 $(call multi_depend, $(user-cmulti), , -objs) 42 $(user-cobjs): $(obj)/%.o: $(src)/%.c FORCE [all …]
|
| /linux/include/linux/ |
| H A D | ipmi.h | 48 struct ipmi_user *user; member 120 struct ipmi_user **user); 130 void ipmi_destroy_user(struct ipmi_user *user); 133 int ipmi_get_version(struct ipmi_user *user, 146 int ipmi_set_my_address(struct ipmi_user *user, 149 int ipmi_get_my_address(struct ipmi_user *user, 152 int ipmi_set_my_LUN(struct ipmi_user *user, 155 int ipmi_get_my_LUN(struct ipmi_user *user, 172 int ipmi_request_settime(struct ipmi_user *user, 190 int ipmi_request_supply_msgs(struct ipmi_user *user, [all …]
|
| /linux/drivers/fsi/ |
| H A D | fsi-sbefifo.c | 782 static void sbefifo_release_command(struct sbefifo_user *user) in sbefifo_release_command() argument 784 if (is_vmalloc_addr(user->pending_cmd)) in sbefifo_release_command() 785 vfree(user->pending_cmd); in sbefifo_release_command() 786 user->pending_cmd = NULL; in sbefifo_release_command() 787 user->pending_len = 0; in sbefifo_release_command() 793 struct sbefifo_user *user; in sbefifo_user_open() local 795 user = kzalloc(sizeof(struct sbefifo_user), GFP_KERNEL); in sbefifo_user_open() 796 if (!user) in sbefifo_user_open() 799 file->private_data = user; in sbefifo_user_open() 800 user->sbefifo = sbefifo; in sbefifo_user_open() [all …]
|
| /linux/drivers/iommu/iommufd/ |
| H A D | pages.c | 765 static void pfn_reader_user_init(struct pfn_reader_user *user, in pfn_reader_user_init() argument 768 user->upages = NULL; in pfn_reader_user_init() 769 user->upages_len = 0; in pfn_reader_user_init() 770 user->upages_start = 0; in pfn_reader_user_init() 771 user->upages_end = 0; in pfn_reader_user_init() 772 user->locked = -1; in pfn_reader_user_init() 773 user->gup_flags = FOLL_LONGTERM; in pfn_reader_user_init() 775 user->gup_flags |= FOLL_WRITE; in pfn_reader_user_init() 777 user->file = (pages->type == IOPT_ADDRESS_FILE) ? pages->file : NULL; in pfn_reader_user_init() 778 user->ufolios = NULL; in pfn_reader_user_init() [all …]
|
| /linux/drivers/net/ethernet/freescale/enetc/ |
| H A D | ntmp.c | 110 static int netc_xmit_ntmp_cmd(struct ntmp_user *user, union netc_cbd *cbd) in netc_xmit_ntmp_cmd() argument 121 cbdr = &user->ring[0]; in netc_xmit_ntmp_cmd() 154 dev_err(user->dev, "Command BD error: 0x%04x\n", status); in netc_xmit_ntmp_cmd() 233 static int ntmp_delete_entry_by_id(struct ntmp_user *user, int tbl_id, in ntmp_delete_entry_by_id() argument 238 .dev = user->dev, in ntmp_delete_entry_by_id() 253 err = netc_xmit_ntmp_cmd(user, &cbd); in ntmp_delete_entry_by_id() 255 dev_err(user->dev, in ntmp_delete_entry_by_id() 264 static int ntmp_query_entry_by_id(struct ntmp_user *user, int tbl_id, in ntmp_query_entry_by_id() argument 280 err = netc_xmit_ntmp_cmd(user, &cbd); in ntmp_query_entry_by_id() 282 dev_err(user->dev, in ntmp_query_entry_by_id() [all …]
|
| /linux/arch/m68k/ifpsp060/ |
| H A D | os.S | 57 | Each IO routine checks to see if the memory write/read is to/from user 60 | for user mode applications. 71 | working in user space 80 | a1 - user destination address 82 | 0x4(%a6),bit5 - 1 = supervisor mode, 0 = user mode 110 | a0 - user source address 113 | 0x4(%a6),bit5 - 1 = supervisor mode, 0 = user mode 140 | Read a data byte from user memory. 143 | a0 - user source address 144 | 0x4(%a6),bit5 - 1 = supervisor mode, 0 = user mode [all …]
|
| /linux/drivers/char/ipmi/ |
| H A D | ipmi_poweroff.c | 45 static void (*specific_poweroff_func)(struct ipmi_user *user); 118 static int ipmi_request_wait_for_response(struct ipmi_user *user, in ipmi_request_wait_for_response() argument 127 rv = ipmi_request_supply_msgs(user, addr, 0, send_msg, &comp, in ipmi_request_wait_for_response() 138 static int ipmi_request_in_rc_mode(struct ipmi_user *user, in ipmi_request_in_rc_mode() argument 145 rv = ipmi_request_supply_msgs(user, addr, 0, send_msg, NULL, in ipmi_request_in_rc_mode() 156 ipmi_poll_interface(user); in ipmi_request_in_rc_mode() 178 static void (*atca_oem_poweroff_hook)(struct ipmi_user *user); 180 static void pps_poweroff_atca(struct ipmi_user *user) in pps_poweroff_atca() argument 198 rv = ipmi_request_in_rc_mode(user, in pps_poweroff_atca() 207 static int ipmi_atca_detect(struct ipmi_user *user) in ipmi_atca_detect() argument [all …]
|
| H A D | ipmi_msghandler.c | 41 static struct ipmi_recv_msg *ipmi_alloc_recv_msg(struct ipmi_user *user); 43 struct ipmi_user *user); 206 struct ipmi_user *user; member 606 struct ipmi_user *user = container_of(ref, struct ipmi_user, refcount); in free_ipmi_user() local 609 owner = user->intf->owner; in free_ipmi_user() 610 kref_put(&user->intf->refcount, intf_free); in free_ipmi_user() 612 vfree(user); in free_ipmi_user() 615 static void release_ipmi_user(struct ipmi_user *user) in release_ipmi_user() argument 617 kref_put(&user->refcount, free_ipmi_user); in release_ipmi_user() 620 static struct ipmi_user *acquire_ipmi_user(struct ipmi_user *user) in acquire_ipmi_user() argument [all …]
|
| /linux/security/keys/ |
| H A D | key.c | 52 struct key_user *candidate = NULL, *user; in key_user_lookup() local 63 user = rb_entry(parent, struct key_user, node); in key_user_lookup() 65 if (uid_lt(uid, user->uid)) in key_user_lookup() 67 else if (uid_gt(uid, user->uid)) in key_user_lookup() 79 user = NULL; in key_user_lookup() 104 user = candidate; in key_user_lookup() 109 refcount_inc(&user->usage); in key_user_lookup() 113 return user; in key_user_lookup() 119 void key_user_put(struct key_user *user) in key_user_put() argument 121 if (refcount_dec_and_lock(&user->usage, &key_user_lock)) { in key_user_put() [all …]
|
| /linux/net/bridge/netfilter/ |
| H A D | ebt_mark_m.c | 48 const struct compat_ebt_mark_m_info *user = src; in mark_mt_compat_from_user() local 51 kern->mark = user->mark; in mark_mt_compat_from_user() 52 kern->mask = user->mask; in mark_mt_compat_from_user() 53 kern->invert = user->invert; in mark_mt_compat_from_user() 54 kern->bitmask = user->bitmask; in mark_mt_compat_from_user() 59 struct compat_ebt_mark_m_info __user *user = dst; in mark_mt_compat_to_user() local 62 if (put_user(kern->mark, &user->mark) || in mark_mt_compat_to_user() 63 put_user(kern->mask, &user->mask) || in mark_mt_compat_to_user() 64 put_user(kern->invert, &user->invert) || in mark_mt_compat_to_user() 65 put_user(kern->bitmask, &user->bitmask)) in mark_mt_compat_to_user()
|
| /linux/include/linux/fsl/ |
| H A D | ntmp.h | 70 int ntmp_maft_add_entry(struct ntmp_user *user, u32 entry_id, 72 int ntmp_maft_query_entry(struct ntmp_user *user, u32 entry_id, 74 int ntmp_maft_delete_entry(struct ntmp_user *user, u32 entry_id); 75 int ntmp_rsst_update_entry(struct ntmp_user *user, const u32 *table, 77 int ntmp_rsst_query_entry(struct ntmp_user *user, 90 static inline int ntmp_maft_add_entry(struct ntmp_user *user, u32 entry_id, in ntmp_maft_add_entry() argument 96 static inline int ntmp_maft_query_entry(struct ntmp_user *user, u32 entry_id, in ntmp_maft_query_entry() argument 102 static inline int ntmp_maft_delete_entry(struct ntmp_user *user, u32 entry_id) in ntmp_maft_delete_entry() argument 107 static inline int ntmp_rsst_update_entry(struct ntmp_user *user, in ntmp_rsst_update_entry() argument 113 static inline int ntmp_rsst_query_entry(struct ntmp_user *user, in ntmp_rsst_query_entry() argument
|
| /linux/arch/mips/kernel/ |
| H A D | unaligned.c | 118 bool user = user_mode(regs); in emulate_load_store_insn() local 128 __get_inst32(&insn.word, pc, user); in emulate_load_store_insn() 171 if (user && !access_ok(addr, 4)) in emulate_load_store_insn() 180 if (user && !access_ok(addr, 2)) in emulate_load_store_insn() 189 if (user && !access_ok(addr, 2)) in emulate_load_store_insn() 209 if (user && !access_ok(addr, 4)) in emulate_load_store_insn() 218 if (user && !access_ok(addr, 2)) in emulate_load_store_insn() 290 if (user && !access_ok(addr, 2)) in emulate_load_store_insn() 293 if (IS_ENABLED(CONFIG_EVA) && user) in emulate_load_store_insn() 305 if (user && !access_ok(addr, 4)) in emulate_load_store_insn() [all …]
|
| /linux/drivers/gpu/drm/nouveau/nvkm/engine/disp/ |
| H A D | chan.c | 55 *addr = base + chan->func->user(chan, size); in nvkm_disp_chan_map() 156 if (disp->chan[chan->chid.user] == chan) in nvkm_disp_chan_dtor() 157 disp->chan[chan->chid.user] = NULL; in nvkm_disp_chan_dtor() 178 const struct nvkm_disp_chan_user *user = NULL; in nvkm_disp_chan_new_() local 183 for (i = 0; disp->func->user[i].ctor; i++) { in nvkm_disp_chan_new_() 184 if (disp->func->user[i].base.oclass == oclass->base.oclass) { in nvkm_disp_chan_new_() 185 user = disp->func->user[i].chan; in nvkm_disp_chan_new_() 190 if (WARN_ON(!user)) in nvkm_disp_chan_new_() 195 if (args->v0.id >= nr || !args->v0.pushbuf != !user->func->push) in nvkm_disp_chan_new_() 203 chan->func = user->func; in nvkm_disp_chan_new_() [all …]
|
| /linux/Documentation/gpu/amdgpu/ |
| H A D | userq.rst | 9 user processes manage their submissions to the GPU engines directly, bypassing 28 scheduling firmware which handles dynamically scheduling user queues on the 29 available hardware queue slots. When the number of user queues outnumbers the 31 unmaps queues based on priority and time quanta. The state of each user queue 33 buffer in GPU accessible memory that stores the state of a user queue. The 35 Queue Descriptor) when a user queue is mapped. Each user queue requires a 41 device's MMIO BARs which can be mapped to specific user queues. When the 50 512 user queues. A subset of each page is reserved for each IP type supported 51 on the device. The user can query the doorbell ranges for each IP via the INFO 54 When an application wants to create a user queue, it allocates the necessary [all …]
|