/linux/drivers/infiniband/hw/mlx4/ |
H A D | mcg.c | 50 #define mcg_warn_group(group, format, arg...) \ argument 52 (group)->name, group->demux->port, ## arg) 54 #define mcg_debug_group(group, format, arg...) \ argument 56 (group)->name, (group)->demux->port, ## arg) 58 #define mcg_error_group(group, format, arg...) \ argument 59 pr_err(" %16s: " format, (group)->name, ## arg) 136 struct mcast_group *group; member 144 mcg_warn_group(group, "did not expect to reach zero\n"); \ 166 struct mcast_group *group; in mcast_find() local 170 group = rb_entry(node, struct mcast_group, node); in mcast_find() [all …]
|
/linux/fs/notify/ |
H A D | group.c | 20 * Final freeing of a group 22 static void fsnotify_final_destroy_group(struct fsnotify_group *group) in fsnotify_final_destroy_group() argument 24 if (group->ops->free_group_priv) in fsnotify_final_destroy_group() 25 group->ops->free_group_priv(group); in fsnotify_final_destroy_group() 27 mem_cgroup_put(group->memcg); in fsnotify_final_destroy_group() 28 mutex_destroy(&group->mark_mutex); in fsnotify_final_destroy_group() 30 kfree(group); in fsnotify_final_destroy_group() 34 * Stop queueing new events for this group. Once this function returns 35 * fsnotify_add_event() will not add any new events to the group's queue. 37 void fsnotify_group_stop_queueing(struct fsnotify_group *group) in fsnotify_group_stop_queueing() argument [all …]
|
H A D | notification.c | 7 * Basic idea behind the notification queue: An fsnotify group (like inotify) 10 * event to the group notify queue. Since a single event might need to be on 11 * multiple group's notification queues we can't add the event directly to each 17 * another group a new event_holder (from fsnotify_event_holder_cachep) will be 50 void fsnotify_destroy_event(struct fsnotify_group *group, in fsnotify_destroy_event() argument 53 /* Overflow events are per-group and we don't want to free them */ in fsnotify_destroy_event() 54 if (!event || event == group->overflow_event) in fsnotify_destroy_event() 63 spin_lock(&group->notification_lock); in fsnotify_destroy_event() 65 spin_unlock(&group->notification_lock); in fsnotify_destroy_event() 67 group->ops->free_event(group, event); in fsnotify_destroy_event() [all …]
|
/linux/drivers/infiniband/core/ |
H A D | multicast.c | 117 struct mcast_group *group; member 133 struct mcast_group *group; in mcast_find() local 137 group = rb_entry(node, struct mcast_group, node); in mcast_find() 138 ret = memcmp(mgid->raw, group->rec.mgid.raw, sizeof *mgid); in mcast_find() 140 return group; in mcast_find() 151 struct mcast_group *group, in mcast_insert() argument 163 ret = memcmp(group->rec.mgid.raw, cur_group->rec.mgid.raw, in mcast_insert() 164 sizeof group->rec.mgid); in mcast_insert() 174 rb_link_node(&group->node, parent, link); in mcast_insert() 175 rb_insert_color(&group->node, &port->table); in mcast_insert() [all …]
|
/linux/drivers/gpu/drm/panthor/ |
H A D | panthor_sched.c | 41 * The scheduling happens at the scheduling group level, each group 50 * rotating the groups passed to the firmware so every group gets 66 * queue ring-buffer, and the group is scheduled for execution if it 69 * Kernel-side group scheduling is timeslice-based. When we have less 72 * groups than slots, we let each group a chance to execute stuff for 74 * to schedule. The group selection algorithm is based on 79 * group/queue state that would be based on information we wouldn't have 81 * reason we don't do 'cooperative' scheduling (encoding FW group slot 83 * drm_gpu_scheduler::prepare_job() hook, and treating group rotation as 99 * struct panthor_csg_slot - Command stream group slot [all …]
|
/linux/arch/sparc/kernel/ |
H A D | hvapi.c | 19 unsigned long group; member 28 { .group = HV_GRP_SUN4V, .flags = FLAG_PRE_API }, 29 { .group = HV_GRP_CORE, .flags = FLAG_PRE_API }, 30 { .group = HV_GRP_INTR, }, 31 { .group = HV_GRP_SOFT_STATE, }, 32 { .group = HV_GRP_TM, }, 33 { .group = HV_GRP_PCI, .flags = FLAG_PRE_API }, 34 { .group = HV_GRP_LDOM, }, 35 { .group = HV_GRP_SVC_CHAN, .flags = FLAG_PRE_API }, 36 { .group = HV_GRP_NCS, .flags = FLAG_PRE_API }, [all …]
|
/linux/Documentation/filesystems/ext4/ |
H A D | blockgroup.rst | 6 The layout of a standard block group is approximately as follows (each 13 * - Group 0 Padding 15 - Group Descriptors 30 For the special case of block group 0, the first 1024 bytes are unused, 37 The ext4 driver primarily works with the superblock and the group 38 descriptors that are found in block group 0. Redundant copies of the 39 superblock and group descriptors are written to some of the block groups 42 paragraph for more details). If the group does not have a redundant 43 copy, the block group begins with the data block bitmap. Note also that 45 GDT block” space after the block group descriptors and before the start [all …]
|
H A D | group_descr.rst | 3 Block Group Descriptors 6 Each block group on the filesystem has one of these descriptors 7 associated with it. As noted in the Layout section above, the group 8 descriptors (if present) are the second item in the block group. The 9 standard configuration is for each block group to contain a full copy of 10 the block group descriptor table unless the sparse_super feature flag 13 Notice how the group descriptor records the location of both bitmaps and 15 group, the only data structures with fixed locations are the superblock 16 and the group descriptor table. The flex_bg mechanism uses this 17 property to group several block groups into a flex group and lay out all [all …]
|
/linux/tools/testing/selftests/drivers/net/netdevsim/ |
H A D | nexthop.sh | 158 $IP nexthop add id 10 group 1/2 159 nexthop_check "id 10" "id 10 group 1/2 trap" 160 check_err $? "Unexpected nexthop group entry" 169 $IP nexthop add id 10 group 1,20/2,39 170 nexthop_check "id 10" "id 10 group 1,20/2,39 trap" 171 check_err $? "Unexpected weighted nexthop group entry" 180 log_test "Nexthop group add and delete" 194 $IP nexthop add id 10 group 1/2 &> /dev/null 195 check_fail $? "Nexthop group addition succeeded when should fail" 200 log_test "Nexthop group add failure" [all …]
|
/linux/drivers/net/ethernet/microchip/sparx5/ |
H A D | sparx5_sdlb.c | 59 static void sparx5_sdlb_group_disable(struct sparx5 *sparx5, u32 group) in sparx5_sdlb_group_disable() argument 63 ANA_AC_SDLB_PUP_CTRL(group)); in sparx5_sdlb_group_disable() 66 static void sparx5_sdlb_group_enable(struct sparx5 *sparx5, u32 group) in sparx5_sdlb_group_enable() argument 70 ANA_AC_SDLB_PUP_CTRL(group)); in sparx5_sdlb_group_enable() 73 static u32 sparx5_sdlb_group_get_first(struct sparx5 *sparx5, u32 group) in sparx5_sdlb_group_get_first() argument 77 val = spx5_rd(sparx5, ANA_AC_SDLB_XLB_START(group)); in sparx5_sdlb_group_get_first() 82 static u32 sparx5_sdlb_group_get_next(struct sparx5 *sparx5, u32 group, in sparx5_sdlb_group_get_next() argument 92 static bool sparx5_sdlb_group_is_first(struct sparx5 *sparx5, u32 group, in sparx5_sdlb_group_is_first() argument 95 return lb == sparx5_sdlb_group_get_first(sparx5, group); in sparx5_sdlb_group_is_first() 98 static bool sparx5_sdlb_group_is_last(struct sparx5 *sparx5, u32 group, in sparx5_sdlb_group_is_last() argument [all …]
|
H A D | sparx5_qos.c | 79 static u32 sparx5_lg_get_leak_time(struct sparx5 *sparx5, u32 layer, u32 group) in sparx5_lg_get_leak_time() argument 83 value = spx5_rd(sparx5, HSCH_HSCH_TIMER_CFG(layer, group)); in sparx5_lg_get_leak_time() 87 static void sparx5_lg_set_leak_time(struct sparx5 *sparx5, u32 layer, u32 group, in sparx5_lg_set_leak_time() argument 91 HSCH_HSCH_TIMER_CFG(layer, group)); in sparx5_lg_set_leak_time() 94 static u32 sparx5_lg_get_first(struct sparx5 *sparx5, u32 layer, u32 group) in sparx5_lg_get_first() argument 98 value = spx5_rd(sparx5, HSCH_HSCH_LEAK_CFG(layer, group)); in sparx5_lg_get_first() 102 static u32 sparx5_lg_get_next(struct sparx5 *sparx5, u32 layer, u32 group, in sparx5_lg_get_next() argument 112 static u32 sparx5_lg_get_last(struct sparx5 *sparx5, u32 layer, u32 group) in sparx5_lg_get_last() argument 116 itr = sparx5_lg_get_first(sparx5, layer, group); in sparx5_lg_get_last() 119 next = sparx5_lg_get_next(sparx5, layer, group, itr); in sparx5_lg_get_last() [all …]
|
/linux/drivers/gpio/ |
H A D | gpio-lpc32xx.c | 168 static inline u32 gpreg_read(struct lpc32xx_gpio_chip *group, unsigned long offset) in gpreg_read() argument 170 return __raw_readl(group->reg_base + offset); in gpreg_read() 173 static inline void gpreg_write(struct lpc32xx_gpio_chip *group, u32 val, unsigned long offset) in gpreg_write() argument 175 __raw_writel(val, group->reg_base + offset); in gpreg_write() 178 static void __set_gpio_dir_p012(struct lpc32xx_gpio_chip *group, in __set_gpio_dir_p012() argument 182 gpreg_write(group, GPIO012_PIN_TO_BIT(pin), in __set_gpio_dir_p012() 183 group->gpio_grp->dir_clr); in __set_gpio_dir_p012() 185 gpreg_write(group, GPIO012_PIN_TO_BIT(pin), in __set_gpio_dir_p012() 186 group->gpio_grp->dir_set); in __set_gpio_dir_p012() 189 static void __set_gpio_dir_p3(struct lpc32xx_gpio_chip *group, in __set_gpio_dir_p3() argument [all …]
|
H A D | gpio-104-dio-48e.c | 195 "PPI Group 0 Port A 0", "PPI Group 0 Port A 1", "PPI Group 0 Port A 2", 196 "PPI Group 0 Port A 3", "PPI Group 0 Port A 4", "PPI Group 0 Port A 5", 197 "PPI Group 0 Port A 6", "PPI Group 0 Port A 7", "PPI Group 0 Port B 0", 198 "PPI Group 0 Port B 1", "PPI Group 0 Port B 2", "PPI Group 0 Port B 3", 199 "PPI Group 0 Port B 4", "PPI Group 0 Port B 5", "PPI Group 0 Port B 6", 200 "PPI Group 0 Port B 7", "PPI Group 0 Port C 0", "PPI Group 0 Port C 1", 201 "PPI Group 0 Port C 2", "PPI Group 0 Port C 3", "PPI Group 0 Port C 4", 202 "PPI Group 0 Port C 5", "PPI Group 0 Port C 6", "PPI Group 0 Port C 7", 203 "PPI Group 1 Port A 0", "PPI Group 1 Port A 1", "PPI Group 1 Port A 2", 204 "PPI Group 1 Port A 3", "PPI Group 1 Port A 4", "PPI Group 1 Port A 5", [all …]
|
/linux/drivers/net/ethernet/mellanox/mlx5/core/esw/ |
H A D | qos.c | 45 static int esw_qos_group_config(struct mlx5_eswitch *esw, struct mlx5_esw_rate_group *group, in esw_qos_group_config() argument 53 group->tsar_ix, in esw_qos_group_config() 56 NL_SET_ERR_MSG_MOD(extack, "E-Switch modify group TSAR element failed"); in esw_qos_group_config() 58 trace_mlx5_esw_group_qos_config(dev, group, group->tsar_ix, bw_share, max_rate); in esw_qos_group_config() 91 struct mlx5_esw_rate_group *group, in esw_qos_calculate_min_rate_divider() argument 100 struct mlx5_esw_rate_group *group; in esw_qos_calculate_min_rate_divider() local 102 list_for_each_entry(group, &esw->qos.groups, list) { in esw_qos_calculate_min_rate_divider() 103 if (group->min_rate < max_guarantee) in esw_qos_calculate_min_rate_divider() 105 max_guarantee = group->min_rate; in esw_qos_calculate_min_rate_divider() 110 evport->qos.group != group || evport->qos.min_rate < max_guarantee) in esw_qos_calculate_min_rate_divider() [all …]
|
/linux/tools/testing/selftests/net/ |
H A D | fib_nexthops.sh | 347 run_cmd "$IP nexthop add id $grpidstr group $grpstr" 348 check_nexthop "id $grpidstr" "id $grpidstr group $grpstr" 369 # create a resilient group with $buckets buckets and dump them 371 run_cmd "$IP nexthop add id 1000 group 100 type resilient buckets $buckets" 437 echo "SKIP: iproute2 too old, missing resilient nexthop group support" 455 # create group with multiple nexthops 458 run_cmd "$IP nexthop add id 102 group 61/62 fdb" 459 check_nexthop "id 102" "id 102 group 61/62 fdb" 460 log_test $? 0 "Fdb Nexthop group with multiple nexthops" 462 ## get nexthop group [all …]
|
/linux/kernel/sched/ |
H A D | psi.c | 179 static void group_init(struct psi_group *group) in group_init() argument 183 group->enabled = true; in group_init() 185 seqcount_init(&per_cpu_ptr(group->pcpu, cpu)->seq); in group_init() 186 group->avg_last_update = sched_clock(); in group_init() 187 group->avg_next_update = group->avg_last_update + psi_period; in group_init() 188 mutex_init(&group->avgs_lock); in group_init() 191 INIT_LIST_HEAD(&group->avg_triggers); in group_init() 192 memset(group->avg_nr_triggers, 0, sizeof(group->avg_nr_triggers)); in group_init() 193 INIT_DELAYED_WORK(&group->avgs_work, psi_avgs_work); in group_init() 196 atomic_set(&group->rtpoll_scheduled, 0); in group_init() [all …]
|
/linux/kernel/time/ |
H A D | timer_migration.c | 22 * lowest level group contains CPUs, the next level groups of CPU groups 43 * Each group has a designated migrator CPU/group as long as a CPU/group is 44 * active in the group. This designated role is necessary to avoid that all 45 * active CPUs in a group try to migrate expired timers from other CPUs, 48 * When a CPU is awake, it checks in it's own timer tick the group 53 * If it finds expired timers in one of the group queues it pulls them over 55 * group and the parent groups if required. 60 * CPU does not queue an event in the LVL0 group. If the next migratable 62 * in the LVL0 group. In both cases the CPU marks itself idle in the LVL0 63 * group. [all …]
|
/linux/include/trace/events/ |
H A D | timer_migration.h | 11 /* Group events */ 14 TP_PROTO(struct tmigr_group *group), 16 TP_ARGS(group), 19 __field( void *, group ) 25 __entry->group = group; 26 __entry->lvl = group->level; 27 __entry->numa_node = group->numa_node; 30 TP_printk("group=%p lvl=%d numa=%d", 31 __entry->group, __entry->lvl, __entry->numa_node) 58 TP_printk("group=%p groupmask=%0x parent=%p lvl=%d numa=%d num_children=%d", [all …]
|
/linux/Documentation/devicetree/bindings/pinctrl/ |
H A D | marvell,armada-37xx-pinctrl.txt | 32 group: jtag 36 group sdio0 40 group emmc_nb 44 group pwm0 48 group pwm1 52 group pwm2 56 group pwm3 60 group pmic1 64 group pmic0 68 group i2c2 [all …]
|
/linux/fs/ext2/ |
H A D | ialloc.c | 31 * blocks groups. Each group contains 1 bitmap block for blocks, 1 bitmap 34 * The file system contains group descriptors which are located after the 66 static void ext2_release_inode(struct super_block *sb, int group, int dir) in ext2_release_inode() argument 71 desc = ext2_get_group_desc(sb, group, &bh); in ext2_release_inode() 74 "can't get descriptor for group %d", group); in ext2_release_inode() 78 spin_lock(sb_bgl_lock(EXT2_SB(sb), group)); in ext2_release_inode() 82 spin_unlock(sb_bgl_lock(EXT2_SB(sb), group)); in ext2_release_inode() 180 * Figure out the offset within the block group inode table in ext2_preread_inode() 191 * a directory, then a forward search is made for a block group with both 193 * the groups with above-average free space, that group with the fewest [all …]
|
/linux/fs/ext4/ |
H A D | resize.c | 108 ext4_group_t group) { in ext4_group_overhead_blocks() argument 110 overhead = ext4_bg_num_gdb(sb, group); in ext4_group_overhead_blocks() 111 if (ext4_bg_has_super(sb, group)) in ext4_group_overhead_blocks() 127 ext4_group_t group = input->group; in verify_group_input() local 135 if (group != sbi->s_groups_count) { in verify_group_input() 136 ext4_warning(sb, "Cannot add at group %u (only %u groups)", in verify_group_input() 137 input->group, sbi->s_groups_count); in verify_group_input() 141 overhead = ext4_group_overhead_blocks(sb, group); in verify_group_input() 148 printk(KERN_DEBUG "EXT4-fs: adding %s group in verify_group_input() 313 ext4_group_t group; ext4_alloc_group_tables() local 471 ext4_group_t group; set_flexbg_block_bitmap() local 540 ext4_group_t group, count; setup_new_flex_group_blocks() local 819 add_new_gdb(handle_t * handle,struct inode * inode,ext4_group_t group) add_new_gdb() argument 961 add_new_gdb_meta_bg(struct super_block * sb,handle_t * handle,ext4_group_t group) add_new_gdb_meta_bg() argument 1018 reserve_backup_gdb(handle_t * handle,struct inode * inode,ext4_group_t group) reserve_backup_gdb() argument 1114 ext4_set_block_group_nr(struct super_block * sb,char * data,ext4_group_t group) ext4_set_block_group_nr() argument 1148 ext4_group_t group = 0; update_backups() local 1252 ext4_add_new_descs(handle_t * handle,struct super_block * sb,ext4_group_t group,struct inode * resize_inode,ext4_group_t count) ext4_add_new_descs() argument 1345 ext4_group_t group; ext4_setup_new_descs() local 1543 ext4_group_t group; ext4_flex_group_add() local 1634 ext4_group_t group; ext4_setup_next_flex_gd() local 1841 ext4_group_t group; ext4_group_extend() local [all...] |
/linux/drivers/net/ethernet/netronome/nfp/flower/ |
H A D | lag_conf.c | 6 /* LAG group config flags. */ 33 * @group_id: Group ID applicable 34 * @group_inst: Group instance number - increment when group is reused 35 * @members: Array of 32-bit words listing all active group members 49 * struct nfp_fl_lag_group - list entry for each LAG group 50 * @group_id: Assigned group ID for host/kernel sync 51 * @group_inst: Group instance in case of ID reuse 53 * @master_ndev: Group master Netdev 54 * @dirty: Marked if the group needs synced to HW 55 * @offloaded: Marked if the group is currently offloaded to NIC [all …]
|
/linux/tools/testing/selftests/tc-testing/tc-tests/actions/ |
H A D | sample.json | 20 "cmdUnderTest": "$TC actions add action sample rate 10 group 1 index 2", 23 "matchPattern": "action order [0-9]+: sample rate 1/10 group 1.*index 2 ref", 47 "cmdUnderTest": "$TC actions add action sample rate 700 group 2 continue index 2", 50 "matchPattern": "action order [0-9]+: sample rate 1/700 group 2 continue.*index 2 ref", 74 "cmdUnderTest": "$TC actions add action sample rate 10000 group 11 drop index 22", 77 "matchPattern": "action order [0-9]+: sample rate 1/10000 group 11 drop.*index 22 ref", 101 "cmdUnderTest": "$TC actions add action sample rate 20000 group 72 reclassify index 100", 104 … "matchPattern": "action order [0-9]+: sample rate 1/20000 group 72 reclassify.*index 100 ref", 128 "cmdUnderTest": "$TC actions add action sample rate 20 group 2 pipe index 100", 131 "matchPattern": "action order [0-9]+: sample rate 1/20 group 2 pipe.*index 100 ref", [all …]
|
/linux/include/linux/ |
H A D | fsnotify_backend.h | 122 * Each group much define these ops. The fsnotify infrastructure will call 123 * these operations for each relevant group. 125 * handle_event - main call for a group to handle an fs event 126 * @group: group to notify 135 * @iter_info: array of marks from this group that are interested in the event 149 * free_group_priv - called when a group refcnt hits 0 to clean up the private union 150 * freeing_mark - called when a mark is being destroyed for some reason. The group 156 int (*handle_event)(struct fsnotify_group *group, u32 mask, 163 void (*free_group_priv)(struct fsnotify_group *group); 164 void (*freeing_mark)(struct fsnotify_mark *mark, struct fsnotify_group *group); [all …]
|
/linux/fs/nilfs2/ |
H A D | alloc.c | 21 * nilfs_palloc_groups_per_desc_block - get the number of groups that a group 63 * Number of blocks in a group including entry blocks in nilfs_palloc_init_blockgroup() 77 * nilfs_palloc_group - get group number and offset from an entry number 80 * @offset: pointer to store offset number in the group 85 __u64 group = nr; in nilfs_palloc_group() local 87 *offset = do_div(group, nilfs_palloc_entries_per_group(inode)); in nilfs_palloc_group() 88 return group; in nilfs_palloc_group() 92 * nilfs_palloc_desc_blkoff - get block offset of a group descriptor block 94 * @group: group number 97 * block which contains a descriptor of the specified group. [all …]
|