/linux/drivers/net/ethernet/intel/ice/ |
H A D | ice_lag.c | 32 static void ice_lag_set_primary(struct ice_lag *lag) in ice_lag_set_primary() argument 34 struct ice_pf *pf = lag->pf; in ice_lag_set_primary() 39 if (lag->role != ICE_LAG_UNSET && lag->role != ICE_LAG_BACKUP) { in ice_lag_set_primary() 41 netdev_name(lag->netdev)); in ice_lag_set_primary() 45 lag->role = ICE_LAG_PRIMARY; in ice_lag_set_primary() 52 static void ice_lag_set_backup(struct ice_lag *lag) in ice_lag_set_backup() argument 54 struct ice_pf *pf = lag->pf; in ice_lag_set_backup() 59 if (lag->role != ICE_LAG_UNSET && lag->role != ICE_LAG_PRIMARY) { in ice_lag_set_backup() 61 netdev_name(lag->netdev)); in ice_lag_set_backup() 65 lag->role = ICE_LAG_BACKUP; in ice_lag_set_backup() [all …]
|
H A D | ice_dcb_nl.c | 73 if (pf->lag && pf->lag->bonded) { in ice_dcbnl_setets() 178 if (pf->lag && pf->lag->bonded) { in ice_dcbnl_setdcbx() 274 if (pf->lag && pf->lag->bonded) { in ice_dcbnl_setpfc() 341 if (pf->lag && pf->lag->bonded) { in ice_dcbnl_set_pfc_cfg() 402 if (pf->lag && pf->lag->bonded) { in ice_dcbnl_setstate() 479 if (pf->lag && pf->lag->bonded) { in ice_dcbnl_set_pg_tc_cfg_tx() 538 if (pf->lag && pf->lag->bonded) { in ice_dcbnl_set_pg_bwg_cfg_tx() 763 if (pf->lag && pf->lag->bonded) { in ice_dcbnl_setapp() 879 if (pf->lag && pf->lag->bonded) { in ice_dcbnl_delapp() 985 if (pf->lag && pf->lag->bonded) { in ice_dcbnl_cee_set_all()
|
H A D | ice_vf_lib.c | 861 struct ice_lag *lag; in ice_reset_vf() local 886 lag = pf->lag; in ice_reset_vf() 888 if (lag && lag->bonded && lag->primary) { in ice_reset_vf() 889 act_prt = lag->active_port; in ice_reset_vf() 891 lag->upper_netdev) in ice_reset_vf() 892 ice_lag_move_vf_nodes_cfg(lag, act_prt, pri_prt); in ice_reset_vf() 981 if (lag && lag->bonded && lag->primary && in ice_reset_vf() 983 ice_lag_move_vf_nodes_cfg(lag, pri_prt, act_prt); in ice_reset_vf()
|
H A D | ice_lag.h | 57 struct ice_lag *lag; member 72 void ice_lag_move_vf_nodes_cfg(struct ice_lag *lag, u8 src_prt, u8 dst_prt);
|
H A D | ice_virtchnl.c | 1974 struct ice_lag *lag; in ice_vc_cfg_qs_msg() local 1979 lag = pf->lag; in ice_vc_cfg_qs_msg() 1983 if (lag && lag->bonded && lag->primary) { in ice_vc_cfg_qs_msg() 1984 act_prt = lag->active_port; in ice_vc_cfg_qs_msg() 1986 lag->upper_netdev) in ice_vc_cfg_qs_msg() 1987 ice_lag_move_vf_nodes_cfg(lag, act_prt, pri_prt); in ice_vc_cfg_qs_msg() 2113 if (lag && lag->bonded && lag->primary && in ice_vc_cfg_qs_msg() 2115 ice_lag_move_vf_nodes_cfg(lag, pri_prt, act_prt); in ice_vc_cfg_qs_msg() 2132 if (lag && lag->bonded && lag->primary && in ice_vc_cfg_qs_msg() 2134 ice_lag_move_vf_nodes_cfg(lag, pri_prt, act_prt); in ice_vc_cfg_qs_msg()
|
/linux/drivers/net/ethernet/netronome/nfp/flower/ |
H A D | lag_conf.c | 86 static unsigned int nfp_fl_get_next_pkt_number(struct nfp_fl_lag *lag) in nfp_fl_get_next_pkt_number() argument 88 lag->pkt_num++; in nfp_fl_get_next_pkt_number() 89 lag->pkt_num &= NFP_FL_LAG_PKT_NUMBER_MASK; in nfp_fl_get_next_pkt_number() 91 return lag->pkt_num; in nfp_fl_get_next_pkt_number() 94 static void nfp_fl_increment_version(struct nfp_fl_lag *lag) in nfp_fl_increment_version() argument 97 lag->batch_ver += 2; in nfp_fl_increment_version() 98 lag->batch_ver &= NFP_FL_LAG_VERSION_MASK; in nfp_fl_increment_version() 101 if (!lag->batch_ver) in nfp_fl_increment_version() 102 lag->batch_ver += 2; in nfp_fl_increment_version() 106 nfp_fl_lag_group_create(struct nfp_fl_lag *lag, struct net_device *master) in nfp_fl_lag_group_create() argument [all …]
|
H A D | main.h | 155 struct nfp_tun_neigh_lag lag; member 171 struct nfp_tun_neigh_lag lag; member 655 void nfp_flower_lag_init(struct nfp_fl_lag *lag); 656 void nfp_flower_lag_cleanup(struct nfp_fl_lag *lag); 657 int nfp_flower_lag_reset(struct nfp_fl_lag *lag); 670 struct nfp_tun_neigh_lag *lag);
|
H A D | tunnel_conf.c | 494 struct nfp_tun_neigh_lag *lag; in nfp_tun_write_neigh() local 515 lag = &payload->lag; in nfp_tun_write_neigh() 526 lag = &payload->lag; in nfp_tun_write_neigh() 536 memcpy(lag, &lag_info, sizeof(struct nfp_tun_neigh_lag)); in nfp_tun_write_neigh()
|
/linux/drivers/net/ethernet/microchip/lan966x/ |
H A D | lan966x_lag.c | 10 int p, lag, i; in lan966x_lag_set_aggr_pgids() local 39 for (lag = 0; lag < lan966x->num_phys_ports; ++lag) { in lan966x_lag_set_aggr_pgids() 40 struct lan966x_port *port = lan966x->ports[lag]; in lan966x_lag_set_aggr_pgids() 46 if (!port || !port->bond || (visited & BIT(lag))) in lan966x_lag_set_aggr_pgids() 81 for (p = lag; p < lan966x->num_phys_ports; p++) { in lan966x_lag_set_aggr_pgids() 267 struct netdev_lag_lower_state_info *lag = info->lower_state_info; in lan966x_lag_port_changelowerstate() local 275 is_active = lag->link_up && lag->tx_enabled; in lan966x_lag_port_changelowerstate() 333 bool lan966x_lag_first_port(struct net_device *lag, struct net_device *dev) in lan966x_lag_first_port() argument 339 if (port->bond != lag) in lan966x_lag_first_port() 342 bond_mask = lan966x_lag_get_mask(lan966x, lag); in lan966x_lag_first_port()
|
H A D | lan966x_mac.c | 25 bool lag; member 217 mac_entry->lag = port->bond ? true : false; in lan966x_mac_alloc_entry() 346 mac_entry->lag) { in lan966x_mac_lag_replace_port_entry() 369 mac_entry->lag) { in lan966x_mac_lag_remove_port_entry()
|
/linux/Documentation/scheduler/ |
H A D | sched-eevdf.rst | 15 time to each task, creating a "lag" value that can be used to determine 17 with a positive lag is owed CPU time, while a negative lag means the task 18 has exceeded its portion. EEVDF picks tasks with lag greater or equal to 24 There are ongoing discussions on how to manage lag, especially for sleeping 27 by sleeping briefly to reset their negative lag: when a task sleeps, it 29 lag to decay over VRT. Hence, long-sleeping tasks eventually have their lag
|
/linux/net/dsa/ |
H A D | switch.c | 354 static int dsa_switch_do_lag_fdb_add(struct dsa_switch *ds, struct dsa_lag *lag, in dsa_switch_do_lag_fdb_add() argument 361 mutex_lock(&lag->fdb_lock); in dsa_switch_do_lag_fdb_add() 363 a = dsa_mac_addr_find(&lag->fdbs, addr, vid, db); in dsa_switch_do_lag_fdb_add() 366 trace_dsa_lag_fdb_add_bump(lag->dev, addr, vid, &db, in dsa_switch_do_lag_fdb_add() 377 err = ds->ops->lag_fdb_add(ds, *lag, addr, vid, db); in dsa_switch_do_lag_fdb_add() 378 trace_dsa_lag_fdb_add_hw(lag->dev, addr, vid, &db, err); in dsa_switch_do_lag_fdb_add() 388 list_add_tail(&a->list, &lag->fdbs); in dsa_switch_do_lag_fdb_add() 391 mutex_unlock(&lag->fdb_lock); in dsa_switch_do_lag_fdb_add() 396 static int dsa_switch_do_lag_fdb_del(struct dsa_switch *ds, struct dsa_lag *lag, in dsa_switch_do_lag_fdb_del() argument 403 mutex_lock(&lag->fdb_lock); in dsa_switch_do_lag_fdb_del() [all …]
|
H A D | tag_dsa.c | 261 struct dsa_lag *lag; in dsa_rcv_ll() local 267 lag = dsa_lag_by_id(cpu_dp->dst, source_port + 1); in dsa_rcv_ll() 268 skb->dev = lag ? lag->dev : NULL; in dsa_rcv_ll()
|
H A D | trace.c | 15 sprintf(buf, "lag %s id %d", db->lag.dev->name, db->lag.id); in dsa_db_print()
|
H A D | dsa.h | 20 void dsa_lag_map(struct dsa_switch_tree *dst, struct dsa_lag *lag); 21 void dsa_lag_unmap(struct dsa_switch_tree *dst, struct dsa_lag *lag);
|
H A D | switch.h | 63 struct dsa_lag *lag; member 79 struct dsa_lag lag; member
|
/linux/drivers/net/ethernet/mellanox/mlxsw/ |
H A D | pci_hw.h | 143 MLXSW_ITEM32(pci, cqe0, lag, 0x00, 23, 1); 144 MLXSW_ITEM32(pci, cqe12, lag, 0x00, 24, 1); 145 mlxsw_pci_cqe_item_helpers(lag, 0, 12, 12);
|
/linux/drivers/net/ethernet/mellanox/mlx5/core/lag/ |
H A D | mpesw.h | 22 struct mlx5_lag *lag; member
|
H A D | mpesw.c | 136 struct mlx5_lag *ldev = mpesww->lag; in mlx5_mpesw_work() 176 work->lag = ldev; in mlx5_lag_mpesw_queue_work()
|
H A D | lag.h | 80 return dev->priv.lag; in mlx5_lag_dev()
|
/linux/drivers/net/ethernet/mscc/ |
H A D | ocelot.c | 2275 int i, port, lag; in ocelot_set_aggr_pgids() local 2303 for (lag = 0; lag < ocelot->num_phys_ports; lag++) { in ocelot_set_aggr_pgids() 2304 struct net_device *bond = ocelot->ports[lag]->bond; in ocelot_set_aggr_pgids() 2309 if (!bond || (visited & BIT(lag))) in ocelot_set_aggr_pgids() 2341 for (port = lag; port < ocelot->num_phys_ports; port++) { in ocelot_set_aggr_pgids() 2371 int lag = ocelot_bond_get_id(ocelot, bond); in ocelot_setup_logical_port_ids() local 2374 ANA_PORT_PORT_CFG_PORTID_VAL(lag), in ocelot_setup_logical_port_ids() 2459 int lag) in ocelot_migrate_lag_fdbs() argument 2477 err = ocelot_mact_learn(ocelot, lag, fdb->addr, fdb->vid, in ocelot_migrate_lag_fdbs() 2556 int lag, err; in ocelot_lag_fdb_add() local [all …]
|
/linux/drivers/infiniband/core/ |
H A D | Makefile | 15 trace.o lag.o
|
/linux/drivers/net/dsa/mv88e6xxx/ |
H A D | chip.c | 1770 if (dp && dp->lag) { in mv88e6xxx_pvt_map() 1815 if (dsa_to_port(chip->ds, port)->lag) in mv88e6xxx_port_fast_age_fid() 6828 struct dsa_lag lag, in mv88e6xxx_lag_can_offload() argument 6841 if (!lag.id) in mv88e6xxx_lag_can_offload() 6844 dsa_lag_foreach_port(dp, ds->dst, &lag) in mv88e6xxx_lag_can_offload() 6870 static int mv88e6xxx_lag_sync_map(struct dsa_switch *ds, struct dsa_lag lag) in mv88e6xxx_lag_sync_map() argument 6878 id = lag.id - 1; in mv88e6xxx_lag_sync_map() 6884 dsa_lag_foreach_port(dp, ds->dst, &lag) in mv88e6xxx_lag_sync_map() 6930 struct dsa_lag *lag; in mv88e6xxx_lag_sync_masks() local 6940 if (!dp->lag) in mv88e6xxx_lag_sync_masks() [all …]
|
/linux/drivers/net/ethernet/marvell/prestera/ |
H A D | prestera_switchdev.c | 893 if (pr_port->lag->lag_id == lag_id) { in prestera_br_port_lag_mdb_mc_enable_sync() 1775 struct prestera_lag *lag; in prestera_fdb_event() local 1784 lag = prestera_lag_by_id(sw, evt->fdb_evt.dest.lag_id); in prestera_fdb_event() 1785 if (lag) in prestera_fdb_event() 1786 dev = lag->dev; in prestera_fdb_event()
|
H A D | prestera.h | 136 struct prestera_lag *lag; member
|