| /linux/fs/jffs2/ |
| H A D | readinode.c | 31 static int check_node_data(struct jffs2_sb_info *c, struct jffs2_tmp_dnode_info *tn) in check_node_data() argument 33 struct jffs2_raw_node_ref *ref = tn->fn->raw; in check_node_data() 40 BUG_ON(tn->csize == 0); in check_node_data() 44 len = tn->csize; in check_node_data() 51 if (adj >= tn->csize) { in check_node_data() 53 ref_offset(ref), tn->csize, ofs); in check_node_data() 62 ref_offset(ref), tn->csize, tn->partial_crc, tn->data_crc, ofs - len, ofs, len); in check_node_data() 69 JFFS2_WARNING("MTD point returned len too short: %zu instead of %u.\n", retlen, tn->csize); in check_node_data() 99 crc = crc32(tn->partial_crc, buffer, len); in check_node_data() 107 if (crc != tn->data_crc) { in check_node_data() [all …]
|
| H A D | nodelist.h | 354 #define tn_next(tn) rb_entry(rb_next(&(tn)->rb), struct jffs2_tmp_dnode_info, rb) argument 355 #define tn_prev(tn) rb_entry(rb_prev(&(tn)->rb), struct jffs2_tmp_dnode_info, rb) argument 356 #define tn_parent(tn) rb_entry(rb_parent(&(tn)->rb), struct jffs2_tmp_dnode_info, rb) argument 357 #define tn_left(tn) rb_entry((tn)->rb.rb_left, struct jffs2_tmp_dnode_info, rb) argument 358 #define tn_right(tn) rb_entry((tn)->rb.rb_right, struct jffs2_tmp_dnode_info, rb) argument 359 #define tn_erase(tn, list) rb_erase(&tn->rb, list) argument
|
| /linux/net/tipc/ |
| H A D | core.c | 55 struct tipc_net *tn = net_generic(net, tipc_net_id); in tipc_init_net() local 58 tn->net_id = 4711; in tipc_init_net() 59 tn->node_addr = 0; in tipc_init_net() 60 tn->trial_addr = 0; in tipc_init_net() 61 tn->addr_trial_end = 0; in tipc_init_net() 62 tn->capabilities = TIPC_NODE_CAPABILITIES; in tipc_init_net() 63 INIT_WORK(&tn->work, tipc_net_finalize_work); in tipc_init_net() 64 memset(tn->node_id, 0, sizeof(tn->node_id)); in tipc_init_net() 65 memset(tn->node_id_string, 0, sizeof(tn->node_id_string)); in tipc_init_net() 66 tn->mon_threshold = TIPC_DEF_MON_THRESHOLD; in tipc_init_net() [all …]
|
| H A D | discover.c | 84 struct tipc_net *tn = tipc_net(net); in tipc_disc_init_msg() local 89 tipc_msg_init(tn->trial_addr, hdr, LINK_CONFIG, mtyp, in tipc_disc_init_msg() 93 msg_set_node_sig(hdr, tn->random); in tipc_disc_init_msg() 96 msg_set_bc_netid(hdr, tn->net_id); in tipc_disc_init_msg() 98 msg_set_peer_net_hash(hdr, tipc_net_hash_mixes(net, tn->random)); in tipc_disc_init_msg() 150 struct tipc_net *tn = tipc_net(net); in tipc_disc_addr_trial_msg() local 152 bool trial = time_before(jiffies, tn->addr_trial_end) && !self; in tipc_disc_addr_trial_msg() 159 if (dst != tn->trial_addr) in tipc_disc_addr_trial_msg() 163 tn->trial_addr = sugg_addr; in tipc_disc_addr_trial_msg() 165 tn->addr_trial_end = jiffies + msecs_to_jiffies(1000); in tipc_disc_addr_trial_msg() [all …]
|
| H A D | net.c | 128 struct tipc_net *tn = tipc_net(net); in tipc_net_finalize() local 135 if (cmpxchg(&tn->node_addr, 0, addr)) in tipc_net_finalize() 146 struct tipc_net *tn = container_of(work, struct tipc_net, work); in tipc_net_finalize_work() local 149 tipc_net_finalize(tipc_link_net(tn->bcl), tn->trial_addr); in tipc_net_finalize_work() 168 struct tipc_net *tn = net_generic(net, tipc_net_id); in __tipc_nl_add_net() local 169 u64 *w0 = (u64 *)&tn->node_id[0]; in __tipc_nl_add_net() 170 u64 *w1 = (u64 *)&tn->node_id[8]; in __tipc_nl_add_net() 183 if (nla_put_u32(msg->skb, TIPC_NLA_NET_ID, tn->net_id)) in __tipc_nl_add_net() 231 struct tipc_net *tn = tipc_net(net); in __tipc_nl_net_set() local 255 tn->net_id = val; in __tipc_nl_net_set() [all …]
|
| H A D | name_distr.c | 209 struct tipc_net *tn = tipc_net(net); in tipc_named_node_up() local 214 spin_lock_bh(&tn->nametbl_lock); in tipc_named_node_up() 218 spin_unlock_bh(&tn->nametbl_lock); in tipc_named_node_up() 237 struct tipc_net *tn = tipc_net(net); in tipc_publ_purge() local 243 spin_lock_bh(&tn->nametbl_lock); in tipc_publ_purge() 247 spin_unlock_bh(&tn->nametbl_lock); in tipc_publ_purge() 256 struct tipc_net *tn = tipc_net(net); in tipc_publ_notify() local 262 spin_lock_bh(&tn->nametbl_lock); in tipc_publ_notify() 265 spin_unlock_bh(&tn->nametbl_lock); in tipc_publ_notify() 367 struct tipc_net *tn = tipc_net(net); in tipc_named_rcv() local [all …]
|
| H A D | node.c | 332 struct tipc_net *tn = tipc_net(net); in tipc_node_find() local 337 hlist_for_each_entry_rcu(node, &tn->node_htable[thash], hash) { in tipc_node_find() 354 struct tipc_net *tn = tipc_net(net); in tipc_node_find_by_id() local 359 list_for_each_entry_rcu(n, &tn->node_list, list) { in tipc_node_find_by_id() 472 struct tipc_net *tn = net_generic(net, tipc_net_id); in tipc_node_create() local 479 spin_lock_bh(&tn->node_list_lock); in tipc_node_create() 503 &tn->node_htable[tipc_hashfn(addr)]); in tipc_node_create() 505 list_for_each_entry_rcu(temp_node, &tn->node_list, list) { in tipc_node_create() 528 tn->capabilities = TIPC_NODE_CAPABILITIES; in tipc_node_create() 529 list_for_each_entry_rcu(temp_node, &tn->node_list, list) { in tipc_node_create() [all …]
|
| H A D | bearer.c | 64 struct tipc_net *tn = tipc_net(net); in bearer_get() local 66 return rcu_dereference(tn->bearer_list[bearer_id]); in bearer_get() 183 struct tipc_net *tn = tipc_net(net); in tipc_bearer_find() local 188 b = rtnl_dereference(tn->bearer_list[i]); in tipc_bearer_find() 202 struct tipc_net *tn = tipc_net(net); in tipc_bearer_get_name() local 208 b = rtnl_dereference(tn->bearer_list[bearer_id]); in tipc_bearer_get_name() 252 struct tipc_net *tn = tipc_net(net); in tipc_enable_bearer() local 288 b = rtnl_dereference(tn->bearer_list[i]); in tipc_enable_bearer() 364 rcu_assign_pointer(tn->bearer_list[bearer_id], b); in tipc_enable_bearer() 409 struct tipc_net *tn = tipc_net(net); in bearer_disable() local [all …]
|
| H A D | name_table.c | 575 struct tipc_net *tn = tipc_net(net); in tipc_nametbl_lookup_anycast() local 576 bool legacy = tn->legacy_addr_format; in tipc_nametbl_lookup_anycast() 766 struct tipc_net *tn = tipc_net(net); in tipc_nametbl_publish() local 771 spin_lock_bh(&tn->nametbl_lock); in tipc_nametbl_publish() 785 spin_unlock_bh(&tn->nametbl_lock); in tipc_nametbl_publish() 804 struct tipc_net *tn = tipc_net(net); in tipc_nametbl_withdraw() local 809 spin_lock_bh(&tn->nametbl_lock); in tipc_nametbl_withdraw() 819 spin_unlock_bh(&tn->nametbl_lock); in tipc_nametbl_withdraw() 831 struct tipc_net *tn = tipc_net(sub->net); in tipc_nametbl_subscribe() local 839 spin_lock_bh(&tn->nametbl_lock); in tipc_nametbl_subscribe() [all …]
|
| /linux/net/ipv4/ |
| H A D | fib_trie.c | 174 static struct key_vector *resize(struct trie *t, struct key_vector *tn); 195 #define node_parent(tn) rtnl_dereference(tn_info(tn)->parent) argument 196 #define get_child(tn, i) rtnl_dereference((tn)->tnode[i]) argument 199 #define node_parent_rcu(tn) rcu_dereference_rtnl(tn_info(tn)->parent) argument 200 #define get_child_rcu(tn, i) rcu_dereference_rtnl((tn)->tnode[i]) argument 214 static inline unsigned long child_length(const struct key_vector *tn) in child_length() argument 216 return (1ul << tn->bits) & ~(1ul); in child_length() 374 struct key_vector *tn; in tnode_new() local 392 tn = tnode->kv; in tnode_new() 393 tn->key = (shift < KEYLENGTH) ? (key >> shift) << shift : 0; in tnode_new() [all …]
|
| /linux/net/netfilter/ |
| H A D | xt_TEE.c | 65 struct tee_net *tn = net_generic(net, tee_net_id); in tee_netdev_event() local 68 mutex_lock(&tn->lock); in tee_netdev_event() 69 list_for_each_entry(priv, &tn->priv_list, list) { in tee_netdev_event() 87 mutex_unlock(&tn->lock); in tee_netdev_event() 94 struct tee_net *tn = net_generic(par->net, tee_net_id); in tee_tg_check() local 122 mutex_lock(&tn->lock); in tee_tg_check() 123 list_add(&priv->list, &tn->priv_list); in tee_tg_check() 124 mutex_unlock(&tn->lock); in tee_tg_check() 134 struct tee_net *tn = net_generic(par->net, tee_net_id); in tee_tg_destroy() local 138 mutex_lock(&tn->lock); in tee_tg_destroy() [all …]
|
| H A D | nf_conntrack_proto_tcp.c | 491 const struct nf_tcp_net *tn = nf_tcp_pernet(nf_ct_net(ct)); in nf_tcp_log_invalid() local 496 be_liberal = sender->flags & IP_CT_TCP_FLAG_BE_LIBERAL || tn->tcp_be_liberal; in nf_tcp_log_invalid() 729 const struct nf_tcp_net *tn; in nf_tcp_handle_invalid() local 760 tn = nf_tcp_pernet(nf_ct_net(ct)); in nf_tcp_handle_invalid() 763 timeouts = tn->timeouts; in nf_tcp_handle_invalid() 845 const struct nf_tcp_net *tn = nf_tcp_pernet(net); in tcp_new() local 869 } else if (tn->tcp_loose == 0) { in tcp_new() 938 const struct nf_tcp_net *tn; in nf_conntrack_tcp_set_closing() local 940 tn = nf_tcp_pernet(nf_ct_net(ct)); in nf_conntrack_tcp_set_closing() 941 timeouts = tn->timeouts; in nf_conntrack_tcp_set_closing() [all …]
|
| H A D | nf_flow_table_core.c | 206 const struct nf_tcp_net *tn = nf_tcp_pernet(net); in flow_offload_fixup_ct() 216 timeout = READ_ONCE(tn->timeouts[TCP_CONNTRACK_CLOSE]); in flow_offload_fixup_ct() 221 timeout = READ_ONCE(tn->timeouts[tcp_state]); in flow_offload_fixup_ct() 224 offload_timeout = READ_ONCE(tn->offload_timeout); in flow_offload_fixup_ct() 227 const struct nf_udp_net *tn = nf_udp_pernet(net); in flow_offload_fixup_ct() 232 timeout = READ_ONCE(tn->timeouts[state]); in flow_offload_fixup_ct() 234 offload_timeout = READ_ONCE(tn->offload_timeout); in flow_offload_fixup_ct() 311 struct nf_tcp_net *tn = nf_tcp_pernet(net); in flow_offload_get_timeout() local 313 timeout = tn->offload_timeout; in flow_offload_get_timeout() 315 struct nf_udp_net *tn in flow_offload_get_timeout() 202 const struct nf_tcp_net *tn = nf_tcp_pernet(net); flow_offload_fixup_ct() local 223 const struct nf_udp_net *tn = nf_udp_pernet(net); flow_offload_fixup_ct() local 307 struct nf_tcp_net *tn = nf_tcp_pernet(net); flow_offload_get_timeout() local [all...] |
| /linux/drivers/scsi/sym53c8xx_2/ |
| H A D | sym_nvram.c | 84 Symbios_target *tn = &nvram->target[target]; in sym_Symbios_setup_target() local 86 if (!(tn->flags & SYMBIOS_QUEUE_TAGS_ENABLED)) in sym_Symbios_setup_target() 88 if (!(tn->flags & SYMBIOS_DISCONNECT_ENABLE)) in sym_Symbios_setup_target() 90 if (!(tn->flags & SYMBIOS_SCAN_AT_BOOT_TIME)) in sym_Symbios_setup_target() 92 if (!(tn->flags & SYMBIOS_SCAN_LUNS)) in sym_Symbios_setup_target() 94 tp->usr_period = (tn->sync_period + 3) / 4; in sym_Symbios_setup_target() 95 tp->usr_width = (tn->bus_width == 0x8) ? 0 : 1; in sym_Symbios_setup_target() 108 struct Tekram_target *tn = &nvram->target[target]; in sym_Tekram_setup_target() local 110 if (tn->flags & TEKRAM_TAGGED_COMMANDS) { in sym_Tekram_setup_target() 114 if (tn->flags & TEKRAM_DISCONNECT_ENABLE) in sym_Tekram_setup_target() [all …]
|
| /linux/include/net/ |
| H A D | act_api.h | 158 int tc_action_net_init(struct net *net, struct tc_action_net *tn, in tc_action_net_init() argument 163 tn->idrinfo = kmalloc_obj(*tn->idrinfo); in tc_action_net_init() 164 if (!tn->idrinfo) in tc_action_net_init() 166 tn->ops = ops; in tc_action_net_init() 167 tn->idrinfo->net = net; in tc_action_net_init() 168 mutex_init(&tn->idrinfo->lock); in tc_action_net_init() 169 idr_init(&tn->idrinfo->action_idr); in tc_action_net_init() 182 struct tc_action_net *tn = net_generic(net, id); in tc_action_net_exit() local 184 tcf_idrinfo_destroy(tn->ops, tn->idrinfo); in tc_action_net_exit() 185 kfree(tn->idrinfo); in tc_action_net_exit() [all …]
|
| /linux/net/sched/ |
| H A D | act_vlan.c | 119 struct tc_action_net *tn = net_generic(net, act_vlan_ops.net_id); in tcf_vlan_init() local 147 err = tcf_idr_check_alloc(tn, &index, a, bind); in tcf_vlan_init() 163 tcf_idr_cleanup(tn, index); in tcf_vlan_init() 171 tcf_idr_cleanup(tn, index); in tcf_vlan_init() 185 tcf_idr_cleanup(tn, index); in tcf_vlan_init() 203 tcf_idr_cleanup(tn, index); in tcf_vlan_init() 211 tcf_idr_cleanup(tn, index); in tcf_vlan_init() 217 ret = tcf_idr_create_from_flags(tn, index, est, a, in tcf_vlan_init() 220 tcf_idr_cleanup(tn, index); in tcf_vlan_init() 433 struct tc_action_net *tn = net_generic(net, act_vlan_ops.net_id); in vlan_init_net() local [all …]
|
| H A D | act_skbmod.c | 106 struct tc_action_net *tn = net_generic(net, act_skbmod_ops.net_id); in tcf_skbmod_init() local 154 err = tcf_idr_check_alloc(tn, &index, a, bind); in tcf_skbmod_init() 165 tcf_idr_cleanup(tn, index); in tcf_skbmod_init() 170 ret = tcf_idr_create(tn, index, est, a, in tcf_skbmod_init() 173 tcf_idr_cleanup(tn, index); in tcf_skbmod_init() 292 struct tc_action_net *tn = net_generic(net, act_skbmod_ops.net_id); in skbmod_init_net() local 294 return tc_action_net_init(net, tn, &act_skbmod_ops); in skbmod_init_net()
|
| H A D | act_connmark.c | 103 struct tc_action_net *tn = net_generic(net, act_connmark_ops.net_id); in tcf_connmark_init() local 130 ret = tcf_idr_check_alloc(tn, &index, a, bind); in tcf_connmark_init() 132 ret = tcf_idr_create_from_flags(tn, index, est, a, in tcf_connmark_init() 135 tcf_idr_cleanup(tn, index); in tcf_connmark_init() 253 struct tc_action_net *tn = net_generic(net, act_connmark_ops.net_id); in connmark_init_net() local 255 return tc_action_net_init(net, tn, &act_connmark_ops); in connmark_init_net()
|
| H A D | act_gact.c | 58 struct tc_action_net *tn = net_generic(net, act_gact_ops.net_id); in tcf_gact_init() local 100 err = tcf_idr_check_alloc(tn, &index, a, bind); in tcf_gact_init() 102 ret = tcf_idr_create_from_flags(tn, index, est, a, in tcf_gact_init() 105 tcf_idr_cleanup(tn, index); in tcf_gact_init() 303 struct tc_action_net *tn = net_generic(net, act_gact_ops.net_id); in gact_init_net() local 305 return tc_action_net_init(net, tn, &act_gact_ops); in gact_init_net()
|
| H A D | act_sample.c | 41 struct tc_action_net *tn = net_generic(net, act_sample_ops.net_id); in tcf_sample_init() local 64 err = tcf_idr_check_alloc(tn, &index, a, bind); in tcf_sample_init() 72 ret = tcf_idr_create(tn, index, est, a, in tcf_sample_init() 75 tcf_idr_cleanup(tn, index); in tcf_sample_init() 335 struct tc_action_net *tn = net_generic(net, act_sample_ops.net_id); in sample_init_net() local 337 return tc_action_net_init(net, tn, &act_sample_ops); in sample_init_net()
|
| H A D | act_nat.c | 39 struct tc_action_net *tn = net_generic(net, act_nat_ops.net_id); in tcf_nat_init() local 61 err = tcf_idr_check_alloc(tn, &index, a, bind); in tcf_nat_init() 63 ret = tcf_idr_create_from_flags(tn, index, est, a, &act_nat_ops, in tcf_nat_init() 66 tcf_idr_cleanup(tn, index); in tcf_nat_init() 330 struct tc_action_net *tn = net_generic(net, act_nat_ops.net_id); in nat_init_net() local 332 return tc_action_net_init(net, tn, &act_nat_ops); in nat_init_net()
|
| H A D | act_mirred.c | 105 struct tc_action_net *tn = net_generic(net, act_mirred_ops.net_id); in tcf_mirred_init() local 130 err = tcf_idr_check_alloc(tn, &index, a, bind); in tcf_mirred_init() 143 tcf_idr_cleanup(tn, index); in tcf_mirred_init() 158 tcf_idr_cleanup(tn, index); in tcf_mirred_init() 165 tcf_idr_cleanup(tn, index); in tcf_mirred_init() 170 ret = tcf_idr_create_from_flags(tn, index, est, a, in tcf_mirred_init() 173 tcf_idr_cleanup(tn, index); in tcf_mirred_init() 657 struct tc_action_net *tn = net_generic(net, act_mirred_ops.net_id); in mirred_init_net() local 659 return tc_action_net_init(net, tn, &act_mirred_ops); in mirred_init_net()
|
| /linux/arch/powerpc/sysdev/ |
| H A D | mpic_msgr.c | 149 struct device_node *tn; in mpic_msgr_block_number() local 153 tn = of_find_node_by_path(prop->value); in mpic_msgr_block_number() 154 if (node == tn) { in mpic_msgr_block_number() 155 of_node_put(tn); in mpic_msgr_block_number() 158 of_node_put(tn); in mpic_msgr_block_number()
|
| /linux/arch/arm/vfp/ |
| H A D | vfpsingle.c | 219 int tn, tm = 0; in vfp_propagate_nan() local 221 tn = vfp_single_type(vsn); in vfp_propagate_nan() 237 if (tn == VFP_SNAN || (tm != VFP_SNAN && tn == VFP_QNAN)) in vfp_propagate_nan() 252 return tn == VFP_SNAN || tm == VFP_SNAN ? FPSCR_IOC : VFP_NAN_FLAG; in vfp_propagate_nan() 734 int tn, tm; in vfp_single_fadd_nonnumber() local 736 tn = vfp_single_type(vsn); in vfp_single_fadd_nonnumber() 739 if (tn & tm & VFP_INFINITY) { in vfp_single_fadd_nonnumber() 755 } else if (tn & VFP_INFINITY && tm & VFP_NUMBER) { in vfp_single_fadd_nonnumber() 1057 int tm, tn; in vfp_single_fdiv() local 1066 tn = vfp_single_type(&vsn); in vfp_single_fdiv() [all …]
|
| H A D | vfpdouble.c | 211 int tn, tm = 0; in vfp_propagate_nan() local 213 tn = vfp_double_type(vdn); in vfp_propagate_nan() 229 if (tn == VFP_SNAN || (tm != VFP_SNAN && tn == VFP_QNAN)) in vfp_propagate_nan() 244 return tn == VFP_SNAN || tm == VFP_SNAN ? FPSCR_IOC : VFP_NAN_FLAG; in vfp_propagate_nan() 688 int tn, tm; in vfp_double_fadd_nonnumber() local 690 tn = vfp_double_type(vdn); in vfp_double_fadd_nonnumber() 693 if (tn & tm & VFP_INFINITY) { in vfp_double_fadd_nonnumber() 709 } else if (tn & VFP_INFINITY && tm & VFP_NUMBER) { in vfp_double_fadd_nonnumber() 1011 int tm, tn; in vfp_double_fdiv() local 1018 tn = vfp_double_type(&vdn); in vfp_double_fdiv() [all …]
|