Home
last modified time | relevance | path

Searched refs:ctx (Results 1 – 25 of 133) sorted by relevance

123456

/net/sunrpc/auth_gss/
Dgss_krb5_mech.c150 struct krb5_ctx *ctx, struct crypto_sync_skcipher **res) in get_key() argument
178 *res = crypto_alloc_sync_skcipher(ctx->gk5e->encrypt_name, 0, 0); in get_key()
181 "crypto algorithm %s\n", ctx->gk5e->encrypt_name); in get_key()
187 "crypto algorithm %s\n", ctx->gk5e->encrypt_name); in get_key()
204 gss_import_v1_context(const void *p, const void *end, struct krb5_ctx *ctx) in gss_import_v1_context() argument
210 p = simple_get_bytes(p, end, &ctx->initiate, sizeof(ctx->initiate)); in gss_import_v1_context()
215 ctx->enctype = ENCTYPE_DES_CBC_RAW; in gss_import_v1_context()
217 ctx->gk5e = get_gss_krb5_enctype(ctx->enctype); in gss_import_v1_context()
218 if (ctx->gk5e == NULL) { in gss_import_v1_context()
250 ctx->endtime = (time64_t)time32; in gss_import_v1_context()
[all …]
Dgss_krb5_seal.c73 setup_token(struct krb5_ctx *ctx, struct xdr_netobj *token) in setup_token() argument
77 int body_size = GSS_KRB5_TOK_HDR_LEN + ctx->gk5e->cksumlength; in setup_token()
79 token->len = g_token_size(&ctx->mech_used, body_size); in setup_token()
82 g_make_token_header(&ctx->mech_used, body_size, (unsigned char **)&ptr); in setup_token()
91 *ptr++ = (__force u16)cpu_to_le16(ctx->gk5e->signalg); in setup_token()
99 setup_token_v2(struct krb5_ctx *ctx, struct xdr_netobj *token) in setup_token_v2() argument
105 if ((ctx->flags & KRB5_CTX_FLAG_INITIATOR) == 0) in setup_token_v2()
107 if (ctx->flags & KRB5_CTX_FLAG_ACCEPTOR_SUBKEY) in setup_token_v2()
122 token->len = GSS_KRB5_TOK_HDR_LEN + ctx->gk5e->cksumlength; in setup_token_v2()
127 gss_get_mic_v1(struct krb5_ctx *ctx, struct xdr_buf *text, in gss_get_mic_v1() argument
[all …]
Dgss_krb5_unseal.c74 gss_verify_mic_v1(struct krb5_ctx *ctx, in gss_verify_mic_v1() argument
91 if (g_verify_token_header(&ctx->mech_used, &bodysize, &ptr, in gss_verify_mic_v1()
102 if (signalg != ctx->gk5e->signalg) in gss_verify_mic_v1()
112 if (ctx->gk5e->keyed_cksum) in gss_verify_mic_v1()
113 cksumkey = ctx->cksum; in gss_verify_mic_v1()
117 if (make_checksum(ctx, ptr, 8, message_buffer, 0, in gss_verify_mic_v1()
122 ctx->gk5e->cksumlength)) in gss_verify_mic_v1()
129 if (now > ctx->endtime) in gss_verify_mic_v1()
134 if (krb5_get_seq_num(ctx, ptr + GSS_KRB5_TOK_HDR_LEN, ptr + 8, in gss_verify_mic_v1()
138 if ((ctx->initiate && direction != 0xff) || in gss_verify_mic_v1()
[all …]
Dauth_gss.c97 gss_get_ctx(struct gss_cl_ctx *ctx) in gss_get_ctx() argument
99 refcount_inc(&ctx->count); in gss_get_ctx()
100 return ctx; in gss_get_ctx()
104 gss_put_ctx(struct gss_cl_ctx *ctx) in gss_put_ctx() argument
106 if (refcount_dec_and_test(&ctx->count)) in gss_put_ctx()
107 gss_free_ctx(ctx); in gss_put_ctx()
116 gss_cred_set_ctx(struct rpc_cred *cred, struct gss_cl_ctx *ctx) in gss_cred_set_ctx() argument
122 gss_get_ctx(ctx); in gss_cred_set_ctx()
123 rcu_assign_pointer(gss_cred->gc_ctx, ctx); in gss_cred_set_ctx()
133 struct gss_cl_ctx *ctx = NULL; in gss_cred_get_ctx() local
[all …]
/net/mac80211/
Dchan.c16 struct ieee80211_chanctx *ctx) in ieee80211_chanctx_num_assigned() argument
23 list_for_each_entry(sdata, &ctx->assigned_vifs, assigned_chanctx_list) in ieee80211_chanctx_num_assigned()
30 struct ieee80211_chanctx *ctx) in ieee80211_chanctx_num_reserved() argument
37 list_for_each_entry(sdata, &ctx->reserved_vifs, reserved_chanctx_list) in ieee80211_chanctx_num_reserved()
44 struct ieee80211_chanctx *ctx) in ieee80211_chanctx_refcount() argument
46 return ieee80211_chanctx_num_assigned(local, ctx) + in ieee80211_chanctx_refcount()
47 ieee80211_chanctx_num_reserved(local, ctx); in ieee80211_chanctx_refcount()
52 struct ieee80211_chanctx *ctx; in ieee80211_num_chanctx() local
57 list_for_each_entry(ctx, &local->chanctx_list, list) in ieee80211_num_chanctx()
85 struct ieee80211_chanctx *ctx, in ieee80211_chanctx_reserved_chandef() argument
[all …]
Dtkip.c82 static void tkip_mixing_phase1(const u8 *tk, struct tkip_ctx *ctx, in tkip_mixing_phase1() argument
86 u16 *p1k = ctx->p1k; in tkip_mixing_phase1()
102 ctx->state = TKIP_STATE_PHASE1_DONE; in tkip_mixing_phase1()
103 ctx->p1k_iv32 = tsc_IV32; in tkip_mixing_phase1()
106 static void tkip_mixing_phase2(const u8 *tk, struct tkip_ctx *ctx, in tkip_mixing_phase2() argument
110 const u16 *p1k = ctx->p1k; in tkip_mixing_phase2()
155 struct tkip_ctx *ctx = &key->u.tkip.tx; in ieee80211_compute_tkip_p1k() local
167 if (ctx->p1k_iv32 != iv32 || ctx->state == TKIP_STATE_NOT_INIT) in ieee80211_compute_tkip_p1k()
168 tkip_mixing_phase1(tk, ctx, sdata->vif.addr, iv32); in ieee80211_compute_tkip_p1k()
176 struct tkip_ctx *ctx = &key->u.tkip.tx; in ieee80211_get_tkip_p1k_iv() local
[all …]
/net/tls/
Dtls_main.c68 void update_sk_prot(struct sock *sk, struct tls_context *ctx) in update_sk_prot() argument
73 &tls_prots[ip_ver][ctx->tx_conf][ctx->rx_conf]); in update_sk_prot()
75 &tls_proto_ops[ip_ver][ctx->tx_conf][ctx->rx_conf]); in update_sk_prot()
104 struct tls_context *ctx, in tls_push_sg() argument
118 ctx->in_tcp_sendpages = true; in tls_push_sg()
137 ctx->partially_sent_offset = offset; in tls_push_sg()
138 ctx->partially_sent_record = (void *)sg; in tls_push_sg()
139 ctx->in_tcp_sendpages = false; in tls_push_sg()
153 ctx->in_tcp_sendpages = false; in tls_push_sg()
160 struct tls_context *ctx = tls_get_ctx(sk); in tls_handle_open_record() local
[all …]
Dtls_device.c56 static void tls_device_free_ctx(struct tls_context *ctx) in tls_device_free_ctx() argument
58 if (ctx->tx_conf == TLS_HW) { in tls_device_free_ctx()
59 kfree(tls_offload_ctx_tx(ctx)); in tls_device_free_ctx()
60 kfree(ctx->tx.rec_seq); in tls_device_free_ctx()
61 kfree(ctx->tx.iv); in tls_device_free_ctx()
64 if (ctx->rx_conf == TLS_HW) in tls_device_free_ctx()
65 kfree(tls_offload_ctx_rx(ctx)); in tls_device_free_ctx()
67 tls_ctx_free(NULL, ctx); in tls_device_free_ctx()
74 struct tls_context *ctx = offload_ctx->ctx; in tls_device_tx_del_task() local
75 struct net_device *netdev = ctx->netdev; in tls_device_tx_del_task()
[all …]
Dtls_sw.c131 static int padding_length(struct tls_sw_context_rx *ctx, in padding_length() argument
156 ctx->control = content_type; in padding_length()
166 struct tls_sw_context_rx *ctx; in tls_decrypt_done() local
176 ctx = tls_sw_ctx_rx(tls_ctx); in tls_decrypt_done()
184 ctx->async_wait.err = err; in tls_decrypt_done()
190 pad = padding_length(ctx, prot, skb); in tls_decrypt_done()
192 ctx->async_wait.err = pad; in tls_decrypt_done()
219 spin_lock_bh(&ctx->decrypt_compl_lock); in tls_decrypt_done()
220 pending = atomic_dec_return(&ctx->decrypt_pending); in tls_decrypt_done()
222 if (!pending && ctx->async_notify) in tls_decrypt_done()
[all …]
Dtls_toe.c47 struct tls_context *ctx = tls_get_ctx(sk); in tls_toe_sk_destruct() local
49 ctx->sk_destruct(sk); in tls_toe_sk_destruct()
52 tls_ctx_free(sk, ctx); in tls_toe_sk_destruct()
58 struct tls_context *ctx; in tls_toe_bypass() local
64 ctx = tls_ctx_create(sk); in tls_toe_bypass()
65 if (!ctx) in tls_toe_bypass()
68 ctx->sk_destruct = sk->sk_destruct; in tls_toe_bypass()
70 ctx->rx_conf = TLS_HW_RECORD; in tls_toe_bypass()
71 ctx->tx_conf = TLS_HW_RECORD; in tls_toe_bypass()
72 update_sk_prot(sk, ctx); in tls_toe_bypass()
[all …]
/net/6lowpan/
Ddebugfs.c19 struct lowpan_iphc_ctx *ctx = data; in lowpan_ctx_flag_active_set() local
25 set_bit(LOWPAN_IPHC_CTX_FLAG_ACTIVE, &ctx->flags); in lowpan_ctx_flag_active_set()
27 clear_bit(LOWPAN_IPHC_CTX_FLAG_ACTIVE, &ctx->flags); in lowpan_ctx_flag_active_set()
44 struct lowpan_iphc_ctx *ctx = data; in lowpan_ctx_flag_c_set() local
50 set_bit(LOWPAN_IPHC_CTX_FLAG_COMPRESSION, &ctx->flags); in lowpan_ctx_flag_c_set()
52 clear_bit(LOWPAN_IPHC_CTX_FLAG_COMPRESSION, &ctx->flags); in lowpan_ctx_flag_c_set()
68 struct lowpan_iphc_ctx *ctx = data; in lowpan_ctx_plen_set() local
70 container_of(ctx, struct lowpan_iphc_ctx_table, table[ctx->id]); in lowpan_ctx_plen_set()
76 ctx->plen = val; in lowpan_ctx_plen_set()
84 struct lowpan_iphc_ctx *ctx = data; in lowpan_ctx_plen_get() local
[all …]
/net/xfrm/
Despintcp.c13 static void handle_nonesp(struct espintcp_ctx *ctx, struct sk_buff *skb, in handle_nonesp() argument
26 skb_queue_tail(&ctx->ike_queue, skb); in handle_nonesp()
27 ctx->saved_data_ready(sk); in handle_nonesp()
54 struct espintcp_ctx *ctx = container_of(strp, struct espintcp_ctx, in espintcp_rcv() local
107 handle_nonesp(ctx, skb, strp->sk); in espintcp_rcv()
136 struct espintcp_ctx *ctx = espintcp_getctx(sk); in espintcp_recvmsg() local
144 skb = __skb_recv_datagram(sk, &ctx->ike_queue, flags, &off, &err); in espintcp_recvmsg()
171 struct espintcp_ctx *ctx = espintcp_getctx(sk); in espintcp_queue_out() local
173 if (skb_queue_len(&ctx->out_queue) >= READ_ONCE(netdev_max_backlog)) in espintcp_queue_out()
176 __skb_queue_tail(&ctx->out_queue, skb); in espintcp_queue_out()
[all …]
/net/netfilter/
Dnf_tables_api.c131 static void nft_ctx_init(struct nft_ctx *ctx, in nft_ctx_init() argument
140 ctx->net = net; in nft_ctx_init()
141 ctx->family = family; in nft_ctx_init()
142 ctx->level = 0; in nft_ctx_init()
143 ctx->table = table; in nft_ctx_init()
144 ctx->chain = chain; in nft_ctx_init()
145 ctx->nla = nla; in nft_ctx_init()
146 ctx->portid = NETLINK_CB(skb).portid; in nft_ctx_init()
147 ctx->report = nlmsg_report(nlh); in nft_ctx_init()
148 ctx->flags = nlh->nlmsg_flags; in nft_ctx_init()
[all …]
Dnfnetlink_osf.c67 struct nf_osf_hdr_ctx *ctx) in nf_osf_match_one() argument
69 const __u8 *optpinit = ctx->optp; in nf_osf_match_one()
75 if (ctx->totlen != f->ss || !nf_osf_ttl(skb, ttl_check, f->ttl)) in nf_osf_match_one()
91 ctx->optsize > MAX_IPOPTLEN || in nf_osf_match_one()
92 ctx->optsize != foptsize) in nf_osf_match_one()
98 if (f->opt[optnum].kind == *ctx->optp) { in nf_osf_match_one()
100 const __u8 *optend = ctx->optp + len; in nf_osf_match_one()
104 switch (*ctx->optp) { in nf_osf_match_one()
106 mss = ctx->optp[3]; in nf_osf_match_one()
108 mss |= ctx->optp[2]; in nf_osf_match_one()
[all …]
Dnft_synproxy.c152 static int nft_synproxy_do_init(const struct nft_ctx *ctx, in nft_synproxy_do_init() argument
156 struct synproxy_net *snet = synproxy_pernet(ctx->net); in nft_synproxy_do_init()
171 err = nf_ct_netns_get(ctx->net, ctx->family); in nft_synproxy_do_init()
175 switch (ctx->family) { in nft_synproxy_do_init()
177 err = nf_synproxy_ipv4_init(snet, ctx->net); in nft_synproxy_do_init()
183 err = nf_synproxy_ipv6_init(snet, ctx->net); in nft_synproxy_do_init()
189 err = nf_synproxy_ipv4_init(snet, ctx->net); in nft_synproxy_do_init()
192 err = nf_synproxy_ipv6_init(snet, ctx->net); in nft_synproxy_do_init()
194 nf_synproxy_ipv4_fini(snet, ctx->net); in nft_synproxy_do_init()
203 nf_ct_netns_put(ctx->net, ctx->family); in nft_synproxy_do_init()
[all …]
Dnft_compat.c31 static int nft_compat_chain_validate_dependency(const struct nft_ctx *ctx, in nft_compat_chain_validate_dependency() argument
35 const struct nft_chain *chain = ctx->chain; in nft_compat_chain_validate_dependency()
44 if (ctx->family != NFPROTO_BRIDGE) in nft_compat_chain_validate_dependency()
144 const struct nft_ctx *ctx, in nft_target_set_tgchk_param() argument
148 par->net = ctx->net; in nft_target_set_tgchk_param()
149 par->table = ctx->table->name; in nft_target_set_tgchk_param()
150 switch (ctx->family) { in nft_target_set_tgchk_param()
172 if (nft_is_base_chain(ctx->chain)) { in nft_target_set_tgchk_param()
174 nft_base_chain(ctx->chain); in nft_target_set_tgchk_param()
181 par->family = ctx->family; in nft_target_set_tgchk_param()
[all …]
Dnft_immediate.c46 static int nft_immediate_init(const struct nft_ctx *ctx, in nft_immediate_init() argument
61 err = nft_data_init(ctx, &priv->data, &desc, tb[NFTA_IMMEDIATE_DATA]); in nft_immediate_init()
67 err = nft_parse_register_store(ctx, tb[NFTA_IMMEDIATE_DREG], in nft_immediate_init()
79 err = nf_tables_bind_chain(ctx, chain); in nft_immediate_init()
95 static void nft_immediate_activate(const struct nft_ctx *ctx, in nft_immediate_activate() argument
112 chain_ctx = *ctx; in nft_immediate_activate()
118 nft_clear(ctx->net, chain); in nft_immediate_activate()
128 static void nft_immediate_chain_deactivate(const struct nft_ctx *ctx, in nft_immediate_chain_deactivate() argument
135 chain_ctx = *ctx; in nft_immediate_chain_deactivate()
142 static void nft_immediate_deactivate(const struct nft_ctx *ctx, in nft_immediate_deactivate() argument
[all …]
Dnf_tables_offload.c53 static void nft_flow_rule_transfer_vlan(struct nft_offload_ctx *ctx, in nft_flow_rule_transfer_vlan() argument
90 struct nft_offload_ctx *ctx; in nft_flow_rule_create() local
113 ctx = kzalloc(sizeof(struct nft_offload_ctx), GFP_KERNEL); in nft_flow_rule_create()
114 if (!ctx) { in nft_flow_rule_create()
118 ctx->net = net; in nft_flow_rule_create()
119 ctx->dep.type = NFT_OFFLOAD_DEP_UNSPEC; in nft_flow_rule_create()
126 err = expr->ops->offload(ctx, flow, expr); in nft_flow_rule_create()
132 nft_flow_rule_transfer_vlan(ctx, flow); in nft_flow_rule_create()
134 flow->proto = ctx->dep.l3num; in nft_flow_rule_create()
135 kfree(ctx); in nft_flow_rule_create()
[all …]
Dnft_objref.c25 static int nft_objref_init(const struct nft_ctx *ctx, in nft_objref_init() argument
30 u8 genmask = nft_genmask_next(ctx->net); in nft_objref_init()
38 obj = nft_obj_lookup(ctx->net, ctx->table, in nft_objref_init()
67 static void nft_objref_deactivate(const struct nft_ctx *ctx, in nft_objref_deactivate() argument
79 static void nft_objref_activate(const struct nft_ctx *ctx, in nft_objref_activate() argument
127 static int nft_objref_map_init(const struct nft_ctx *ctx, in nft_objref_map_init() argument
132 u8 genmask = nft_genmask_next(ctx->net); in nft_objref_map_init()
136 set = nft_set_lookup_global(ctx->net, ctx->table, in nft_objref_map_init()
152 err = nf_tables_bind_set(ctx, set, &priv->binding); in nft_objref_map_init()
174 static void nft_objref_map_deactivate(const struct nft_ctx *ctx, in nft_objref_map_deactivate() argument
[all …]
/net/ipv4/netfilter/
Dnf_nat_snmp_basic_main.c67 static void fast_csum(struct snmp_ctx *ctx, unsigned char offset) in fast_csum() argument
73 memcpy(&s[1], &ctx->from, 4); in fast_csum()
74 memcpy(&s[7], &ctx->to, 4); in fast_csum()
83 memcpy(&s[0], &ctx->from, 4); in fast_csum()
84 memcpy(&s[4], &ctx->to, 4); in fast_csum()
91 *ctx->check = csum_fold(csum_partial(s, size, in fast_csum()
92 ~csum_unfold(*ctx->check))); in fast_csum()
108 struct snmp_ctx *ctx = (struct snmp_ctx *)context; in snmp_helper() local
114 if (*pdata == ctx->from) { in snmp_helper()
116 (void *)&ctx->from, (void *)&ctx->to); in snmp_helper()
[all …]
/net/sctp/
Doutqueue.c797 static void sctp_outq_select_transport(struct sctp_flush_ctx *ctx, in sctp_outq_select_transport() argument
813 if (ctx->transport && sctp_cmp_addr_exact(&chunk->dest, in sctp_outq_select_transport()
814 &ctx->transport->ipaddr)) in sctp_outq_select_transport()
815 new_transport = ctx->transport; in sctp_outq_select_transport()
817 new_transport = sctp_assoc_lookup_paddr(ctx->asoc, in sctp_outq_select_transport()
825 new_transport = ctx->asoc->peer.active_path; in sctp_outq_select_transport()
850 new_transport = ctx->asoc->peer.active_path; in sctp_outq_select_transport()
858 if (new_transport != ctx->transport) { in sctp_outq_select_transport()
859 ctx->transport = new_transport; in sctp_outq_select_transport()
860 ctx->packet = &ctx->transport->packet; in sctp_outq_select_transport()
[all …]
/net/bpf/
Dtest_run.c89 static int bpf_test_run(struct bpf_prog *prog, void *ctx, u32 repeat, in bpf_test_run() argument
117 *retval = bpf_prog_run_xdp(prog, ctx); in bpf_test_run()
119 *retval = bpf_prog_run(prog, ctx); in bpf_test_run()
321 void *ctx; member
331 info->retval = bpf_prog_run(info->prog, info->ctx); in __bpf_prog_test_run_raw_tp()
359 info.ctx = kzalloc(ctx_size_in, GFP_USER); in bpf_prog_test_run_raw_tp()
360 if (!info.ctx) in bpf_prog_test_run_raw_tp()
362 if (copy_from_user(info.ctx, ctx_in, ctx_size_in)) { in bpf_prog_test_run_raw_tp()
367 info.ctx = NULL; in bpf_prog_test_run_raw_tp()
394 kfree(info.ctx); in bpf_prog_test_run_raw_tp()
[all …]
/net/ipv4/
Dtcp_fastopen.c38 struct tcp_fastopen_context *ctx = in tcp_fastopen_ctx_free() local
41 kfree_sensitive(ctx); in tcp_fastopen_ctx_free()
46 struct tcp_fastopen_context *ctx; in tcp_fastopen_destroy_cipher() local
48 ctx = rcu_dereference_protected( in tcp_fastopen_destroy_cipher()
49 inet_csk(sk)->icsk_accept_queue.fastopenq.ctx, 1); in tcp_fastopen_destroy_cipher()
50 if (ctx) in tcp_fastopen_destroy_cipher()
51 call_rcu(&ctx->rcu, tcp_fastopen_ctx_free); in tcp_fastopen_destroy_cipher()
67 struct tcp_fastopen_context *ctx, *octx; in tcp_fastopen_reset_cipher() local
71 ctx = kmalloc(sizeof(*ctx), GFP_KERNEL); in tcp_fastopen_reset_cipher()
72 if (!ctx) { in tcp_fastopen_reset_cipher()
[all …]
/net/ethtool/
Dtunnels.c220 struct ethnl_tunnel_info_dump_ctx *ctx = (void *)cb->ctx; in ethnl_tunnel_info_start() local
224 BUILD_BUG_ON(sizeof(*ctx) > sizeof(cb->ctx)); in ethnl_tunnel_info_start()
226 memset(ctx, 0, sizeof(*ctx)); in ethnl_tunnel_info_start()
228 ret = ethnl_parse_header_dev_get(&ctx->req_info, in ethnl_tunnel_info_start()
232 if (ctx->req_info.dev) { in ethnl_tunnel_info_start()
233 dev_put(ctx->req_info.dev); in ethnl_tunnel_info_start()
234 ctx->req_info.dev = NULL; in ethnl_tunnel_info_start()
242 struct ethnl_tunnel_info_dump_ctx *ctx = (void *)cb->ctx; in ethnl_tunnel_info_dumpit() local
244 int s_idx = ctx->pos_idx; in ethnl_tunnel_info_dumpit()
251 for (h = ctx->pos_hash; h < NETDEV_HASHENTRIES; h++, s_idx = 0) { in ethnl_tunnel_info_dumpit()
[all …]
/net/netlink/
Dgenetlink.c564 struct genl_start_context *ctx = cb->data; in genl_start() local
565 const struct genl_ops *ops = ctx->ops; in genl_start()
573 if (ctx->nlh->nlmsg_len < nlmsg_msg_size(ctx->hdrlen)) in genl_start()
576 attrs = genl_family_rcv_msg_attrs_parse(ctx->family, ctx->nlh, ctx->extack, in genl_start()
577 ops, ctx->hdrlen, in genl_start()
588 info->family = ctx->family; in genl_start()
594 if (!ctx->family->parallel_ops) in genl_start()
597 if (!ctx->family->parallel_ops) in genl_start()
656 struct genl_start_context ctx; in genl_family_rcv_msg_dumpit() local
662 ctx.family = family; in genl_family_rcv_msg_dumpit()
[all …]

123456