Author | Tokens | Token Proportion | Commits | Commit Proportion |
---|---|---|---|---|
Sabrina Dubroca | 2757 | 93.36% | 8 | 20.00% |
David Howells | 66 | 2.24% | 3 | 7.50% |
Linus Torvalds (pre-git) | 20 | 0.68% | 1 | 2.50% |
Herbert Xu | 11 | 0.37% | 2 | 5.00% |
Joy Latten | 11 | 0.37% | 2 | 5.00% |
Hideaki Yoshifuji / 吉藤英明 | 9 | 0.30% | 1 | 2.50% |
Steffen Klassert | 8 | 0.27% | 3 | 7.50% |
David S. Miller | 8 | 0.27% | 2 | 5.00% |
Alexey Dobriyan | 7 | 0.24% | 1 | 2.50% |
Alexey Kuznetsov | 7 | 0.24% | 2 | 5.00% |
Eric Dumazet | 6 | 0.20% | 2 | 5.00% |
Kazunori Miyazawa | 6 | 0.20% | 2 | 5.00% |
Benjamin Coddington | 6 | 0.20% | 1 | 2.50% |
Kunihiro Ishiguro | 5 | 0.17% | 1 | 2.50% |
Rusty Russell | 5 | 0.17% | 1 | 2.50% |
Peilin Ye | 5 | 0.17% | 1 | 2.50% |
Shinta Sugimoto | 4 | 0.14% | 1 | 2.50% |
David Ahern | 3 | 0.10% | 1 | 2.50% |
Kuniyuki Iwashima | 3 | 0.10% | 1 | 2.50% |
Satoru Moriya | 3 | 0.10% | 1 | 2.50% |
Adrian Bunk | 1 | 0.03% | 1 | 2.50% |
Al Viro | 1 | 0.03% | 1 | 2.50% |
Richard Gobert | 1 | 0.03% | 1 | 2.50% |
Total | 2953 | 40 |
// SPDX-License-Identifier: GPL-2.0 #include <net/tcp.h> #include <net/strparser.h> #include <net/xfrm.h> #include <net/esp.h> #include <net/espintcp.h> #include <linux/skmsg.h> #include <net/inet_common.h> #include <trace/events/sock.h> #if IS_ENABLED(CONFIG_IPV6) #include <net/ipv6_stubs.h> #endif #include <net/hotdata.h> static void handle_nonesp(struct espintcp_ctx *ctx, struct sk_buff *skb, struct sock *sk) { if (atomic_read(&sk->sk_rmem_alloc) >= sk->sk_rcvbuf || !sk_rmem_schedule(sk, skb, skb->truesize)) { XFRM_INC_STATS(sock_net(sk), LINUX_MIB_XFRMINERROR); kfree_skb(skb); return; } skb_set_owner_r(skb, sk); memset(skb->cb, 0, sizeof(skb->cb)); skb_queue_tail(&ctx->ike_queue, skb); ctx->saved_data_ready(sk); } static void handle_esp(struct sk_buff *skb, struct sock *sk) { struct tcp_skb_cb *tcp_cb = (struct tcp_skb_cb *)skb->cb; skb_reset_transport_header(skb); /* restore IP CB, we need at least IP6CB->nhoff */ memmove(skb->cb, &tcp_cb->header, sizeof(tcp_cb->header)); rcu_read_lock(); skb->dev = dev_get_by_index_rcu(sock_net(sk), skb->skb_iif); local_bh_disable(); #if IS_ENABLED(CONFIG_IPV6) if (sk->sk_family == AF_INET6) ipv6_stub->xfrm6_rcv_encap(skb, IPPROTO_ESP, 0, TCP_ENCAP_ESPINTCP); else #endif xfrm4_rcv_encap(skb, IPPROTO_ESP, 0, TCP_ENCAP_ESPINTCP); local_bh_enable(); rcu_read_unlock(); } static void espintcp_rcv(struct strparser *strp, struct sk_buff *skb) { struct espintcp_ctx *ctx = container_of(strp, struct espintcp_ctx, strp); struct strp_msg *rxm = strp_msg(skb); int len = rxm->full_len - 2; u32 nonesp_marker; int err; /* keepalive packet? */ if (unlikely(len == 1)) { u8 data; err = skb_copy_bits(skb, rxm->offset + 2, &data, 1); if (err < 0) { XFRM_INC_STATS(sock_net(strp->sk), LINUX_MIB_XFRMINHDRERROR); kfree_skb(skb); return; } if (data == 0xff) { kfree_skb(skb); return; } } /* drop other short messages */ if (unlikely(len <= sizeof(nonesp_marker))) { XFRM_INC_STATS(sock_net(strp->sk), LINUX_MIB_XFRMINHDRERROR); kfree_skb(skb); return; } err = skb_copy_bits(skb, rxm->offset + 2, &nonesp_marker, sizeof(nonesp_marker)); if (err < 0) { XFRM_INC_STATS(sock_net(strp->sk), LINUX_MIB_XFRMINHDRERROR); kfree_skb(skb); return; } /* remove header, leave non-ESP marker/SPI */ if (!pskb_pull(skb, rxm->offset + 2)) { XFRM_INC_STATS(sock_net(strp->sk), LINUX_MIB_XFRMINERROR); kfree_skb(skb); return; } if (pskb_trim(skb, rxm->full_len - 2) != 0) { XFRM_INC_STATS(sock_net(strp->sk), LINUX_MIB_XFRMINERROR); kfree_skb(skb); return; } if (nonesp_marker == 0) handle_nonesp(ctx, skb, strp->sk); else handle_esp(skb, strp->sk); } static int espintcp_parse(struct strparser *strp, struct sk_buff *skb) { struct strp_msg *rxm = strp_msg(skb); __be16 blen; u16 len; int err; if (skb->len < rxm->offset + 2) return 0; err = skb_copy_bits(skb, rxm->offset, &blen, sizeof(blen)); if (err < 0) return err; len = be16_to_cpu(blen); if (len < 2) return -EINVAL; return len; } static int espintcp_recvmsg(struct sock *sk, struct msghdr *msg, size_t len, int flags, int *addr_len) { struct espintcp_ctx *ctx = espintcp_getctx(sk); struct sk_buff *skb; int err = 0; int copied; int off = 0; skb = __skb_recv_datagram(sk, &ctx->ike_queue, flags, &off, &err); if (!skb) { if (err == -EAGAIN && sk->sk_shutdown & RCV_SHUTDOWN) return 0; return err; } copied = len; if (copied > skb->len) copied = skb->len; else if (copied < skb->len) msg->msg_flags |= MSG_TRUNC; err = skb_copy_datagram_msg(skb, 0, msg, copied); if (unlikely(err)) { kfree_skb(skb); return err; } if (flags & MSG_TRUNC) copied = skb->len; kfree_skb(skb); return copied; } int espintcp_queue_out(struct sock *sk, struct sk_buff *skb) { struct espintcp_ctx *ctx = espintcp_getctx(sk); if (skb_queue_len(&ctx->out_queue) >= READ_ONCE(net_hotdata.max_backlog)) return -ENOBUFS; __skb_queue_tail(&ctx->out_queue, skb); return 0; } EXPORT_SYMBOL_GPL(espintcp_queue_out); /* espintcp length field is 2B and length includes the length field's size */ #define MAX_ESPINTCP_MSG (((1 << 16) - 1) - 2) static int espintcp_sendskb_locked(struct sock *sk, struct espintcp_msg *emsg, int flags) { do { int ret; ret = skb_send_sock_locked(sk, emsg->skb, emsg->offset, emsg->len); if (ret < 0) return ret; emsg->len -= ret; emsg->offset += ret; } while (emsg->len > 0); kfree_skb(emsg->skb); memset(emsg, 0, sizeof(*emsg)); return 0; } static int espintcp_sendskmsg_locked(struct sock *sk, struct espintcp_msg *emsg, int flags) { struct msghdr msghdr = { .msg_flags = flags | MSG_SPLICE_PAGES | MSG_MORE, }; struct sk_msg *skmsg = &emsg->skmsg; bool more = flags & MSG_MORE; struct scatterlist *sg; int done = 0; int ret; sg = &skmsg->sg.data[skmsg->sg.start]; do { struct bio_vec bvec; size_t size = sg->length - emsg->offset; int offset = sg->offset + emsg->offset; struct page *p; emsg->offset = 0; if (sg_is_last(sg) && !more) msghdr.msg_flags &= ~MSG_MORE; p = sg_page(sg); retry: bvec_set_page(&bvec, p, size, offset); iov_iter_bvec(&msghdr.msg_iter, ITER_SOURCE, &bvec, 1, size); ret = tcp_sendmsg_locked(sk, &msghdr, size); if (ret < 0) { emsg->offset = offset - sg->offset; skmsg->sg.start += done; return ret; } if (ret != size) { offset += ret; size -= ret; goto retry; } done++; put_page(p); sk_mem_uncharge(sk, sg->length); sg = sg_next(sg); } while (sg); memset(emsg, 0, sizeof(*emsg)); return 0; } static int espintcp_push_msgs(struct sock *sk, int flags) { struct espintcp_ctx *ctx = espintcp_getctx(sk); struct espintcp_msg *emsg = &ctx->partial; int err; if (!emsg->len) return 0; if (ctx->tx_running) return -EAGAIN; ctx->tx_running = 1; if (emsg->skb) err = espintcp_sendskb_locked(sk, emsg, flags); else err = espintcp_sendskmsg_locked(sk, emsg, flags); if (err == -EAGAIN) { ctx->tx_running = 0; return flags & MSG_DONTWAIT ? -EAGAIN : 0; } if (!err) memset(emsg, 0, sizeof(*emsg)); ctx->tx_running = 0; return err; } int espintcp_push_skb(struct sock *sk, struct sk_buff *skb) { struct espintcp_ctx *ctx = espintcp_getctx(sk); struct espintcp_msg *emsg = &ctx->partial; unsigned int len; int offset; if (sk->sk_state != TCP_ESTABLISHED) { kfree_skb(skb); return -ECONNRESET; } offset = skb_transport_offset(skb); len = skb->len - offset; espintcp_push_msgs(sk, 0); if (emsg->len) { kfree_skb(skb); return -ENOBUFS; } skb_set_owner_w(skb, sk); emsg->offset = offset; emsg->len = len; emsg->skb = skb; espintcp_push_msgs(sk, 0); return 0; } EXPORT_SYMBOL_GPL(espintcp_push_skb); static int espintcp_sendmsg(struct sock *sk, struct msghdr *msg, size_t size) { long timeo = sock_sndtimeo(sk, msg->msg_flags & MSG_DONTWAIT); struct espintcp_ctx *ctx = espintcp_getctx(sk); struct espintcp_msg *emsg = &ctx->partial; struct iov_iter pfx_iter; struct kvec pfx_iov = {}; size_t msglen = size + 2; char buf[2] = {0}; int err, end; if (msg->msg_flags & ~MSG_DONTWAIT) return -EOPNOTSUPP; if (size > MAX_ESPINTCP_MSG) return -EMSGSIZE; if (msg->msg_controllen) return -EOPNOTSUPP; lock_sock(sk); err = espintcp_push_msgs(sk, msg->msg_flags & MSG_DONTWAIT); if (err < 0) { if (err != -EAGAIN || !(msg->msg_flags & MSG_DONTWAIT)) err = -ENOBUFS; goto unlock; } sk_msg_init(&emsg->skmsg); while (1) { /* only -ENOMEM is possible since we don't coalesce */ err = sk_msg_alloc(sk, &emsg->skmsg, msglen, 0); if (!err) break; err = sk_stream_wait_memory(sk, &timeo); if (err) goto fail; } *((__be16 *)buf) = cpu_to_be16(msglen); pfx_iov.iov_base = buf; pfx_iov.iov_len = sizeof(buf); iov_iter_kvec(&pfx_iter, ITER_SOURCE, &pfx_iov, 1, pfx_iov.iov_len); err = sk_msg_memcopy_from_iter(sk, &pfx_iter, &emsg->skmsg, pfx_iov.iov_len); if (err < 0) goto fail; err = sk_msg_memcopy_from_iter(sk, &msg->msg_iter, &emsg->skmsg, size); if (err < 0) goto fail; end = emsg->skmsg.sg.end; emsg->len = size; sk_msg_iter_var_prev(end); sg_mark_end(sk_msg_elem(&emsg->skmsg, end)); tcp_rate_check_app_limited(sk); err = espintcp_push_msgs(sk, msg->msg_flags & MSG_DONTWAIT); /* this message could be partially sent, keep it */ release_sock(sk); return size; fail: sk_msg_free(sk, &emsg->skmsg); memset(emsg, 0, sizeof(*emsg)); unlock: release_sock(sk); return err; } static struct proto espintcp_prot __ro_after_init; static struct proto_ops espintcp_ops __ro_after_init; static struct proto espintcp6_prot; static struct proto_ops espintcp6_ops; static DEFINE_MUTEX(tcpv6_prot_mutex); static void espintcp_data_ready(struct sock *sk) { struct espintcp_ctx *ctx = espintcp_getctx(sk); trace_sk_data_ready(sk); strp_data_ready(&ctx->strp); } static void espintcp_tx_work(struct work_struct *work) { struct espintcp_ctx *ctx = container_of(work, struct espintcp_ctx, work); struct sock *sk = ctx->strp.sk; lock_sock(sk); if (!ctx->tx_running) espintcp_push_msgs(sk, 0); release_sock(sk); } static void espintcp_write_space(struct sock *sk) { struct espintcp_ctx *ctx = espintcp_getctx(sk); schedule_work(&ctx->work); ctx->saved_write_space(sk); } static void espintcp_destruct(struct sock *sk) { struct espintcp_ctx *ctx = espintcp_getctx(sk); ctx->saved_destruct(sk); kfree(ctx); } bool tcp_is_ulp_esp(struct sock *sk) { return sk->sk_prot == &espintcp_prot || sk->sk_prot == &espintcp6_prot; } EXPORT_SYMBOL_GPL(tcp_is_ulp_esp); static void build_protos(struct proto *espintcp_prot, struct proto_ops *espintcp_ops, const struct proto *orig_prot, const struct proto_ops *orig_ops); static int espintcp_init_sk(struct sock *sk) { struct inet_connection_sock *icsk = inet_csk(sk); struct strp_callbacks cb = { .rcv_msg = espintcp_rcv, .parse_msg = espintcp_parse, }; struct espintcp_ctx *ctx; int err; /* sockmap is not compatible with espintcp */ if (sk->sk_user_data) return -EBUSY; ctx = kzalloc(sizeof(*ctx), GFP_KERNEL); if (!ctx) return -ENOMEM; err = strp_init(&ctx->strp, sk, &cb); if (err) goto free; __sk_dst_reset(sk); strp_check_rcv(&ctx->strp); skb_queue_head_init(&ctx->ike_queue); skb_queue_head_init(&ctx->out_queue); if (sk->sk_family == AF_INET) { sk->sk_prot = &espintcp_prot; sk->sk_socket->ops = &espintcp_ops; } else { mutex_lock(&tcpv6_prot_mutex); if (!espintcp6_prot.recvmsg) build_protos(&espintcp6_prot, &espintcp6_ops, sk->sk_prot, sk->sk_socket->ops); mutex_unlock(&tcpv6_prot_mutex); sk->sk_prot = &espintcp6_prot; sk->sk_socket->ops = &espintcp6_ops; } ctx->saved_data_ready = sk->sk_data_ready; ctx->saved_write_space = sk->sk_write_space; ctx->saved_destruct = sk->sk_destruct; sk->sk_data_ready = espintcp_data_ready; sk->sk_write_space = espintcp_write_space; sk->sk_destruct = espintcp_destruct; rcu_assign_pointer(icsk->icsk_ulp_data, ctx); INIT_WORK(&ctx->work, espintcp_tx_work); /* avoid using task_frag */ sk->sk_allocation = GFP_ATOMIC; sk->sk_use_task_frag = false; return 0; free: kfree(ctx); return err; } static void espintcp_release(struct sock *sk) { struct espintcp_ctx *ctx = espintcp_getctx(sk); struct sk_buff_head queue; struct sk_buff *skb; __skb_queue_head_init(&queue); skb_queue_splice_init(&ctx->out_queue, &queue); while ((skb = __skb_dequeue(&queue))) espintcp_push_skb(sk, skb); tcp_release_cb(sk); } static void espintcp_close(struct sock *sk, long timeout) { struct espintcp_ctx *ctx = espintcp_getctx(sk); struct espintcp_msg *emsg = &ctx->partial; strp_stop(&ctx->strp); sk->sk_prot = &tcp_prot; barrier(); cancel_work_sync(&ctx->work); strp_done(&ctx->strp); skb_queue_purge(&ctx->out_queue); skb_queue_purge(&ctx->ike_queue); if (emsg->len) { if (emsg->skb) kfree_skb(emsg->skb); else sk_msg_free(sk, &emsg->skmsg); } tcp_close(sk, timeout); } static __poll_t espintcp_poll(struct file *file, struct socket *sock, poll_table *wait) { __poll_t mask = datagram_poll(file, sock, wait); struct sock *sk = sock->sk; struct espintcp_ctx *ctx = espintcp_getctx(sk); if (!skb_queue_empty(&ctx->ike_queue)) mask |= EPOLLIN | EPOLLRDNORM; return mask; } static void build_protos(struct proto *espintcp_prot, struct proto_ops *espintcp_ops, const struct proto *orig_prot, const struct proto_ops *orig_ops) { memcpy(espintcp_prot, orig_prot, sizeof(struct proto)); memcpy(espintcp_ops, orig_ops, sizeof(struct proto_ops)); espintcp_prot->sendmsg = espintcp_sendmsg; espintcp_prot->recvmsg = espintcp_recvmsg; espintcp_prot->close = espintcp_close; espintcp_prot->release_cb = espintcp_release; espintcp_ops->poll = espintcp_poll; } static struct tcp_ulp_ops espintcp_ulp __read_mostly = { .name = "espintcp", .owner = THIS_MODULE, .init = espintcp_init_sk, }; void __init espintcp_init(void) { build_protos(&espintcp_prot, &espintcp_ops, &tcp_prot, &inet_stream_ops); tcp_register_ulp(&espintcp_ulp); }
Information contained on this website is for historical information purposes only and does not indicate or represent copyright ownership.
Created with Cregit http://github.com/cregit/cregit
Version 2.0-RC1