Release 4.12 include/linux/netpoll.h
/*
* Common code for low-level network console, dump, and debugger code
*
* Derived from netconsole, kgdb-over-ethernet, and netdump patches
*/
#ifndef _LINUX_NETPOLL_H
#define _LINUX_NETPOLL_H
#include <linux/netdevice.h>
#include <linux/interrupt.h>
#include <linux/rcupdate.h>
#include <linux/list.h>
union inet_addr {
__u32 all[4];
__be32 ip;
__be32 ip6[4];
struct in_addr in;
struct in6_addr in6;
};
struct netpoll {
struct net_device *dev;
char dev_name[IFNAMSIZ];
const char *name;
union inet_addr local_ip, remote_ip;
bool ipv6;
u16 local_port, remote_port;
u8 remote_mac[ETH_ALEN];
struct work_struct cleanup_work;
};
struct netpoll_info {
atomic_t refcnt;
struct semaphore dev_lock;
struct sk_buff_head txq;
struct delayed_work tx_work;
struct netpoll *netpoll;
struct rcu_head rcu;
};
#ifdef CONFIG_NETPOLL
extern void netpoll_poll_disable(struct net_device *dev);
extern void netpoll_poll_enable(struct net_device *dev);
#else
static inline void netpoll_poll_disable(struct net_device *dev) { return; }
Contributors
Person | Tokens | Prop | Commits | CommitProp |
Neil Horman | 10 | 76.92% | 1 | 33.33% |
Ding Tianhong | 2 | 15.38% | 1 | 33.33% |
Eric W. Biedermann | 1 | 7.69% | 1 | 33.33% |
Total | 13 | 100.00% | 3 | 100.00% |
static inline void netpoll_poll_enable(struct net_device *dev) { return; }
Contributors
Person | Tokens | Prop | Commits | CommitProp |
Neil Horman | 12 | 92.31% | 1 | 50.00% |
Eric W. Biedermann | 1 | 7.69% | 1 | 50.00% |
Total | 13 | 100.00% | 2 | 100.00% |
#endif
void netpoll_send_udp(struct netpoll *np, const char *msg, int len);
void netpoll_print_options(struct netpoll *np);
int netpoll_parse_options(struct netpoll *np, char *opt);
int __netpoll_setup(struct netpoll *np, struct net_device *ndev);
int netpoll_setup(struct netpoll *np);
void __netpoll_cleanup(struct netpoll *np);
void __netpoll_free_async(struct netpoll *np);
void netpoll_cleanup(struct netpoll *np);
void netpoll_send_skb_on_dev(struct netpoll *np, struct sk_buff *skb,
struct net_device *dev);
static inline void netpoll_send_skb(struct netpoll *np, struct sk_buff *skb)
{
unsigned long flags;
local_irq_save(flags);
netpoll_send_skb_on_dev(np, skb, np->dev);
local_irq_restore(flags);
}
Contributors
Person | Tokens | Prop | Commits | CommitProp |
Neil Horman | 28 | 66.67% | 1 | 50.00% |
Américo Wang | 14 | 33.33% | 1 | 50.00% |
Total | 42 | 100.00% | 2 | 100.00% |
#ifdef CONFIG_NETPOLL
static inline void *netpoll_poll_lock(struct napi_struct *napi)
{
struct net_device *dev = napi->dev;
if (dev && dev->npinfo) {
int owner = smp_processor_id();
while (cmpxchg(&napi->poll_owner, -1, owner) != -1)
cpu_relax();
return napi;
}
return NULL;
}
Contributors
Person | Tokens | Prop | Commits | CommitProp |
Matt Mackall | 29 | 44.62% | 2 | 40.00% |
Eric Dumazet | 20 | 30.77% | 1 | 20.00% |
Stephen Hemminger | 15 | 23.08% | 1 | 20.00% |
Jeff Moyer | 1 | 1.54% | 1 | 20.00% |
Total | 65 | 100.00% | 5 | 100.00% |
static inline void netpoll_poll_unlock(void *have)
{
struct napi_struct *napi = have;
if (napi)
smp_store_release(&napi->poll_owner, -1);
}
Contributors
Person | Tokens | Prop | Commits | CommitProp |
Matt Mackall | 21 | 63.64% | 2 | 40.00% |
Jeff Moyer | 4 | 12.12% | 1 | 20.00% |
Eric Dumazet | 4 | 12.12% | 1 | 20.00% |
Stephen Hemminger | 4 | 12.12% | 1 | 20.00% |
Total | 33 | 100.00% | 5 | 100.00% |
static inline bool netpoll_tx_running(struct net_device *dev)
{
return irqs_disabled();
}
Contributors
Person | Tokens | Prop | Commits | CommitProp |
Herbert Xu | 15 | 93.75% | 1 | 50.00% |
Américo Wang | 1 | 6.25% | 1 | 50.00% |
Total | 16 | 100.00% | 2 | 100.00% |
#else
static inline void *netpoll_poll_lock(struct napi_struct *napi)
{
return NULL;
}
Contributors
Person | Tokens | Prop | Commits | CommitProp |
Stephen Hemminger | 13 | 81.25% | 1 | 50.00% |
Matt Mackall | 3 | 18.75% | 1 | 50.00% |
Total | 16 | 100.00% | 2 | 100.00% |
static inline void netpoll_poll_unlock(void *have)
{
}
Contributors
Person | Tokens | Prop | Commits | CommitProp |
Stephen Hemminger | 7 | 70.00% | 1 | 50.00% |
Matt Mackall | 3 | 30.00% | 1 | 50.00% |
Total | 10 | 100.00% | 2 | 100.00% |
static inline void netpoll_netdev_init(struct net_device *dev)
{
}
Contributors
Person | Tokens | Prop | Commits | CommitProp |
Stephen Hemminger | 11 | 100.00% | 1 | 100.00% |
Total | 11 | 100.00% | 1 | 100.00% |
static inline bool netpoll_tx_running(struct net_device *dev)
{
return false;
}
Contributors
Person | Tokens | Prop | Commits | CommitProp |
Herbert Xu | 13 | 86.67% | 1 | 50.00% |
Américo Wang | 2 | 13.33% | 1 | 50.00% |
Total | 15 | 100.00% | 2 | 100.00% |
#endif
#endif
Overall Contributors
Person | Tokens | Prop | Commits | CommitProp |
Matt Mackall | 148 | 29.66% | 4 | 14.29% |
Neil Horman | 85 | 17.03% | 4 | 14.29% |
Américo Wang | 75 | 15.03% | 5 | 17.86% |
Stephen Hemminger | 64 | 12.83% | 4 | 14.29% |
Herbert Xu | 46 | 9.22% | 2 | 7.14% |
Eric W. Biedermann | 29 | 5.81% | 2 | 7.14% |
Eric Dumazet | 24 | 4.81% | 1 | 3.57% |
Jeff Moyer | 10 | 2.00% | 2 | 7.14% |
Satyam Sharma | 9 | 1.80% | 1 | 3.57% |
Jiri Pirko | 5 | 1.00% | 1 | 3.57% |
Ding Tianhong | 3 | 0.60% | 1 | 3.57% |
David Howells | 1 | 0.20% | 1 | 3.57% |
Total | 499 | 100.00% | 28 | 100.00% |
Information contained on this website is for historical information purposes only and does not indicate or represent copyright ownership.