Author | Tokens | Token Proportion | Commits | Commit Proportion |
---|---|---|---|---|
Frank Blaschka | 7804 | 56.11% | 18 | 9.89% |
Julian Wiedmann | 3270 | 23.51% | 91 | 50.00% |
Ursula Braun-Krahl | 814 | 5.85% | 20 | 10.99% |
Lakhvich Dmitriy | 629 | 4.52% | 1 | 0.55% |
Einar Lueck | 615 | 4.42% | 3 | 1.65% |
Thomas Richter | 170 | 1.22% | 6 | 3.30% |
Carsten Otte | 137 | 0.98% | 1 | 0.55% |
Jiri Pirko | 79 | 0.57% | 4 | 2.20% |
David S. Miller | 73 | 0.52% | 4 | 2.20% |
Klaus-Dieter Wacker | 66 | 0.47% | 1 | 0.55% |
Stefan Raspl | 54 | 0.39% | 4 | 2.20% |
Eugene Crosser | 40 | 0.29% | 3 | 1.65% |
Kittipon Meesompop | 33 | 0.24% | 2 | 1.10% |
Hans Wippel | 22 | 0.16% | 3 | 1.65% |
Patrick McHardy | 21 | 0.15% | 3 | 1.65% |
Peter Tiedemann | 17 | 0.12% | 1 | 0.55% |
Michał Mirosław | 16 | 0.12% | 1 | 0.55% |
Sachin P. Sant | 11 | 0.08% | 1 | 0.55% |
Sebastian Ott | 9 | 0.06% | 1 | 0.55% |
Heiko Carstens | 7 | 0.05% | 2 | 1.10% |
Hideaki Yoshifuji / 吉藤英明 | 3 | 0.02% | 1 | 0.55% |
Tejun Heo | 3 | 0.02% | 1 | 0.55% |
Eric Dumazet | 3 | 0.02% | 1 | 0.55% |
Gustavo A. R. Silva | 2 | 0.01% | 1 | 0.55% |
Ding Tianhong | 2 | 0.01% | 1 | 0.55% |
Joe Perches | 2 | 0.01% | 2 | 1.10% |
Paul Gortmaker | 2 | 0.01% | 1 | 0.55% |
Petr Machata | 2 | 0.01% | 1 | 0.55% |
Wilfried Klaebe | 1 | 0.01% | 1 | 0.55% |
Peter Senna Tschudin | 1 | 0.01% | 1 | 0.55% |
Greg Kroah-Hartman | 1 | 0.01% | 1 | 0.55% |
Total | 13909 | 182 |
// SPDX-License-Identifier: GPL-2.0 /* * Copyright IBM Corp. 2007, 2009 * Author(s): Utz Bacher <utz.bacher@de.ibm.com>, * Frank Pavlic <fpavlic@de.ibm.com>, * Thomas Spatzier <tspat@de.ibm.com>, * Frank Blaschka <frank.blaschka@de.ibm.com> */ #define KMSG_COMPONENT "qeth" #define pr_fmt(fmt) KMSG_COMPONENT ": " fmt #include <linux/module.h> #include <linux/moduleparam.h> #include <linux/bitops.h> #include <linux/string.h> #include <linux/errno.h> #include <linux/kernel.h> #include <linux/etherdevice.h> #include <linux/ip.h> #include <linux/in.h> #include <linux/ipv6.h> #include <linux/inetdevice.h> #include <linux/igmp.h> #include <linux/slab.h> #include <linux/if_ether.h> #include <linux/if_vlan.h> #include <linux/skbuff.h> #include <net/ip.h> #include <net/arp.h> #include <net/route.h> #include <net/ipv6.h> #include <net/ip6_route.h> #include <net/ip6_fib.h> #include <net/iucv/af_iucv.h> #include <linux/hashtable.h> #include "qeth_l3.h" static int qeth_l3_set_offline(struct ccwgroup_device *); static void qeth_l3_set_rx_mode(struct net_device *dev); static int qeth_l3_register_addr_entry(struct qeth_card *, struct qeth_ipaddr *); static int qeth_l3_deregister_addr_entry(struct qeth_card *, struct qeth_ipaddr *); static void qeth_l3_ipaddr4_to_string(const __u8 *addr, char *buf) { sprintf(buf, "%pI4", addr); } static void qeth_l3_ipaddr6_to_string(const __u8 *addr, char *buf) { sprintf(buf, "%pI6", addr); } void qeth_l3_ipaddr_to_string(enum qeth_prot_versions proto, const __u8 *addr, char *buf) { if (proto == QETH_PROT_IPV4) qeth_l3_ipaddr4_to_string(addr, buf); else if (proto == QETH_PROT_IPV6) qeth_l3_ipaddr6_to_string(addr, buf); } static struct qeth_ipaddr *qeth_l3_get_addr_buffer(enum qeth_prot_versions prot) { struct qeth_ipaddr *addr = kmalloc(sizeof(*addr), GFP_ATOMIC); if (addr) qeth_l3_init_ipaddr(addr, QETH_IP_TYPE_NORMAL, prot); return addr; } static struct qeth_ipaddr *qeth_l3_find_addr_by_ip(struct qeth_card *card, struct qeth_ipaddr *query) { u64 key = qeth_l3_ipaddr_hash(query); struct qeth_ipaddr *addr; if (query->is_multicast) { hash_for_each_possible(card->ip_mc_htable, addr, hnode, key) if (qeth_l3_addr_match_ip(addr, query)) return addr; } else { hash_for_each_possible(card->ip_htable, addr, hnode, key) if (qeth_l3_addr_match_ip(addr, query)) return addr; } return NULL; } static void qeth_l3_convert_addr_to_bits(u8 *addr, u8 *bits, int len) { int i, j; u8 octet; for (i = 0; i < len; ++i) { octet = addr[i]; for (j = 7; j >= 0; --j) { bits[i*8 + j] = octet & 1; octet >>= 1; } } } static bool qeth_l3_is_addr_covered_by_ipato(struct qeth_card *card, struct qeth_ipaddr *addr) { struct qeth_ipato_entry *ipatoe; u8 addr_bits[128] = {0, }; u8 ipatoe_bits[128] = {0, }; int rc = 0; if (!card->ipato.enabled) return false; if (addr->type != QETH_IP_TYPE_NORMAL) return false; qeth_l3_convert_addr_to_bits((u8 *) &addr->u, addr_bits, (addr->proto == QETH_PROT_IPV4)? 4:16); list_for_each_entry(ipatoe, &card->ipato.entries, entry) { if (addr->proto != ipatoe->proto) continue; qeth_l3_convert_addr_to_bits(ipatoe->addr, ipatoe_bits, (ipatoe->proto == QETH_PROT_IPV4) ? 4 : 16); if (addr->proto == QETH_PROT_IPV4) rc = !memcmp(addr_bits, ipatoe_bits, min(32, ipatoe->mask_bits)); else rc = !memcmp(addr_bits, ipatoe_bits, min(128, ipatoe->mask_bits)); if (rc) break; } /* invert? */ if ((addr->proto == QETH_PROT_IPV4) && card->ipato.invert4) rc = !rc; else if ((addr->proto == QETH_PROT_IPV6) && card->ipato.invert6) rc = !rc; return rc; } static int qeth_l3_delete_ip(struct qeth_card *card, struct qeth_ipaddr *tmp_addr) { int rc = 0; struct qeth_ipaddr *addr; if (tmp_addr->type == QETH_IP_TYPE_RXIP) QETH_CARD_TEXT(card, 2, "delrxip"); else if (tmp_addr->type == QETH_IP_TYPE_VIPA) QETH_CARD_TEXT(card, 2, "delvipa"); else QETH_CARD_TEXT(card, 2, "delip"); if (tmp_addr->proto == QETH_PROT_IPV4) QETH_CARD_HEX(card, 4, &tmp_addr->u.a4.addr, 4); else { QETH_CARD_HEX(card, 4, &tmp_addr->u.a6.addr, 8); QETH_CARD_HEX(card, 4, ((char *)&tmp_addr->u.a6.addr) + 8, 8); } addr = qeth_l3_find_addr_by_ip(card, tmp_addr); if (!addr || !qeth_l3_addr_match_all(addr, tmp_addr)) return -ENOENT; addr->ref_counter--; if (addr->type == QETH_IP_TYPE_NORMAL && addr->ref_counter > 0) return rc; if (addr->in_progress) return -EINPROGRESS; if (qeth_card_hw_is_reachable(card)) rc = qeth_l3_deregister_addr_entry(card, addr); hash_del(&addr->hnode); kfree(addr); return rc; } static int qeth_l3_add_ip(struct qeth_card *card, struct qeth_ipaddr *tmp_addr) { int rc = 0; struct qeth_ipaddr *addr; char buf[40]; if (tmp_addr->type == QETH_IP_TYPE_RXIP) QETH_CARD_TEXT(card, 2, "addrxip"); else if (tmp_addr->type == QETH_IP_TYPE_VIPA) QETH_CARD_TEXT(card, 2, "addvipa"); else QETH_CARD_TEXT(card, 2, "addip"); if (tmp_addr->proto == QETH_PROT_IPV4) QETH_CARD_HEX(card, 4, &tmp_addr->u.a4.addr, 4); else { QETH_CARD_HEX(card, 4, &tmp_addr->u.a6.addr, 8); QETH_CARD_HEX(card, 4, ((char *)&tmp_addr->u.a6.addr) + 8, 8); } addr = qeth_l3_find_addr_by_ip(card, tmp_addr); if (addr) { if (tmp_addr->type != QETH_IP_TYPE_NORMAL) return -EADDRINUSE; if (qeth_l3_addr_match_all(addr, tmp_addr)) { addr->ref_counter++; return 0; } qeth_l3_ipaddr_to_string(tmp_addr->proto, (u8 *)&tmp_addr->u, buf); dev_warn(&card->gdev->dev, "Registering IP address %s failed\n", buf); return -EADDRINUSE; } else { addr = qeth_l3_get_addr_buffer(tmp_addr->proto); if (!addr) return -ENOMEM; memcpy(addr, tmp_addr, sizeof(struct qeth_ipaddr)); addr->ref_counter = 1; if (qeth_l3_is_addr_covered_by_ipato(card, addr)) { QETH_CARD_TEXT(card, 2, "tkovaddr"); addr->ipato = 1; } hash_add(card->ip_htable, &addr->hnode, qeth_l3_ipaddr_hash(addr)); if (!qeth_card_hw_is_reachable(card)) { addr->disp_flag = QETH_DISP_ADDR_ADD; return 0; } /* qeth_l3_register_addr_entry can go to sleep * if we add a IPV4 addr. It is caused by the reason * that SETIP ipa cmd starts ARP staff for IPV4 addr. * Thus we should unlock spinlock, and make a protection * using in_progress variable to indicate that there is * an hardware operation with this IPV4 address */ if (addr->proto == QETH_PROT_IPV4) { addr->in_progress = 1; mutex_unlock(&card->ip_lock); rc = qeth_l3_register_addr_entry(card, addr); mutex_lock(&card->ip_lock); addr->in_progress = 0; } else rc = qeth_l3_register_addr_entry(card, addr); if (!rc || rc == -EADDRINUSE || rc == -ENETDOWN) { addr->disp_flag = QETH_DISP_ADDR_DO_NOTHING; if (addr->ref_counter < 1) { qeth_l3_deregister_addr_entry(card, addr); hash_del(&addr->hnode); kfree(addr); } } else { hash_del(&addr->hnode); kfree(addr); } } return rc; } static int qeth_l3_modify_ip(struct qeth_card *card, struct qeth_ipaddr *addr, bool add) { int rc; mutex_lock(&card->ip_lock); rc = add ? qeth_l3_add_ip(card, addr) : qeth_l3_delete_ip(card, addr); mutex_unlock(&card->ip_lock); return rc; } static void qeth_l3_drain_rx_mode_cache(struct qeth_card *card) { struct qeth_ipaddr *addr; struct hlist_node *tmp; int i; hash_for_each_safe(card->ip_mc_htable, i, tmp, addr, hnode) { hash_del(&addr->hnode); kfree(addr); } } static void qeth_l3_clear_ip_htable(struct qeth_card *card, int recover) { struct qeth_ipaddr *addr; struct hlist_node *tmp; int i; QETH_CARD_TEXT(card, 4, "clearip"); mutex_lock(&card->ip_lock); hash_for_each_safe(card->ip_htable, i, tmp, addr, hnode) { if (!recover) { hash_del(&addr->hnode); kfree(addr); continue; } addr->disp_flag = QETH_DISP_ADDR_ADD; } mutex_unlock(&card->ip_lock); } static void qeth_l3_recover_ip(struct qeth_card *card) { struct qeth_ipaddr *addr; struct hlist_node *tmp; int i; int rc; QETH_CARD_TEXT(card, 4, "recovrip"); mutex_lock(&card->ip_lock); hash_for_each_safe(card->ip_htable, i, tmp, addr, hnode) { if (addr->disp_flag == QETH_DISP_ADDR_ADD) { if (addr->proto == QETH_PROT_IPV4) { addr->in_progress = 1; mutex_unlock(&card->ip_lock); rc = qeth_l3_register_addr_entry(card, addr); mutex_lock(&card->ip_lock); addr->in_progress = 0; } else rc = qeth_l3_register_addr_entry(card, addr); if (!rc) { addr->disp_flag = QETH_DISP_ADDR_DO_NOTHING; if (addr->ref_counter < 1) qeth_l3_delete_ip(card, addr); } else { hash_del(&addr->hnode); kfree(addr); } } } mutex_unlock(&card->ip_lock); } static int qeth_l3_setdelip_cb(struct qeth_card *card, struct qeth_reply *reply, unsigned long data) { struct qeth_ipa_cmd *cmd = (struct qeth_ipa_cmd *) data; switch (cmd->hdr.return_code) { case IPA_RC_SUCCESS: return 0; case IPA_RC_DUPLICATE_IP_ADDRESS: return -EADDRINUSE; case IPA_RC_MC_ADDR_NOT_FOUND: return -ENOENT; case IPA_RC_LAN_OFFLINE: return -ENETDOWN; default: return -EIO; } } static int qeth_l3_send_setdelmc(struct qeth_card *card, struct qeth_ipaddr *addr, int ipacmd) { struct qeth_cmd_buffer *iob; struct qeth_ipa_cmd *cmd; QETH_CARD_TEXT(card, 4, "setdelmc"); iob = qeth_get_ipacmd_buffer(card, ipacmd, addr->proto); if (!iob) return -ENOMEM; cmd = __ipa_cmd(iob); ether_addr_copy(cmd->data.setdelipm.mac, addr->mac); if (addr->proto == QETH_PROT_IPV6) memcpy(cmd->data.setdelipm.ip6, &addr->u.a6.addr, sizeof(struct in6_addr)); else memcpy(&cmd->data.setdelipm.ip4, &addr->u.a4.addr, 4); return qeth_send_ipa_cmd(card, iob, qeth_l3_setdelip_cb, NULL); } static void qeth_l3_fill_netmask(u8 *netmask, unsigned int len) { int i, j; for (i = 0; i < 16; i++) { j = (len) - (i * 8); if (j >= 8) netmask[i] = 0xff; else if (j > 0) netmask[i] = (u8)(0xFF00 >> j); else netmask[i] = 0; } } static u32 qeth_l3_get_setdelip_flags(struct qeth_ipaddr *addr, bool set) { switch (addr->type) { case QETH_IP_TYPE_RXIP: return (set) ? QETH_IPA_SETIP_TAKEOVER_FLAG : 0; case QETH_IP_TYPE_VIPA: return (set) ? QETH_IPA_SETIP_VIPA_FLAG : QETH_IPA_DELIP_VIPA_FLAG; default: return (set && addr->ipato) ? QETH_IPA_SETIP_TAKEOVER_FLAG : 0; } } static int qeth_l3_send_setdelip(struct qeth_card *card, struct qeth_ipaddr *addr, enum qeth_ipa_cmds ipacmd) { struct qeth_cmd_buffer *iob; struct qeth_ipa_cmd *cmd; __u8 netmask[16]; u32 flags; QETH_CARD_TEXT(card, 4, "setdelip"); iob = qeth_get_ipacmd_buffer(card, ipacmd, addr->proto); if (!iob) return -ENOMEM; cmd = __ipa_cmd(iob); flags = qeth_l3_get_setdelip_flags(addr, ipacmd == IPA_CMD_SETIP); QETH_CARD_TEXT_(card, 4, "flags%02X", flags); if (addr->proto == QETH_PROT_IPV6) { memcpy(cmd->data.setdelip6.ip_addr, &addr->u.a6.addr, sizeof(struct in6_addr)); qeth_l3_fill_netmask(netmask, addr->u.a6.pfxlen); memcpy(cmd->data.setdelip6.mask, netmask, sizeof(struct in6_addr)); cmd->data.setdelip6.flags = flags; } else { memcpy(cmd->data.setdelip4.ip_addr, &addr->u.a4.addr, 4); memcpy(cmd->data.setdelip4.mask, &addr->u.a4.mask, 4); cmd->data.setdelip4.flags = flags; } return qeth_send_ipa_cmd(card, iob, qeth_l3_setdelip_cb, NULL); } static int qeth_l3_send_setrouting(struct qeth_card *card, enum qeth_routing_types type, enum qeth_prot_versions prot) { int rc; struct qeth_ipa_cmd *cmd; struct qeth_cmd_buffer *iob; QETH_CARD_TEXT(card, 4, "setroutg"); iob = qeth_get_ipacmd_buffer(card, IPA_CMD_SETRTG, prot); if (!iob) return -ENOMEM; cmd = __ipa_cmd(iob); cmd->data.setrtg.type = (type); rc = qeth_send_ipa_cmd(card, iob, NULL, NULL); return rc; } static int qeth_l3_correct_routing_type(struct qeth_card *card, enum qeth_routing_types *type, enum qeth_prot_versions prot) { if (IS_IQD(card)) { switch (*type) { case NO_ROUTER: case PRIMARY_CONNECTOR: case SECONDARY_CONNECTOR: case MULTICAST_ROUTER: return 0; default: goto out_inval; } } else { switch (*type) { case NO_ROUTER: case PRIMARY_ROUTER: case SECONDARY_ROUTER: return 0; case MULTICAST_ROUTER: if (qeth_is_ipafunc_supported(card, prot, IPA_OSA_MC_ROUTER)) return 0; default: goto out_inval; } } out_inval: *type = NO_ROUTER; return -EINVAL; } int qeth_l3_setrouting_v4(struct qeth_card *card) { int rc; QETH_CARD_TEXT(card, 3, "setrtg4"); rc = qeth_l3_correct_routing_type(card, &card->options.route4.type, QETH_PROT_IPV4); if (rc) return rc; rc = qeth_l3_send_setrouting(card, card->options.route4.type, QETH_PROT_IPV4); if (rc) { card->options.route4.type = NO_ROUTER; QETH_DBF_MESSAGE(2, "Error (%#06x) while setting routing type on device %x. Type set to 'no router'.\n", rc, CARD_DEVID(card)); } return rc; } int qeth_l3_setrouting_v6(struct qeth_card *card) { int rc = 0; QETH_CARD_TEXT(card, 3, "setrtg6"); if (!qeth_is_supported(card, IPA_IPV6)) return 0; rc = qeth_l3_correct_routing_type(card, &card->options.route6.type, QETH_PROT_IPV6); if (rc) return rc; rc = qeth_l3_send_setrouting(card, card->options.route6.type, QETH_PROT_IPV6); if (rc) { card->options.route6.type = NO_ROUTER; QETH_DBF_MESSAGE(2, "Error (%#06x) while setting routing type on device %x. Type set to 'no router'.\n", rc, CARD_DEVID(card)); } return rc; } /* * IP address takeover related functions */ /** * qeth_l3_update_ipato() - Update 'takeover' property, for all NORMAL IPs. * * Caller must hold ip_lock. */ void qeth_l3_update_ipato(struct qeth_card *card) { struct qeth_ipaddr *addr; unsigned int i; hash_for_each(card->ip_htable, i, addr, hnode) { if (addr->type != QETH_IP_TYPE_NORMAL) continue; addr->ipato = qeth_l3_is_addr_covered_by_ipato(card, addr); } } static void qeth_l3_clear_ipato_list(struct qeth_card *card) { struct qeth_ipato_entry *ipatoe, *tmp; mutex_lock(&card->ip_lock); list_for_each_entry_safe(ipatoe, tmp, &card->ipato.entries, entry) { list_del(&ipatoe->entry); kfree(ipatoe); } qeth_l3_update_ipato(card); mutex_unlock(&card->ip_lock); } int qeth_l3_add_ipato_entry(struct qeth_card *card, struct qeth_ipato_entry *new) { struct qeth_ipato_entry *ipatoe; int rc = 0; QETH_CARD_TEXT(card, 2, "addipato"); mutex_lock(&card->ip_lock); list_for_each_entry(ipatoe, &card->ipato.entries, entry) { if (ipatoe->proto != new->proto) continue; if (!memcmp(ipatoe->addr, new->addr, (ipatoe->proto == QETH_PROT_IPV4)? 4:16) && (ipatoe->mask_bits == new->mask_bits)) { rc = -EEXIST; break; } } if (!rc) { list_add_tail(&new->entry, &card->ipato.entries); qeth_l3_update_ipato(card); } mutex_unlock(&card->ip_lock); return rc; } int qeth_l3_del_ipato_entry(struct qeth_card *card, enum qeth_prot_versions proto, u8 *addr, int mask_bits) { struct qeth_ipato_entry *ipatoe, *tmp; int rc = -ENOENT; QETH_CARD_TEXT(card, 2, "delipato"); mutex_lock(&card->ip_lock); list_for_each_entry_safe(ipatoe, tmp, &card->ipato.entries, entry) { if (ipatoe->proto != proto) continue; if (!memcmp(ipatoe->addr, addr, (proto == QETH_PROT_IPV4)? 4:16) && (ipatoe->mask_bits == mask_bits)) { list_del(&ipatoe->entry); qeth_l3_update_ipato(card); kfree(ipatoe); rc = 0; } } mutex_unlock(&card->ip_lock); return rc; } int qeth_l3_modify_rxip_vipa(struct qeth_card *card, bool add, const u8 *ip, enum qeth_ip_types type, enum qeth_prot_versions proto) { struct qeth_ipaddr addr; qeth_l3_init_ipaddr(&addr, type, proto); if (proto == QETH_PROT_IPV4) memcpy(&addr.u.a4.addr, ip, 4); else memcpy(&addr.u.a6.addr, ip, 16); return qeth_l3_modify_ip(card, &addr, add); } int qeth_l3_modify_hsuid(struct qeth_card *card, bool add) { struct qeth_ipaddr addr; unsigned int i; qeth_l3_init_ipaddr(&addr, QETH_IP_TYPE_NORMAL, QETH_PROT_IPV6); addr.u.a6.addr.s6_addr[0] = 0xfe; addr.u.a6.addr.s6_addr[1] = 0x80; for (i = 0; i < 8; i++) addr.u.a6.addr.s6_addr[8+i] = card->options.hsuid[i]; return qeth_l3_modify_ip(card, &addr, add); } static int qeth_l3_register_addr_entry(struct qeth_card *card, struct qeth_ipaddr *addr) { char buf[50]; int rc = 0; int cnt = 3; if (card->options.sniffer) return 0; if (addr->proto == QETH_PROT_IPV4) { QETH_CARD_TEXT(card, 2, "setaddr4"); QETH_CARD_HEX(card, 3, &addr->u.a4.addr, sizeof(int)); } else if (addr->proto == QETH_PROT_IPV6) { QETH_CARD_TEXT(card, 2, "setaddr6"); QETH_CARD_HEX(card, 3, &addr->u.a6.addr, 8); QETH_CARD_HEX(card, 3, ((char *)&addr->u.a6.addr) + 8, 8); } else { QETH_CARD_TEXT(card, 2, "setaddr?"); QETH_CARD_HEX(card, 3, addr, sizeof(struct qeth_ipaddr)); } do { if (addr->is_multicast) rc = qeth_l3_send_setdelmc(card, addr, IPA_CMD_SETIPM); else rc = qeth_l3_send_setdelip(card, addr, IPA_CMD_SETIP); if (rc) QETH_CARD_TEXT(card, 2, "failed"); } while ((--cnt > 0) && rc); if (rc) { QETH_CARD_TEXT(card, 2, "FAILED"); qeth_l3_ipaddr_to_string(addr->proto, (u8 *)&addr->u, buf); dev_warn(&card->gdev->dev, "Registering IP address %s failed\n", buf); } return rc; } static int qeth_l3_deregister_addr_entry(struct qeth_card *card, struct qeth_ipaddr *addr) { int rc = 0; if (card->options.sniffer) return 0; if (addr->proto == QETH_PROT_IPV4) { QETH_CARD_TEXT(card, 2, "deladdr4"); QETH_CARD_HEX(card, 3, &addr->u.a4.addr, sizeof(int)); } else if (addr->proto == QETH_PROT_IPV6) { QETH_CARD_TEXT(card, 2, "deladdr6"); QETH_CARD_HEX(card, 3, &addr->u.a6.addr, 8); QETH_CARD_HEX(card, 3, ((char *)&addr->u.a6.addr) + 8, 8); } else { QETH_CARD_TEXT(card, 2, "deladdr?"); QETH_CARD_HEX(card, 3, addr, sizeof(struct qeth_ipaddr)); } if (addr->is_multicast) rc = qeth_l3_send_setdelmc(card, addr, IPA_CMD_DELIPM); else rc = qeth_l3_send_setdelip(card, addr, IPA_CMD_DELIP); if (rc) QETH_CARD_TEXT(card, 2, "failed"); return rc; } static int qeth_l3_setadapter_parms(struct qeth_card *card) { int rc = 0; QETH_DBF_TEXT(SETUP, 2, "setadprm"); if (qeth_adp_supported(card, IPA_SETADP_ALTER_MAC_ADDRESS)) { rc = qeth_setadpparms_change_macaddr(card); if (rc) dev_warn(&card->gdev->dev, "Reading the adapter MAC" " address failed\n"); } return rc; } static int qeth_l3_start_ipa_arp_processing(struct qeth_card *card) { int rc; QETH_CARD_TEXT(card, 3, "ipaarp"); if (!qeth_is_supported(card, IPA_ARP_PROCESSING)) { dev_info(&card->gdev->dev, "ARP processing not supported on %s!\n", QETH_CARD_IFNAME(card)); return 0; } rc = qeth_send_simple_setassparms(card, IPA_ARP_PROCESSING, IPA_CMD_ASS_START, 0); if (rc) { dev_warn(&card->gdev->dev, "Starting ARP processing support for %s failed\n", QETH_CARD_IFNAME(card)); } return rc; } static int qeth_l3_start_ipa_source_mac(struct qeth_card *card) { int rc; QETH_CARD_TEXT(card, 3, "stsrcmac"); if (!qeth_is_supported(card, IPA_SOURCE_MAC)) { dev_info(&card->gdev->dev, "Inbound source MAC-address not supported on %s\n", QETH_CARD_IFNAME(card)); return -EOPNOTSUPP; } rc = qeth_send_simple_setassparms(card, IPA_SOURCE_MAC, IPA_CMD_ASS_START, 0); if (rc) dev_warn(&card->gdev->dev, "Starting source MAC-address support for %s failed\n", QETH_CARD_IFNAME(card)); return rc; } static int qeth_l3_start_ipa_vlan(struct qeth_card *card) { int rc = 0; QETH_CARD_TEXT(card, 3, "strtvlan"); if (!qeth_is_supported(card, IPA_FULL_VLAN)) { dev_info(&card->gdev->dev, "VLAN not supported on %s\n", QETH_CARD_IFNAME(card)); return -EOPNOTSUPP; } rc = qeth_send_simple_setassparms(card, IPA_VLAN_PRIO, IPA_CMD_ASS_START, 0); if (rc) { dev_warn(&card->gdev->dev, "Starting VLAN support for %s failed\n", QETH_CARD_IFNAME(card)); } else { dev_info(&card->gdev->dev, "VLAN enabled\n"); } return rc; } static int qeth_l3_start_ipa_multicast(struct qeth_card *card) { int rc; QETH_CARD_TEXT(card, 3, "stmcast"); if (!qeth_is_supported(card, IPA_MULTICASTING)) { dev_info(&card->gdev->dev, "Multicast not supported on %s\n", QETH_CARD_IFNAME(card)); return -EOPNOTSUPP; } rc = qeth_send_simple_setassparms(card, IPA_MULTICASTING, IPA_CMD_ASS_START, 0); if (rc) { dev_warn(&card->gdev->dev, "Starting multicast support for %s failed\n", QETH_CARD_IFNAME(card)); } else { dev_info(&card->gdev->dev, "Multicast enabled\n"); card->dev->flags |= IFF_MULTICAST; } return rc; } static int qeth_l3_softsetup_ipv6(struct qeth_card *card) { int rc; QETH_CARD_TEXT(card, 3, "softipv6"); if (IS_IQD(card)) goto out; rc = qeth_send_simple_setassparms(card, IPA_IPV6, IPA_CMD_ASS_START, 3); if (rc) { dev_err(&card->gdev->dev, "Activating IPv6 support for %s failed\n", QETH_CARD_IFNAME(card)); return rc; } rc = qeth_send_simple_setassparms_v6(card, IPA_IPV6, IPA_CMD_ASS_START, 0); if (rc) { dev_err(&card->gdev->dev, "Activating IPv6 support for %s failed\n", QETH_CARD_IFNAME(card)); return rc; } rc = qeth_send_simple_setassparms_v6(card, IPA_PASSTHRU, IPA_CMD_ASS_START, 0); if (rc) { dev_warn(&card->gdev->dev, "Enabling the passthrough mode for %s failed\n", QETH_CARD_IFNAME(card)); return rc; } out: dev_info(&card->gdev->dev, "IPV6 enabled\n"); return 0; } static int qeth_l3_start_ipa_ipv6(struct qeth_card *card) { QETH_CARD_TEXT(card, 3, "strtipv6"); if (!qeth_is_supported(card, IPA_IPV6)) { dev_info(&card->gdev->dev, "IPv6 not supported on %s\n", QETH_CARD_IFNAME(card)); return 0; } return qeth_l3_softsetup_ipv6(card); } static int qeth_l3_start_ipa_broadcast(struct qeth_card *card) { int rc; QETH_CARD_TEXT(card, 3, "stbrdcst"); card->info.broadcast_capable = 0; if (!qeth_is_supported(card, IPA_FILTERING)) { dev_info(&card->gdev->dev, "Broadcast not supported on %s\n", QETH_CARD_IFNAME(card)); rc = -EOPNOTSUPP; goto out; } rc = qeth_send_simple_setassparms(card, IPA_FILTERING, IPA_CMD_ASS_START, 0); if (rc) { dev_warn(&card->gdev->dev, "Enabling broadcast filtering for " "%s failed\n", QETH_CARD_IFNAME(card)); goto out; } rc = qeth_send_simple_setassparms(card, IPA_FILTERING, IPA_CMD_ASS_CONFIGURE, 1); if (rc) { dev_warn(&card->gdev->dev, "Setting up broadcast filtering for %s failed\n", QETH_CARD_IFNAME(card)); goto out; } card->info.broadcast_capable = QETH_BROADCAST_WITH_ECHO; dev_info(&card->gdev->dev, "Broadcast enabled\n"); rc = qeth_send_simple_setassparms(card, IPA_FILTERING, IPA_CMD_ASS_ENABLE, 1); if (rc) { dev_warn(&card->gdev->dev, "Setting up broadcast echo " "filtering for %s failed\n", QETH_CARD_IFNAME(card)); goto out; } card->info.broadcast_capable = QETH_BROADCAST_WITHOUT_ECHO; out: if (card->info.broadcast_capable) card->dev->flags |= IFF_BROADCAST; else card->dev->flags &= ~IFF_BROADCAST; return rc; } static int qeth_l3_start_ipassists(struct qeth_card *card) { QETH_CARD_TEXT(card, 3, "strtipas"); if (qeth_set_access_ctrl_online(card, 0)) return -EIO; qeth_l3_start_ipa_arp_processing(card); /* go on*/ qeth_l3_start_ipa_source_mac(card); /* go on*/ qeth_l3_start_ipa_vlan(card); /* go on*/ qeth_l3_start_ipa_multicast(card); /* go on*/ qeth_l3_start_ipa_ipv6(card); /* go on*/ qeth_l3_start_ipa_broadcast(card); /* go on*/ return 0; } static int qeth_l3_iqd_read_initial_mac_cb(struct qeth_card *card, struct qeth_reply *reply, unsigned long data) { struct qeth_ipa_cmd *cmd = (struct qeth_ipa_cmd *) data; if (cmd->hdr.return_code) return -EIO; ether_addr_copy(card->dev->dev_addr, cmd->data.create_destroy_addr.unique_id); return 0; } static int qeth_l3_iqd_read_initial_mac(struct qeth_card *card) { int rc = 0; struct qeth_cmd_buffer *iob; struct qeth_ipa_cmd *cmd; QETH_DBF_TEXT(SETUP, 2, "hsrmac"); iob = qeth_get_ipacmd_buffer(card, IPA_CMD_CREATE_ADDR, QETH_PROT_IPV6); if (!iob) return -ENOMEM; cmd = __ipa_cmd(iob); *((__u16 *) &cmd->data.create_destroy_addr.unique_id[6]) = card->info.unique_id; rc = qeth_send_ipa_cmd(card, iob, qeth_l3_iqd_read_initial_mac_cb, NULL); return rc; } static int qeth_l3_get_unique_id_cb(struct qeth_card *card, struct qeth_reply *reply, unsigned long data) { struct qeth_ipa_cmd *cmd = (struct qeth_ipa_cmd *) data; if (cmd->hdr.return_code == 0) { card->info.unique_id = *((__u16 *) &cmd->data.create_destroy_addr.unique_id[6]); return 0; } card->info.unique_id = UNIQUE_ID_IF_CREATE_ADDR_FAILED | UNIQUE_ID_NOT_BY_CARD; dev_warn(&card->gdev->dev, "The network adapter failed to generate a unique ID\n"); return -EIO; } static int qeth_l3_get_unique_id(struct qeth_card *card) { int rc = 0; struct qeth_cmd_buffer *iob; struct qeth_ipa_cmd *cmd; QETH_DBF_TEXT(SETUP, 2, "guniqeid"); if (!qeth_is_supported(card, IPA_IPV6)) { card->info.unique_id = UNIQUE_ID_IF_CREATE_ADDR_FAILED | UNIQUE_ID_NOT_BY_CARD; return 0; } iob = qeth_get_ipacmd_buffer(card, IPA_CMD_CREATE_ADDR, QETH_PROT_IPV6); if (!iob) return -ENOMEM; cmd = __ipa_cmd(iob); *((__u16 *) &cmd->data.create_destroy_addr.unique_id[6]) = card->info.unique_id; rc = qeth_send_ipa_cmd(card, iob, qeth_l3_get_unique_id_cb, NULL); return rc; } static int qeth_diags_trace_cb(struct qeth_card *card, struct qeth_reply *reply, unsigned long data) { struct qeth_ipa_cmd *cmd; __u16 rc; QETH_DBF_TEXT(SETUP, 2, "diastrcb"); cmd = (struct qeth_ipa_cmd *)data; rc = cmd->hdr.return_code; if (rc) QETH_CARD_TEXT_(card, 2, "dxter%x", rc); switch (cmd->data.diagass.action) { case QETH_DIAGS_CMD_TRACE_QUERY: break; case QETH_DIAGS_CMD_TRACE_DISABLE: switch (rc) { case 0: case IPA_RC_INVALID_SUBCMD: card->info.promisc_mode = SET_PROMISC_MODE_OFF; dev_info(&card->gdev->dev, "The HiperSockets network " "traffic analyzer is deactivated\n"); break; default: break; } break; case QETH_DIAGS_CMD_TRACE_ENABLE: switch (rc) { case 0: card->info.promisc_mode = SET_PROMISC_MODE_ON; dev_info(&card->gdev->dev, "The HiperSockets network " "traffic analyzer is activated\n"); break; case IPA_RC_HARDWARE_AUTH_ERROR: dev_warn(&card->gdev->dev, "The device is not " "authorized to run as a HiperSockets network " "traffic analyzer\n"); break; case IPA_RC_TRACE_ALREADY_ACTIVE: dev_warn(&card->gdev->dev, "A HiperSockets " "network traffic analyzer is already " "active in the HiperSockets LAN\n"); break; default: break; } break; default: QETH_DBF_MESSAGE(2, "Unknown sniffer action (%#06x) on device %x\n", cmd->data.diagass.action, CARD_DEVID(card)); } return rc ? -EIO : 0; } static int qeth_diags_trace(struct qeth_card *card, enum qeth_diags_trace_cmds diags_cmd) { struct qeth_cmd_buffer *iob; struct qeth_ipa_cmd *cmd; QETH_DBF_TEXT(SETUP, 2, "diagtrac"); iob = qeth_get_ipacmd_buffer(card, IPA_CMD_SET_DIAG_ASS, 0); if (!iob) return -ENOMEM; cmd = __ipa_cmd(iob); cmd->data.diagass.subcmd_len = 16; cmd->data.diagass.subcmd = QETH_DIAGS_CMD_TRACE; cmd->data.diagass.type = QETH_DIAGS_TYPE_HIPERSOCKET; cmd->data.diagass.action = diags_cmd; return qeth_send_ipa_cmd(card, iob, qeth_diags_trace_cb, NULL); } static void qeth_l3_add_mc_to_hash(struct qeth_card *card, struct in_device *in4_dev) { struct ip_mc_list *im4; struct qeth_ipaddr *tmp, *ipm; QETH_CARD_TEXT(card, 4, "addmc"); tmp = qeth_l3_get_addr_buffer(QETH_PROT_IPV4); if (!tmp) return; for (im4 = rcu_dereference(in4_dev->mc_list); im4 != NULL; im4 = rcu_dereference(im4->next_rcu)) { ip_eth_mc_map(im4->multiaddr, tmp->mac); tmp->u.a4.addr = be32_to_cpu(im4->multiaddr); tmp->is_multicast = 1; ipm = qeth_l3_find_addr_by_ip(card, tmp); if (ipm) { /* for mcast, by-IP match means full match */ ipm->disp_flag = QETH_DISP_ADDR_DO_NOTHING; } else { ipm = qeth_l3_get_addr_buffer(QETH_PROT_IPV4); if (!ipm) continue; ether_addr_copy(ipm->mac, tmp->mac); ipm->u.a4.addr = be32_to_cpu(im4->multiaddr); ipm->is_multicast = 1; ipm->disp_flag = QETH_DISP_ADDR_ADD; hash_add(card->ip_mc_htable, &ipm->hnode, qeth_l3_ipaddr_hash(ipm)); } } kfree(tmp); } /* called with rcu_read_lock */ static void qeth_l3_add_vlan_mc(struct qeth_card *card) { struct in_device *in_dev; u16 vid; QETH_CARD_TEXT(card, 4, "addmcvl"); if (!qeth_is_supported(card, IPA_FULL_VLAN)) return; for_each_set_bit(vid, card->active_vlans, VLAN_N_VID) { struct net_device *netdev; netdev = __vlan_find_dev_deep_rcu(card->dev, htons(ETH_P_8021Q), vid); if (netdev == NULL || !(netdev->flags & IFF_UP)) continue; in_dev = __in_dev_get_rcu(netdev); if (!in_dev) continue; qeth_l3_add_mc_to_hash(card, in_dev); } } static void qeth_l3_add_multicast_ipv4(struct qeth_card *card) { struct in_device *in4_dev; QETH_CARD_TEXT(card, 4, "chkmcv4"); rcu_read_lock(); in4_dev = __in_dev_get_rcu(card->dev); if (in4_dev == NULL) goto unlock; qeth_l3_add_mc_to_hash(card, in4_dev); qeth_l3_add_vlan_mc(card); unlock: rcu_read_unlock(); } static void qeth_l3_add_mc6_to_hash(struct qeth_card *card, struct inet6_dev *in6_dev) { struct qeth_ipaddr *ipm; struct ifmcaddr6 *im6; struct qeth_ipaddr *tmp; QETH_CARD_TEXT(card, 4, "addmc6"); tmp = qeth_l3_get_addr_buffer(QETH_PROT_IPV6); if (!tmp) return; for (im6 = in6_dev->mc_list; im6 != NULL; im6 = im6->next) { ipv6_eth_mc_map(&im6->mca_addr, tmp->mac); memcpy(&tmp->u.a6.addr, &im6->mca_addr.s6_addr, sizeof(struct in6_addr)); tmp->is_multicast = 1; ipm = qeth_l3_find_addr_by_ip(card, tmp); if (ipm) { /* for mcast, by-IP match means full match */ ipm->disp_flag = QETH_DISP_ADDR_DO_NOTHING; continue; } ipm = qeth_l3_get_addr_buffer(QETH_PROT_IPV6); if (!ipm) continue; ether_addr_copy(ipm->mac, tmp->mac); memcpy(&ipm->u.a6.addr, &im6->mca_addr.s6_addr, sizeof(struct in6_addr)); ipm->is_multicast = 1; ipm->disp_flag = QETH_DISP_ADDR_ADD; hash_add(card->ip_mc_htable, &ipm->hnode, qeth_l3_ipaddr_hash(ipm)); } kfree(tmp); } /* called with rcu_read_lock */ static void qeth_l3_add_vlan_mc6(struct qeth_card *card) { struct inet6_dev *in_dev; u16 vid; QETH_CARD_TEXT(card, 4, "admc6vl"); if (!qeth_is_supported(card, IPA_FULL_VLAN)) return; for_each_set_bit(vid, card->active_vlans, VLAN_N_VID) { struct net_device *netdev; netdev = __vlan_find_dev_deep_rcu(card->dev, htons(ETH_P_8021Q), vid); if (netdev == NULL || !(netdev->flags & IFF_UP)) continue; in_dev = in6_dev_get(netdev); if (!in_dev) continue; read_lock_bh(&in_dev->lock); qeth_l3_add_mc6_to_hash(card, in_dev); read_unlock_bh(&in_dev->lock); in6_dev_put(in_dev); } } static void qeth_l3_add_multicast_ipv6(struct qeth_card *card) { struct inet6_dev *in6_dev; QETH_CARD_TEXT(card, 4, "chkmcv6"); if (!qeth_is_supported(card, IPA_IPV6)) return ; in6_dev = in6_dev_get(card->dev); if (!in6_dev) return; rcu_read_lock(); read_lock_bh(&in6_dev->lock); qeth_l3_add_mc6_to_hash(card, in6_dev); qeth_l3_add_vlan_mc6(card); read_unlock_bh(&in6_dev->lock); rcu_read_unlock(); in6_dev_put(in6_dev); } static int qeth_l3_vlan_rx_add_vid(struct net_device *dev, __be16 proto, u16 vid) { struct qeth_card *card = dev->ml_priv; set_bit(vid, card->active_vlans); return 0; } static int qeth_l3_vlan_rx_kill_vid(struct net_device *dev, __be16 proto, u16 vid) { struct qeth_card *card = dev->ml_priv; QETH_CARD_TEXT_(card, 4, "kid:%d", vid); clear_bit(vid, card->active_vlans); qeth_l3_set_rx_mode(dev); return 0; } static void qeth_l3_rebuild_skb(struct qeth_card *card, struct sk_buff *skb, struct qeth_hdr *hdr) { if (!(hdr->hdr.l3.flags & QETH_HDR_PASSTHRU)) { u16 prot = (hdr->hdr.l3.flags & QETH_HDR_IPV6) ? ETH_P_IPV6 : ETH_P_IP; unsigned char tg_addr[ETH_ALEN]; skb_reset_network_header(skb); switch (hdr->hdr.l3.flags & QETH_HDR_CAST_MASK) { case QETH_CAST_MULTICAST: if (prot == ETH_P_IP) ip_eth_mc_map(ip_hdr(skb)->daddr, tg_addr); else ipv6_eth_mc_map(&ipv6_hdr(skb)->daddr, tg_addr); QETH_CARD_STAT_INC(card, rx_multicast); break; case QETH_CAST_BROADCAST: ether_addr_copy(tg_addr, card->dev->broadcast); QETH_CARD_STAT_INC(card, rx_multicast); break; default: if (card->options.sniffer) skb->pkt_type = PACKET_OTHERHOST; ether_addr_copy(tg_addr, card->dev->dev_addr); } if (hdr->hdr.l3.ext_flags & QETH_HDR_EXT_SRC_MAC_ADDR) card->dev->header_ops->create(skb, card->dev, prot, tg_addr, &hdr->hdr.l3.next_hop.rx.src_mac, skb->len); else card->dev->header_ops->create(skb, card->dev, prot, tg_addr, "FAKELL", skb->len); } skb->protocol = eth_type_trans(skb, card->dev); /* copy VLAN tag from hdr into skb */ if (!card->options.sniffer && (hdr->hdr.l3.ext_flags & (QETH_HDR_EXT_VLAN_FRAME | QETH_HDR_EXT_INCLUDE_VLAN_TAG))) { u16 tag = (hdr->hdr.l3.ext_flags & QETH_HDR_EXT_VLAN_FRAME) ? hdr->hdr.l3.vlan_id : hdr->hdr.l3.next_hop.rx.vlan_id; __vlan_hwaccel_put_tag(skb, htons(ETH_P_8021Q), tag); } qeth_rx_csum(card, skb, hdr->hdr.l3.ext_flags); } static int qeth_l3_process_inbound_buffer(struct qeth_card *card, int budget, int *done) { struct net_device *dev = card->dev; int work_done = 0; struct sk_buff *skb; struct qeth_hdr *hdr; unsigned int len; __u16 magic; *done = 0; WARN_ON_ONCE(!budget); while (budget) { skb = qeth_core_get_next_skb(card, &card->qdio.in_q->bufs[card->rx.b_index], &card->rx.b_element, &card->rx.e_offset, &hdr); if (!skb) { *done = 1; break; } switch (hdr->hdr.l3.id) { case QETH_HEADER_TYPE_LAYER3: magic = *(__u16 *)skb->data; if (IS_IQD(card) && magic == ETH_P_AF_IUCV) { len = skb->len; dev_hard_header(skb, dev, ETH_P_AF_IUCV, dev->dev_addr, "FAKELL", len); skb->protocol = eth_type_trans(skb, dev); netif_receive_skb(skb); } else { qeth_l3_rebuild_skb(card, skb, hdr); len = skb->len; napi_gro_receive(&card->napi, skb); } break; case QETH_HEADER_TYPE_LAYER2: /* for HiperSockets sniffer */ skb->protocol = eth_type_trans(skb, skb->dev); len = skb->len; netif_receive_skb(skb); break; default: dev_kfree_skb_any(skb); QETH_CARD_TEXT(card, 3, "inbunkno"); QETH_DBF_HEX(CTRL, 3, hdr, sizeof(*hdr)); continue; } work_done++; budget--; QETH_CARD_STAT_INC(card, rx_packets); QETH_CARD_STAT_ADD(card, rx_bytes, len); } return work_done; } static void qeth_l3_stop_card(struct qeth_card *card) { QETH_DBF_TEXT(SETUP, 2, "stopcard"); QETH_DBF_HEX(SETUP, 2, &card, sizeof(void *)); qeth_set_allowed_threads(card, 0, 1); cancel_work_sync(&card->rx_mode_work); qeth_l3_drain_rx_mode_cache(card); if (card->options.sniffer && (card->info.promisc_mode == SET_PROMISC_MODE_ON)) qeth_diags_trace(card, QETH_DIAGS_CMD_TRACE_DISABLE); if (card->state == CARD_STATE_SOFTSETUP) { qeth_l3_clear_ip_htable(card, 1); qeth_clear_ipacmd_list(card); card->state = CARD_STATE_HARDSETUP; } if (card->state == CARD_STATE_HARDSETUP) { qeth_qdio_clear_card(card, 0); qeth_drain_output_queues(card); qeth_clear_working_pool_list(card); card->state = CARD_STATE_DOWN; } if (card->state == CARD_STATE_DOWN) { qeth_clear_cmd_buffers(&card->read); qeth_clear_cmd_buffers(&card->write); } flush_workqueue(card->event_wq); } /* * test for and Switch promiscuous mode (on or off) * either for guestlan or HiperSocket Sniffer */ static void qeth_l3_handle_promisc_mode(struct qeth_card *card) { struct net_device *dev = card->dev; if (((dev->flags & IFF_PROMISC) && (card->info.promisc_mode == SET_PROMISC_MODE_ON)) || (!(dev->flags & IFF_PROMISC) && (card->info.promisc_mode == SET_PROMISC_MODE_OFF))) return; if (IS_VM_NIC(card)) { /* Guestlan trace */ if (qeth_adp_supported(card, IPA_SETADP_SET_PROMISC_MODE)) qeth_setadp_promisc_mode(card); } else if (card->options.sniffer && /* HiperSockets trace */ qeth_adp_supported(card, IPA_SETADP_SET_DIAG_ASSIST)) { if (dev->flags & IFF_PROMISC) { QETH_CARD_TEXT(card, 3, "+promisc"); qeth_diags_trace(card, QETH_DIAGS_CMD_TRACE_ENABLE); } else { QETH_CARD_TEXT(card, 3, "-promisc"); qeth_diags_trace(card, QETH_DIAGS_CMD_TRACE_DISABLE); } } } static void qeth_l3_rx_mode_work(struct work_struct *work) { struct qeth_card *card = container_of(work, struct qeth_card, rx_mode_work); struct qeth_ipaddr *addr; struct hlist_node *tmp; int i, rc; QETH_CARD_TEXT(card, 3, "setmulti"); if (!card->options.sniffer) { qeth_l3_add_multicast_ipv4(card); qeth_l3_add_multicast_ipv6(card); hash_for_each_safe(card->ip_mc_htable, i, tmp, addr, hnode) { switch (addr->disp_flag) { case QETH_DISP_ADDR_DELETE: rc = qeth_l3_deregister_addr_entry(card, addr); if (!rc || rc == -ENOENT) { hash_del(&addr->hnode); kfree(addr); } break; case QETH_DISP_ADDR_ADD: rc = qeth_l3_register_addr_entry(card, addr); if (rc && rc != -ENETDOWN) { hash_del(&addr->hnode); kfree(addr); break; } addr->ref_counter = 1; /* fall through */ default: /* for next call to set_rx_mode(): */ addr->disp_flag = QETH_DISP_ADDR_DELETE; } } if (!qeth_adp_supported(card, IPA_SETADP_SET_PROMISC_MODE)) return; } qeth_l3_handle_promisc_mode(card); } static int qeth_l3_arp_makerc(u16 rc) { switch (rc) { case IPA_RC_SUCCESS: return 0; case QETH_IPA_ARP_RC_NOTSUPP: case QETH_IPA_ARP_RC_Q_NOTSUPP: return -EOPNOTSUPP; case QETH_IPA_ARP_RC_OUT_OF_RANGE: return -EINVAL; case QETH_IPA_ARP_RC_Q_NO_DATA: return -ENOENT; default: return -EIO; } } static int qeth_l3_arp_cmd_cb(struct qeth_card *card, struct qeth_reply *reply, unsigned long data) { struct qeth_ipa_cmd *cmd = (struct qeth_ipa_cmd *) data; qeth_setassparms_cb(card, reply, data); return qeth_l3_arp_makerc(cmd->hdr.return_code); } static int qeth_l3_arp_set_no_entries(struct qeth_card *card, int no_entries) { struct qeth_cmd_buffer *iob; int rc; QETH_CARD_TEXT(card, 3, "arpstnoe"); /* * currently GuestLAN only supports the ARP assist function * IPA_CMD_ASS_ARP_QUERY_INFO, but not IPA_CMD_ASS_ARP_SET_NO_ENTRIES; * thus we say EOPNOTSUPP for this ARP function */ if (IS_VM_NIC(card)) return -EOPNOTSUPP; if (!qeth_is_supported(card, IPA_ARP_PROCESSING)) { return -EOPNOTSUPP; } iob = qeth_get_setassparms_cmd(card, IPA_ARP_PROCESSING, IPA_CMD_ASS_ARP_SET_NO_ENTRIES, 4, QETH_PROT_IPV4); if (!iob) return -ENOMEM; __ipa_cmd(iob)->data.setassparms.data.flags_32bit = (u32) no_entries; rc = qeth_send_ipa_cmd(card, iob, qeth_l3_arp_cmd_cb, NULL); if (rc) QETH_DBF_MESSAGE(2, "Could not set number of ARP entries on device %x: %#x\n", CARD_DEVID(card), rc); return rc; } static __u32 get_arp_entry_size(struct qeth_card *card, struct qeth_arp_query_data *qdata, struct qeth_arp_entrytype *type, __u8 strip_entries) { __u32 rc; __u8 is_hsi; is_hsi = qdata->reply_bits == 5; if (type->ip == QETHARP_IP_ADDR_V4) { QETH_CARD_TEXT(card, 4, "arpev4"); if (strip_entries) { rc = is_hsi ? sizeof(struct qeth_arp_qi_entry5_short) : sizeof(struct qeth_arp_qi_entry7_short); } else { rc = is_hsi ? sizeof(struct qeth_arp_qi_entry5) : sizeof(struct qeth_arp_qi_entry7); } } else if (type->ip == QETHARP_IP_ADDR_V6) { QETH_CARD_TEXT(card, 4, "arpev6"); if (strip_entries) { rc = is_hsi ? sizeof(struct qeth_arp_qi_entry5_short_ipv6) : sizeof(struct qeth_arp_qi_entry7_short_ipv6); } else { rc = is_hsi ? sizeof(struct qeth_arp_qi_entry5_ipv6) : sizeof(struct qeth_arp_qi_entry7_ipv6); } } else { QETH_CARD_TEXT(card, 4, "arpinv"); rc = 0; } return rc; } static int arpentry_matches_prot(struct qeth_arp_entrytype *type, __u16 prot) { return (type->ip == QETHARP_IP_ADDR_V4 && prot == QETH_PROT_IPV4) || (type->ip == QETHARP_IP_ADDR_V6 && prot == QETH_PROT_IPV6); } static int qeth_l3_arp_query_cb(struct qeth_card *card, struct qeth_reply *reply, unsigned long data) { struct qeth_ipa_cmd *cmd; struct qeth_arp_query_data *qdata; struct qeth_arp_query_info *qinfo; int e; int entrybytes_done; int stripped_bytes; __u8 do_strip_entries; QETH_CARD_TEXT(card, 3, "arpquecb"); qinfo = (struct qeth_arp_query_info *) reply->param; cmd = (struct qeth_ipa_cmd *) data; QETH_CARD_TEXT_(card, 4, "%i", cmd->hdr.prot_version); if (cmd->hdr.return_code) { QETH_CARD_TEXT(card, 4, "arpcberr"); QETH_CARD_TEXT_(card, 4, "%i", cmd->hdr.return_code); return qeth_l3_arp_makerc(cmd->hdr.return_code); } if (cmd->data.setassparms.hdr.return_code) { cmd->hdr.return_code = cmd->data.setassparms.hdr.return_code; QETH_CARD_TEXT(card, 4, "setaperr"); QETH_CARD_TEXT_(card, 4, "%i", cmd->hdr.return_code); return qeth_l3_arp_makerc(cmd->hdr.return_code); } qdata = &cmd->data.setassparms.data.query_arp; QETH_CARD_TEXT_(card, 4, "anoen%i", qdata->no_entries); do_strip_entries = (qinfo->mask_bits & QETH_QARP_STRIP_ENTRIES) > 0; stripped_bytes = do_strip_entries ? QETH_QARP_MEDIASPECIFIC_BYTES : 0; entrybytes_done = 0; for (e = 0; e < qdata->no_entries; ++e) { char *cur_entry; __u32 esize; struct qeth_arp_entrytype *etype; cur_entry = &qdata->data + entrybytes_done; etype = &((struct qeth_arp_qi_entry5 *) cur_entry)->type; if (!arpentry_matches_prot(etype, cmd->hdr.prot_version)) { QETH_CARD_TEXT(card, 4, "pmis"); QETH_CARD_TEXT_(card, 4, "%i", etype->ip); break; } esize = get_arp_entry_size(card, qdata, etype, do_strip_entries); QETH_CARD_TEXT_(card, 5, "esz%i", esize); if (!esize) break; if ((qinfo->udata_len - qinfo->udata_offset) < esize) { QETH_CARD_TEXT_(card, 4, "qaer3%i", -ENOSPC); memset(qinfo->udata, 0, 4); return -ENOSPC; } memcpy(qinfo->udata + qinfo->udata_offset, &qdata->data + entrybytes_done + stripped_bytes, esize); entrybytes_done += esize + stripped_bytes; qinfo->udata_offset += esize; ++qinfo->no_entries; } /* check if all replies received ... */ if (cmd->data.setassparms.hdr.seq_no < cmd->data.setassparms.hdr.number_of_replies) return 1; QETH_CARD_TEXT_(card, 4, "nove%i", qinfo->no_entries); memcpy(qinfo->udata, &qinfo->no_entries, 4); /* keep STRIP_ENTRIES flag so the user program can distinguish * stripped entries from normal ones */ if (qinfo->mask_bits & QETH_QARP_STRIP_ENTRIES) qdata->reply_bits |= QETH_QARP_STRIP_ENTRIES; memcpy(qinfo->udata + QETH_QARP_MASK_OFFSET, &qdata->reply_bits, 2); QETH_CARD_TEXT_(card, 4, "rc%i", 0); return 0; } static int qeth_l3_query_arp_cache_info(struct qeth_card *card, enum qeth_prot_versions prot, struct qeth_arp_query_info *qinfo) { struct qeth_cmd_buffer *iob; struct qeth_ipa_cmd *cmd; int rc; QETH_CARD_TEXT_(card, 3, "qarpipv%i", prot); iob = qeth_get_setassparms_cmd(card, IPA_ARP_PROCESSING, IPA_CMD_ASS_ARP_QUERY_INFO, sizeof(struct qeth_arp_query_data) - sizeof(char), prot); if (!iob) return -ENOMEM; cmd = __ipa_cmd(iob); cmd->data.setassparms.data.query_arp.request_bits = 0x000F; rc = qeth_send_ipa_cmd(card, iob, qeth_l3_arp_query_cb, qinfo); if (rc) QETH_DBF_MESSAGE(2, "Error while querying ARP cache on device %x: %#x\n", CARD_DEVID(card), rc); return rc; } static int qeth_l3_arp_query(struct qeth_card *card, char __user *udata) { struct qeth_arp_query_info qinfo = {0, }; int rc; QETH_CARD_TEXT(card, 3, "arpquery"); if (!qeth_is_supported(card,/*IPA_QUERY_ARP_ADDR_INFO*/ IPA_ARP_PROCESSING)) { QETH_CARD_TEXT(card, 3, "arpqnsup"); rc = -EOPNOTSUPP; goto out; } /* get size of userspace buffer and mask_bits -> 6 bytes */ if (copy_from_user(&qinfo, udata, 6)) { rc = -EFAULT; goto out; } qinfo.udata = kzalloc(qinfo.udata_len, GFP_KERNEL); if (!qinfo.udata) { rc = -ENOMEM; goto out; } qinfo.udata_offset = QETH_QARP_ENTRIES_OFFSET; rc = qeth_l3_query_arp_cache_info(card, QETH_PROT_IPV4, &qinfo); if (rc) { if (copy_to_user(udata, qinfo.udata, 4)) rc = -EFAULT; goto free_and_out; } if (qinfo.mask_bits & QETH_QARP_WITH_IPV6) { /* fails in case of GuestLAN QDIO mode */ qeth_l3_query_arp_cache_info(card, QETH_PROT_IPV6, &qinfo); } if (copy_to_user(udata, qinfo.udata, qinfo.udata_len)) { QETH_CARD_TEXT(card, 4, "qactf"); rc = -EFAULT; goto free_and_out; } QETH_CARD_TEXT(card, 4, "qacts"); free_and_out: kfree(qinfo.udata); out: return rc; } static int qeth_l3_arp_modify_entry(struct qeth_card *card, struct qeth_arp_cache_entry *entry, enum qeth_arp_process_subcmds arp_cmd) { struct qeth_arp_cache_entry *cmd_entry; struct qeth_cmd_buffer *iob; int rc; if (arp_cmd == IPA_CMD_ASS_ARP_ADD_ENTRY) QETH_CARD_TEXT(card, 3, "arpadd"); else QETH_CARD_TEXT(card, 3, "arpdel"); /* * currently GuestLAN only supports the ARP assist function * IPA_CMD_ASS_ARP_QUERY_INFO, but not IPA_CMD_ASS_ARP_ADD_ENTRY; * thus we say EOPNOTSUPP for this ARP function */ if (IS_VM_NIC(card)) return -EOPNOTSUPP; if (!qeth_is_supported(card, IPA_ARP_PROCESSING)) { return -EOPNOTSUPP; } iob = qeth_get_setassparms_cmd(card, IPA_ARP_PROCESSING, arp_cmd, sizeof(*cmd_entry), QETH_PROT_IPV4); if (!iob) return -ENOMEM; cmd_entry = &__ipa_cmd(iob)->data.setassparms.data.arp_entry; ether_addr_copy(cmd_entry->macaddr, entry->macaddr); memcpy(cmd_entry->ipaddr, entry->ipaddr, 4); rc = qeth_send_ipa_cmd(card, iob, qeth_l3_arp_cmd_cb, NULL); if (rc) QETH_DBF_MESSAGE(2, "Could not modify (cmd: %#x) ARP entry on device %x: %#x\n", arp_cmd, CARD_DEVID(card), rc); return rc; } static int qeth_l3_arp_flush_cache(struct qeth_card *card) { struct qeth_cmd_buffer *iob; int rc; QETH_CARD_TEXT(card, 3, "arpflush"); /* * currently GuestLAN only supports the ARP assist function * IPA_CMD_ASS_ARP_QUERY_INFO, but not IPA_CMD_ASS_ARP_FLUSH_CACHE; * thus we say EOPNOTSUPP for this ARP function */ if (IS_VM_NIC(card) || IS_IQD(card)) return -EOPNOTSUPP; if (!qeth_is_supported(card, IPA_ARP_PROCESSING)) { return -EOPNOTSUPP; } iob = qeth_get_setassparms_cmd(card, IPA_ARP_PROCESSING, IPA_CMD_ASS_ARP_FLUSH_CACHE, 0, QETH_PROT_IPV4); if (!iob) return -ENOMEM; rc = qeth_send_ipa_cmd(card, iob, qeth_l3_arp_cmd_cb, NULL); if (rc) QETH_DBF_MESSAGE(2, "Could not flush ARP cache on device %x: %#x\n", CARD_DEVID(card), rc); return rc; } static int qeth_l3_do_ioctl(struct net_device *dev, struct ifreq *rq, int cmd) { struct qeth_card *card = dev->ml_priv; struct qeth_arp_cache_entry arp_entry; enum qeth_arp_process_subcmds arp_cmd; int rc = 0; switch (cmd) { case SIOC_QETH_ARP_SET_NO_ENTRIES: if (!capable(CAP_NET_ADMIN)) { rc = -EPERM; break; } rc = qeth_l3_arp_set_no_entries(card, rq->ifr_ifru.ifru_ivalue); break; case SIOC_QETH_ARP_QUERY_INFO: if (!capable(CAP_NET_ADMIN)) { rc = -EPERM; break; } rc = qeth_l3_arp_query(card, rq->ifr_ifru.ifru_data); break; case SIOC_QETH_ARP_ADD_ENTRY: case SIOC_QETH_ARP_REMOVE_ENTRY: if (!capable(CAP_NET_ADMIN)) return -EPERM; if (copy_from_user(&arp_entry, rq->ifr_data, sizeof(arp_entry))) return -EFAULT; arp_cmd = (cmd == SIOC_QETH_ARP_ADD_ENTRY) ? IPA_CMD_ASS_ARP_ADD_ENTRY : IPA_CMD_ASS_ARP_REMOVE_ENTRY; return qeth_l3_arp_modify_entry(card, &arp_entry, arp_cmd); case SIOC_QETH_ARP_FLUSH_CACHE: if (!capable(CAP_NET_ADMIN)) { rc = -EPERM; break; } rc = qeth_l3_arp_flush_cache(card); break; default: rc = -EOPNOTSUPP; } return rc; } static int qeth_l3_get_cast_type(struct sk_buff *skb) { int ipv = qeth_get_ip_version(skb); struct neighbour *n = NULL; struct dst_entry *dst; rcu_read_lock(); dst = skb_dst(skb); if (dst) { struct rt6_info *rt = (struct rt6_info *) dst; dst = dst_check(dst, (ipv == 6) ? rt6_get_cookie(rt) : 0); if (dst) n = dst_neigh_lookup_skb(dst, skb); } if (n) { int cast_type = n->type; rcu_read_unlock(); neigh_release(n); if ((cast_type == RTN_BROADCAST) || (cast_type == RTN_MULTICAST) || (cast_type == RTN_ANYCAST)) return cast_type; return RTN_UNICAST; } rcu_read_unlock(); /* no neighbour (eg AF_PACKET), fall back to target's IP address ... */ switch (ipv) { case 4: if (ipv4_is_lbcast(ip_hdr(skb)->daddr)) return RTN_BROADCAST; return ipv4_is_multicast(ip_hdr(skb)->daddr) ? RTN_MULTICAST : RTN_UNICAST; case 6: return ipv6_addr_is_multicast(&ipv6_hdr(skb)->daddr) ? RTN_MULTICAST : RTN_UNICAST; default: /* ... and MAC address */ return qeth_get_ether_cast_type(skb); } } static u8 qeth_l3_cast_type_to_flag(int cast_type) { if (cast_type == RTN_MULTICAST) return QETH_CAST_MULTICAST; if (cast_type == RTN_ANYCAST) return QETH_CAST_ANYCAST; if (cast_type == RTN_BROADCAST) return QETH_CAST_BROADCAST; return QETH_CAST_UNICAST; } static void qeth_l3_fill_header(struct qeth_qdio_out_q *queue, struct qeth_hdr *hdr, struct sk_buff *skb, int ipv, int cast_type, unsigned int data_len) { struct qeth_hdr_layer3 *l3_hdr = &hdr->hdr.l3; struct vlan_ethhdr *veth = vlan_eth_hdr(skb); struct qeth_card *card = queue->card; struct dst_entry *dst; hdr->hdr.l3.length = data_len; if (skb_is_gso(skb)) { hdr->hdr.l3.id = QETH_HEADER_TYPE_L3_TSO; } else { hdr->hdr.l3.id = QETH_HEADER_TYPE_LAYER3; if (skb->protocol == htons(ETH_P_AF_IUCV)) { l3_hdr->flags = QETH_HDR_IPV6 | QETH_CAST_UNICAST; l3_hdr->next_hop.ipv6_addr.s6_addr16[0] = htons(0xfe80); memcpy(&l3_hdr->next_hop.ipv6_addr.s6_addr32[2], iucv_trans_hdr(skb)->destUserID, 8); return; } if (skb->ip_summed == CHECKSUM_PARTIAL) { qeth_tx_csum(skb, &hdr->hdr.l3.ext_flags, ipv); /* some HW requires combined L3+L4 csum offload: */ if (ipv == 4) hdr->hdr.l3.ext_flags |= QETH_HDR_EXT_CSUM_HDR_REQ; QETH_TXQ_STAT_INC(queue, skbs_csum); } } if (ipv == 4 || IS_IQD(card)) { /* NETIF_F_HW_VLAN_CTAG_TX */ if (skb_vlan_tag_present(skb)) { hdr->hdr.l3.ext_flags |= QETH_HDR_EXT_VLAN_FRAME; hdr->hdr.l3.vlan_id = skb_vlan_tag_get(skb); } } else if (veth->h_vlan_proto == htons(ETH_P_8021Q)) { hdr->hdr.l3.ext_flags |= QETH_HDR_EXT_INCLUDE_VLAN_TAG; hdr->hdr.l3.vlan_id = ntohs(veth->h_vlan_TCI); } l3_hdr->flags = qeth_l3_cast_type_to_flag(cast_type); /* OSA only: */ if (!ipv) { l3_hdr->flags |= QETH_HDR_PASSTHRU; return; } rcu_read_lock(); dst = skb_dst(skb); if (ipv == 4) { struct rtable *rt; if (dst) dst = dst_check(dst, 0); rt = (struct rtable *) dst; *((__be32 *) &hdr->hdr.l3.next_hop.ipv4.addr) = (rt) ? rt_nexthop(rt, ip_hdr(skb)->daddr) : ip_hdr(skb)->daddr; } else { /* IPv6 */ struct rt6_info *rt; if (dst) { rt = (struct rt6_info *) dst; dst = dst_check(dst, rt6_get_cookie(rt)); } rt = (struct rt6_info *) dst; if (rt && !ipv6_addr_any(&rt->rt6i_gateway)) l3_hdr->next_hop.ipv6_addr = rt->rt6i_gateway; else l3_hdr->next_hop.ipv6_addr = ipv6_hdr(skb)->daddr; hdr->hdr.l3.flags |= QETH_HDR_IPV6; if (!IS_IQD(card)) hdr->hdr.l3.flags |= QETH_HDR_PASSTHRU; } rcu_read_unlock(); } static void qeth_l3_fixup_headers(struct sk_buff *skb) { struct iphdr *iph = ip_hdr(skb); /* this is safe, IPv6 traffic takes a different path */ if (skb->ip_summed == CHECKSUM_PARTIAL) iph->check = 0; if (skb_is_gso(skb)) { iph->tot_len = 0; tcp_hdr(skb)->check = ~tcp_v4_check(0, iph->saddr, iph->daddr, 0); } } static int qeth_l3_xmit(struct qeth_card *card, struct sk_buff *skb, struct qeth_qdio_out_q *queue, int ipv, int cast_type) { unsigned int hw_hdr_len; int rc; /* re-use the L2 header area for the HW header: */ hw_hdr_len = skb_is_gso(skb) ? sizeof(struct qeth_hdr_tso) : sizeof(struct qeth_hdr); rc = skb_cow_head(skb, hw_hdr_len - ETH_HLEN); if (rc) return rc; skb_pull(skb, ETH_HLEN); qeth_l3_fixup_headers(skb); return qeth_xmit(card, skb, queue, ipv, cast_type, qeth_l3_fill_header); } static netdev_tx_t qeth_l3_hard_start_xmit(struct sk_buff *skb, struct net_device *dev) { struct qeth_card *card = dev->ml_priv; u16 txq = skb_get_queue_mapping(skb); int ipv = qeth_get_ip_version(skb); struct qeth_qdio_out_q *queue; int tx_bytes = skb->len; int cast_type, rc; if (IS_IQD(card)) { queue = card->qdio.out_qs[qeth_iqd_translate_txq(dev, txq)]; if (card->options.sniffer) goto tx_drop; if ((card->options.cq != QETH_CQ_ENABLED && !ipv) || (card->options.cq == QETH_CQ_ENABLED && skb->protocol != htons(ETH_P_AF_IUCV))) goto tx_drop; if (txq == QETH_IQD_MCAST_TXQ) cast_type = qeth_l3_get_cast_type(skb); else cast_type = RTN_UNICAST; } else { queue = card->qdio.out_qs[txq]; cast_type = qeth_l3_get_cast_type(skb); } if (cast_type == RTN_BROADCAST && !card->info.broadcast_capable) goto tx_drop; if (ipv == 4 || IS_IQD(card)) rc = qeth_l3_xmit(card, skb, queue, ipv, cast_type); else rc = qeth_xmit(card, skb, queue, ipv, cast_type, qeth_l3_fill_header); if (!rc) { QETH_TXQ_STAT_INC(queue, tx_packets); QETH_TXQ_STAT_ADD(queue, tx_bytes, tx_bytes); return NETDEV_TX_OK; } tx_drop: QETH_TXQ_STAT_INC(queue, tx_dropped); kfree_skb(skb); return NETDEV_TX_OK; } static void qeth_l3_set_rx_mode(struct net_device *dev) { struct qeth_card *card = dev->ml_priv; schedule_work(&card->rx_mode_work); } /* * we need NOARP for IPv4 but we want neighbor solicitation for IPv6. Setting * NOARP on the netdevice is no option because it also turns off neighbor * solicitation. For IPv4 we install a neighbor_setup function. We don't want * arp resolution but we want the hard header (packet socket will work * e.g. tcpdump) */ static int qeth_l3_neigh_setup_noarp(struct neighbour *n) { n->nud_state = NUD_NOARP; memcpy(n->ha, "FAKELL", 6); n->output = n->ops->connected_output; return 0; } static int qeth_l3_neigh_setup(struct net_device *dev, struct neigh_parms *np) { if (np->tbl->family == AF_INET) np->neigh_setup = qeth_l3_neigh_setup_noarp; return 0; } static netdev_features_t qeth_l3_osa_features_check(struct sk_buff *skb, struct net_device *dev, netdev_features_t features) { if (qeth_get_ip_version(skb) != 4) features &= ~NETIF_F_HW_VLAN_CTAG_TX; return qeth_features_check(skb, dev, features); } static u16 qeth_l3_iqd_select_queue(struct net_device *dev, struct sk_buff *skb, struct net_device *sb_dev) { return qeth_iqd_select_queue(dev, skb, qeth_l3_get_cast_type(skb), sb_dev); } static u16 qeth_l3_osa_select_queue(struct net_device *dev, struct sk_buff *skb, struct net_device *sb_dev) { struct qeth_card *card = dev->ml_priv; return qeth_get_priority_queue(card, skb); } static const struct net_device_ops qeth_l3_netdev_ops = { .ndo_open = qeth_open, .ndo_stop = qeth_stop, .ndo_get_stats64 = qeth_get_stats64, .ndo_start_xmit = qeth_l3_hard_start_xmit, .ndo_select_queue = qeth_l3_iqd_select_queue, .ndo_validate_addr = eth_validate_addr, .ndo_set_rx_mode = qeth_l3_set_rx_mode, .ndo_do_ioctl = qeth_do_ioctl, .ndo_fix_features = qeth_fix_features, .ndo_set_features = qeth_set_features, .ndo_vlan_rx_add_vid = qeth_l3_vlan_rx_add_vid, .ndo_vlan_rx_kill_vid = qeth_l3_vlan_rx_kill_vid, .ndo_tx_timeout = qeth_tx_timeout, }; static const struct net_device_ops qeth_l3_osa_netdev_ops = { .ndo_open = qeth_open, .ndo_stop = qeth_stop, .ndo_get_stats64 = qeth_get_stats64, .ndo_start_xmit = qeth_l3_hard_start_xmit, .ndo_features_check = qeth_l3_osa_features_check, .ndo_select_queue = qeth_l3_osa_select_queue, .ndo_validate_addr = eth_validate_addr, .ndo_set_rx_mode = qeth_l3_set_rx_mode, .ndo_do_ioctl = qeth_do_ioctl, .ndo_fix_features = qeth_fix_features, .ndo_set_features = qeth_set_features, .ndo_vlan_rx_add_vid = qeth_l3_vlan_rx_add_vid, .ndo_vlan_rx_kill_vid = qeth_l3_vlan_rx_kill_vid, .ndo_tx_timeout = qeth_tx_timeout, .ndo_neigh_setup = qeth_l3_neigh_setup, }; static int qeth_l3_setup_netdev(struct qeth_card *card, bool carrier_ok) { unsigned int headroom; int rc; if (IS_OSD(card) || IS_OSX(card)) { if ((card->info.link_type == QETH_LINK_TYPE_LANE_TR) || (card->info.link_type == QETH_LINK_TYPE_HSTR)) { pr_info("qeth_l3: ignoring TR device\n"); return -ENODEV; } card->dev->netdev_ops = &qeth_l3_osa_netdev_ops; /*IPv6 address autoconfiguration stuff*/ qeth_l3_get_unique_id(card); if (!(card->info.unique_id & UNIQUE_ID_NOT_BY_CARD)) card->dev->dev_id = card->info.unique_id & 0xffff; if (!IS_VM_NIC(card)) { card->dev->features |= NETIF_F_SG; card->dev->hw_features |= NETIF_F_TSO | NETIF_F_RXCSUM | NETIF_F_IP_CSUM; card->dev->vlan_features |= NETIF_F_TSO | NETIF_F_RXCSUM | NETIF_F_IP_CSUM; } if (qeth_is_supported6(card, IPA_OUTBOUND_CHECKSUM_V6)) { card->dev->hw_features |= NETIF_F_IPV6_CSUM; card->dev->vlan_features |= NETIF_F_IPV6_CSUM; } if (qeth_is_supported6(card, IPA_OUTBOUND_TSO)) { card->dev->hw_features |= NETIF_F_TSO6; card->dev->vlan_features |= NETIF_F_TSO6; } /* allow for de-acceleration of NETIF_F_HW_VLAN_CTAG_TX: */ if (card->dev->hw_features & NETIF_F_TSO6) headroom = sizeof(struct qeth_hdr_tso) + VLAN_HLEN; else if (card->dev->hw_features & NETIF_F_TSO) headroom = sizeof(struct qeth_hdr_tso); else headroom = sizeof(struct qeth_hdr) + VLAN_HLEN; } else if (IS_IQD(card)) { card->dev->flags |= IFF_NOARP; card->dev->netdev_ops = &qeth_l3_netdev_ops; headroom = sizeof(struct qeth_hdr) - ETH_HLEN; rc = qeth_l3_iqd_read_initial_mac(card); if (rc) goto out; } else return -ENODEV; card->dev->needed_headroom = headroom; card->dev->features |= NETIF_F_HW_VLAN_CTAG_TX | NETIF_F_HW_VLAN_CTAG_RX | NETIF_F_HW_VLAN_CTAG_FILTER; netif_keep_dst(card->dev); if (card->dev->hw_features & (NETIF_F_TSO | NETIF_F_TSO6)) netif_set_gso_max_size(card->dev, PAGE_SIZE * (QETH_MAX_BUFFER_ELEMENTS(card) - 1)); netif_napi_add(card->dev, &card->napi, qeth_poll, QETH_NAPI_WEIGHT); rc = register_netdev(card->dev); if (!rc && carrier_ok) netif_carrier_on(card->dev); out: if (rc) card->dev->netdev_ops = NULL; return rc; } static const struct device_type qeth_l3_devtype = { .name = "qeth_layer3", .groups = qeth_l3_attr_groups, }; static int qeth_l3_probe_device(struct ccwgroup_device *gdev) { struct qeth_card *card = dev_get_drvdata(&gdev->dev); int rc; hash_init(card->ip_htable); mutex_init(&card->ip_lock); card->cmd_wq = alloc_ordered_workqueue("%s_cmd", 0, dev_name(&gdev->dev)); if (!card->cmd_wq) return -ENOMEM; if (gdev->dev.type == &qeth_generic_devtype) { rc = qeth_l3_create_device_attributes(&gdev->dev); if (rc) { destroy_workqueue(card->cmd_wq); return rc; } } hash_init(card->ip_mc_htable); INIT_WORK(&card->rx_mode_work, qeth_l3_rx_mode_work); return 0; } static void qeth_l3_remove_device(struct ccwgroup_device *cgdev) { struct qeth_card *card = dev_get_drvdata(&cgdev->dev); if (cgdev->dev.type == &qeth_generic_devtype) qeth_l3_remove_device_attributes(&cgdev->dev); qeth_set_allowed_threads(card, 0, 1); wait_event(card->wait_q, qeth_threads_running(card, 0xffffffff) == 0); if (cgdev->state == CCWGROUP_ONLINE) qeth_l3_set_offline(cgdev); cancel_work_sync(&card->close_dev_work); if (qeth_netdev_is_registered(card->dev)) unregister_netdev(card->dev); flush_workqueue(card->cmd_wq); destroy_workqueue(card->cmd_wq); qeth_l3_clear_ip_htable(card, 0); qeth_l3_clear_ipato_list(card); } static int qeth_l3_set_online(struct ccwgroup_device *gdev) { struct qeth_card *card = dev_get_drvdata(&gdev->dev); struct net_device *dev = card->dev; int rc = 0; bool carrier_ok; mutex_lock(&card->discipline_mutex); mutex_lock(&card->conf_mutex); QETH_DBF_TEXT(SETUP, 2, "setonlin"); QETH_DBF_HEX(SETUP, 2, &card, sizeof(void *)); rc = qeth_core_hardsetup_card(card, &carrier_ok); if (rc) { QETH_DBF_TEXT_(SETUP, 2, "2err%04x", rc); rc = -ENODEV; goto out_remove; } if (qeth_is_diagass_supported(card, QETH_DIAGS_CMD_TRAP)) { if (card->info.hwtrap && qeth_hw_trap(card, QETH_DIAGS_TRAP_ARM)) card->info.hwtrap = 0; } else card->info.hwtrap = 0; card->state = CARD_STATE_HARDSETUP; qeth_print_status_message(card); /* softsetup */ QETH_DBF_TEXT(SETUP, 2, "softsetp"); rc = qeth_l3_setadapter_parms(card); if (rc) QETH_DBF_TEXT_(SETUP, 2, "2err%04x", rc); if (!card->options.sniffer) { rc = qeth_l3_start_ipassists(card); if (rc) { QETH_DBF_TEXT_(SETUP, 2, "3err%d", rc); goto out_remove; } rc = qeth_l3_setrouting_v4(card); if (rc) QETH_DBF_TEXT_(SETUP, 2, "4err%04x", rc); rc = qeth_l3_setrouting_v6(card); if (rc) QETH_DBF_TEXT_(SETUP, 2, "5err%04x", rc); } rc = qeth_init_qdio_queues(card); if (rc) { QETH_DBF_TEXT_(SETUP, 2, "6err%d", rc); rc = -ENODEV; goto out_remove; } card->state = CARD_STATE_SOFTSETUP; qeth_set_allowed_threads(card, 0xffffffff, 0); qeth_l3_recover_ip(card); if (!qeth_netdev_is_registered(dev)) { rc = qeth_l3_setup_netdev(card, carrier_ok); if (rc) goto out_remove; } else { rtnl_lock(); if (carrier_ok) netif_carrier_on(dev); else netif_carrier_off(dev); netif_device_attach(dev); qeth_enable_hw_features(dev); if (card->info.open_when_online) { card->info.open_when_online = 0; dev_open(dev, NULL); } rtnl_unlock(); } qeth_trace_features(card); /* let user_space know that device is online */ kobject_uevent(&gdev->dev.kobj, KOBJ_CHANGE); mutex_unlock(&card->conf_mutex); mutex_unlock(&card->discipline_mutex); return 0; out_remove: qeth_l3_stop_card(card); ccw_device_set_offline(CARD_DDEV(card)); ccw_device_set_offline(CARD_WDEV(card)); ccw_device_set_offline(CARD_RDEV(card)); qdio_free(CARD_DDEV(card)); card->state = CARD_STATE_DOWN; mutex_unlock(&card->conf_mutex); mutex_unlock(&card->discipline_mutex); return rc; } static int __qeth_l3_set_offline(struct ccwgroup_device *cgdev, int recovery_mode) { struct qeth_card *card = dev_get_drvdata(&cgdev->dev); int rc = 0, rc2 = 0, rc3 = 0; mutex_lock(&card->discipline_mutex); mutex_lock(&card->conf_mutex); QETH_DBF_TEXT(SETUP, 3, "setoffl"); QETH_DBF_HEX(SETUP, 3, &card, sizeof(void *)); if ((!recovery_mode && card->info.hwtrap) || card->info.hwtrap == 2) { qeth_hw_trap(card, QETH_DIAGS_TRAP_DISARM); card->info.hwtrap = 1; } rtnl_lock(); card->info.open_when_online = card->dev->flags & IFF_UP; dev_close(card->dev); netif_device_detach(card->dev); netif_carrier_off(card->dev); rtnl_unlock(); qeth_l3_stop_card(card); if (card->options.cq == QETH_CQ_ENABLED) { rtnl_lock(); call_netdevice_notifiers(NETDEV_REBOOT, card->dev); rtnl_unlock(); } rc = ccw_device_set_offline(CARD_DDEV(card)); rc2 = ccw_device_set_offline(CARD_WDEV(card)); rc3 = ccw_device_set_offline(CARD_RDEV(card)); if (!rc) rc = (rc2) ? rc2 : rc3; if (rc) QETH_DBF_TEXT_(SETUP, 2, "1err%d", rc); qdio_free(CARD_DDEV(card)); /* let user_space know that device is offline */ kobject_uevent(&cgdev->dev.kobj, KOBJ_CHANGE); mutex_unlock(&card->conf_mutex); mutex_unlock(&card->discipline_mutex); return 0; } static int qeth_l3_set_offline(struct ccwgroup_device *cgdev) { return __qeth_l3_set_offline(cgdev, 0); } static int qeth_l3_recover(void *ptr) { struct qeth_card *card; int rc = 0; card = (struct qeth_card *) ptr; QETH_CARD_TEXT(card, 2, "recover1"); QETH_CARD_HEX(card, 2, &card, sizeof(void *)); if (!qeth_do_run_thread(card, QETH_RECOVER_THREAD)) return 0; QETH_CARD_TEXT(card, 2, "recover2"); dev_warn(&card->gdev->dev, "A recovery process has been started for the device\n"); __qeth_l3_set_offline(card->gdev, 1); rc = qeth_l3_set_online(card->gdev); if (!rc) dev_info(&card->gdev->dev, "Device successfully recovered!\n"); else { ccwgroup_set_offline(card->gdev); dev_warn(&card->gdev->dev, "The qeth device driver " "failed to recover an error on the device\n"); } qeth_clear_thread_start_bit(card, QETH_RECOVER_THREAD); qeth_clear_thread_running_bit(card, QETH_RECOVER_THREAD); return 0; } static int qeth_l3_pm_suspend(struct ccwgroup_device *gdev) { struct qeth_card *card = dev_get_drvdata(&gdev->dev); qeth_set_allowed_threads(card, 0, 1); wait_event(card->wait_q, qeth_threads_running(card, 0xffffffff) == 0); if (gdev->state == CCWGROUP_OFFLINE) return 0; qeth_l3_set_offline(gdev); return 0; } static int qeth_l3_pm_resume(struct ccwgroup_device *gdev) { struct qeth_card *card = dev_get_drvdata(&gdev->dev); int rc; rc = qeth_l3_set_online(gdev); qeth_set_allowed_threads(card, 0xffffffff, 0); if (rc) dev_warn(&card->gdev->dev, "The qeth device driver " "failed to recover an error on the device\n"); return rc; } /* Returns zero if the command is successfully "consumed" */ static int qeth_l3_control_event(struct qeth_card *card, struct qeth_ipa_cmd *cmd) { return 1; } struct qeth_discipline qeth_l3_discipline = { .devtype = &qeth_l3_devtype, .process_rx_buffer = qeth_l3_process_inbound_buffer, .recover = qeth_l3_recover, .setup = qeth_l3_probe_device, .remove = qeth_l3_remove_device, .set_online = qeth_l3_set_online, .set_offline = qeth_l3_set_offline, .freeze = qeth_l3_pm_suspend, .thaw = qeth_l3_pm_resume, .restore = qeth_l3_pm_resume, .do_ioctl = qeth_l3_do_ioctl, .control_event_handler = qeth_l3_control_event, }; EXPORT_SYMBOL_GPL(qeth_l3_discipline); static int qeth_l3_handle_ip_event(struct qeth_card *card, struct qeth_ipaddr *addr, unsigned long event) { switch (event) { case NETDEV_UP: qeth_l3_modify_ip(card, addr, true); return NOTIFY_OK; case NETDEV_DOWN: qeth_l3_modify_ip(card, addr, false); return NOTIFY_OK; default: return NOTIFY_DONE; } } struct qeth_l3_ip_event_work { struct work_struct work; struct qeth_card *card; struct qeth_ipaddr addr; }; #define to_ip_work(w) container_of((w), struct qeth_l3_ip_event_work, work) static void qeth_l3_add_ip_worker(struct work_struct *work) { struct qeth_l3_ip_event_work *ip_work = to_ip_work(work); qeth_l3_modify_ip(ip_work->card, &ip_work->addr, true); kfree(work); } static void qeth_l3_delete_ip_worker(struct work_struct *work) { struct qeth_l3_ip_event_work *ip_work = to_ip_work(work); qeth_l3_modify_ip(ip_work->card, &ip_work->addr, false); kfree(work); } static struct qeth_card *qeth_l3_get_card_from_dev(struct net_device *dev) { if (is_vlan_dev(dev)) dev = vlan_dev_real_dev(dev); if (dev->netdev_ops == &qeth_l3_osa_netdev_ops || dev->netdev_ops == &qeth_l3_netdev_ops) return (struct qeth_card *) dev->ml_priv; return NULL; } static int qeth_l3_ip_event(struct notifier_block *this, unsigned long event, void *ptr) { struct in_ifaddr *ifa = (struct in_ifaddr *)ptr; struct net_device *dev = ifa->ifa_dev->dev; struct qeth_ipaddr addr; struct qeth_card *card; if (dev_net(dev) != &init_net) return NOTIFY_DONE; card = qeth_l3_get_card_from_dev(dev); if (!card) return NOTIFY_DONE; QETH_CARD_TEXT(card, 3, "ipevent"); qeth_l3_init_ipaddr(&addr, QETH_IP_TYPE_NORMAL, QETH_PROT_IPV4); addr.u.a4.addr = be32_to_cpu(ifa->ifa_address); addr.u.a4.mask = be32_to_cpu(ifa->ifa_mask); return qeth_l3_handle_ip_event(card, &addr, event); } static struct notifier_block qeth_l3_ip_notifier = { qeth_l3_ip_event, NULL, }; static int qeth_l3_ip6_event(struct notifier_block *this, unsigned long event, void *ptr) { struct inet6_ifaddr *ifa = (struct inet6_ifaddr *)ptr; struct net_device *dev = ifa->idev->dev; struct qeth_l3_ip_event_work *ip_work; struct qeth_card *card; if (event != NETDEV_UP && event != NETDEV_DOWN) return NOTIFY_DONE; card = qeth_l3_get_card_from_dev(dev); if (!card) return NOTIFY_DONE; QETH_CARD_TEXT(card, 3, "ip6event"); if (!qeth_is_supported(card, IPA_IPV6)) return NOTIFY_DONE; ip_work = kmalloc(sizeof(*ip_work), GFP_ATOMIC); if (!ip_work) return NOTIFY_DONE; if (event == NETDEV_UP) INIT_WORK(&ip_work->work, qeth_l3_add_ip_worker); else INIT_WORK(&ip_work->work, qeth_l3_delete_ip_worker); ip_work->card = card; qeth_l3_init_ipaddr(&ip_work->addr, QETH_IP_TYPE_NORMAL, QETH_PROT_IPV6); ip_work->addr.u.a6.addr = ifa->addr; ip_work->addr.u.a6.pfxlen = ifa->prefix_len; queue_work(card->cmd_wq, &ip_work->work); return NOTIFY_OK; } static struct notifier_block qeth_l3_ip6_notifier = { qeth_l3_ip6_event, NULL, }; static int qeth_l3_register_notifiers(void) { int rc; QETH_DBF_TEXT(SETUP, 5, "regnotif"); rc = register_inetaddr_notifier(&qeth_l3_ip_notifier); if (rc) return rc; rc = register_inet6addr_notifier(&qeth_l3_ip6_notifier); if (rc) { unregister_inetaddr_notifier(&qeth_l3_ip_notifier); return rc; } return 0; } static void qeth_l3_unregister_notifiers(void) { QETH_DBF_TEXT(SETUP, 5, "unregnot"); WARN_ON(unregister_inetaddr_notifier(&qeth_l3_ip_notifier)); WARN_ON(unregister_inet6addr_notifier(&qeth_l3_ip6_notifier)); } static int __init qeth_l3_init(void) { pr_info("register layer 3 discipline\n"); return qeth_l3_register_notifiers(); } static void __exit qeth_l3_exit(void) { qeth_l3_unregister_notifiers(); pr_info("unregister layer 3 discipline\n"); } module_init(qeth_l3_init); module_exit(qeth_l3_exit); MODULE_AUTHOR("Frank Blaschka <frank.blaschka@de.ibm.com>"); MODULE_DESCRIPTION("qeth layer 3 discipline"); MODULE_LICENSE("GPL");
Information contained on this website is for historical information purposes only and does not indicate or represent copyright ownership.
Created with Cregit http://github.com/cregit/cregit
Version 2.0-RC1