Author | Tokens | Token Proportion | Commits | Commit Proportion |
---|---|---|---|---|
Julian Wiedmann | 3494 | 29.92% | 90 | 48.39% |
Alexandra Winter | 2543 | 21.78% | 20 | 10.75% |
Eugene Crosser | 2102 | 18.00% | 12 | 6.45% |
Frank Blaschka | 1426 | 12.21% | 10 | 5.38% |
Hans Wippel | 1285 | 11.00% | 4 | 2.15% |
Lakhvich Dmitriy | 216 | 1.85% | 2 | 1.08% |
Carsten Otte | 205 | 1.76% | 1 | 0.54% |
Einar Lueck | 107 | 0.92% | 2 | 1.08% |
Thomas Richter | 80 | 0.69% | 4 | 2.15% |
Kittipon Meesompop | 50 | 0.43% | 3 | 1.61% |
Stefan Raspl | 24 | 0.21% | 2 | 1.08% |
Ursula Braun-Krahl | 20 | 0.17% | 6 | 3.23% |
Heiko Carstens | 16 | 0.14% | 3 | 1.61% |
Martin Schwidefsky | 14 | 0.12% | 2 | 1.08% |
Jiri Pirko | 14 | 0.12% | 3 | 1.61% |
Thorsten Winkler | 12 | 0.10% | 1 | 0.54% |
Kees Cook | 11 | 0.09% | 1 | 0.54% |
Patrick McHardy | 9 | 0.08% | 2 | 1.08% |
Sebastian Ott | 7 | 0.06% | 2 | 1.08% |
Arnd Bergmann | 6 | 0.05% | 2 | 1.08% |
Michael Ernst | 6 | 0.05% | 1 | 0.54% |
Michael Müller | 6 | 0.05% | 1 | 0.54% |
Vladimir Oltean | 4 | 0.03% | 1 | 0.54% |
Jakub Kiciński | 4 | 0.03% | 2 | 1.08% |
Lennert Buytenhek | 3 | 0.03% | 1 | 0.54% |
Tejun Heo | 3 | 0.03% | 1 | 0.54% |
Vasily Gorbik | 2 | 0.02% | 1 | 0.54% |
Cornelia Huck | 2 | 0.02% | 2 | 1.08% |
Peter Tiedemann | 2 | 0.02% | 1 | 0.54% |
Peter Oberparleiter | 2 | 0.02% | 1 | 0.54% |
Gustavo A. R. Silva | 2 | 0.02% | 1 | 0.54% |
Greg Kroah-Hartman | 1 | 0.01% | 1 | 0.54% |
Total | 11678 | 186 |
// SPDX-License-Identifier: GPL-2.0 /* * Copyright IBM Corp. 2007, 2009 * Author(s): Utz Bacher <utz.bacher@de.ibm.com>, * Frank Pavlic <fpavlic@de.ibm.com>, * Thomas Spatzier <tspat@de.ibm.com>, * Frank Blaschka <frank.blaschka@de.ibm.com> */ #define KMSG_COMPONENT "qeth" #define pr_fmt(fmt) KMSG_COMPONENT ": " fmt #include <linux/module.h> #include <linux/moduleparam.h> #include <linux/string.h> #include <linux/errno.h> #include <linux/kernel.h> #include <linux/slab.h> #include <linux/etherdevice.h> #include <linux/if_bridge.h> #include <linux/list.h> #include <linux/hash.h> #include <linux/hashtable.h> #include <net/switchdev.h> #include <asm/chsc.h> #include <asm/css_chars.h> #include <asm/setup.h> #include "qeth_core.h" #include "qeth_l2.h" static int qeth_l2_setdelmac_makerc(struct qeth_card *card, u16 retcode) { int rc; if (retcode) QETH_CARD_TEXT_(card, 2, "err%04x", retcode); switch (retcode) { case IPA_RC_SUCCESS: rc = 0; break; case IPA_RC_L2_UNSUPPORTED_CMD: rc = -EOPNOTSUPP; break; case IPA_RC_L2_ADDR_TABLE_FULL: rc = -ENOSPC; break; case IPA_RC_L2_DUP_MAC: case IPA_RC_L2_DUP_LAYER3_MAC: rc = -EADDRINUSE; break; case IPA_RC_L2_MAC_NOT_AUTH_BY_HYP: case IPA_RC_L2_MAC_NOT_AUTH_BY_ADP: rc = -EADDRNOTAVAIL; break; case IPA_RC_L2_MAC_NOT_FOUND: rc = -ENOENT; break; default: rc = -EIO; break; } return rc; } static int qeth_l2_send_setdelmac_cb(struct qeth_card *card, struct qeth_reply *reply, unsigned long data) { struct qeth_ipa_cmd *cmd = (struct qeth_ipa_cmd *) data; return qeth_l2_setdelmac_makerc(card, cmd->hdr.return_code); } static int qeth_l2_send_setdelmac(struct qeth_card *card, const __u8 *mac, enum qeth_ipa_cmds ipacmd) { struct qeth_ipa_cmd *cmd; struct qeth_cmd_buffer *iob; QETH_CARD_TEXT(card, 2, "L2sdmac"); iob = qeth_ipa_alloc_cmd(card, ipacmd, QETH_PROT_IPV4, IPA_DATA_SIZEOF(setdelmac)); if (!iob) return -ENOMEM; cmd = __ipa_cmd(iob); cmd->data.setdelmac.mac_length = ETH_ALEN; ether_addr_copy(cmd->data.setdelmac.mac, mac); return qeth_send_ipa_cmd(card, iob, qeth_l2_send_setdelmac_cb, NULL); } static int qeth_l2_send_setmac(struct qeth_card *card, const __u8 *mac) { int rc; QETH_CARD_TEXT(card, 2, "L2Setmac"); rc = qeth_l2_send_setdelmac(card, mac, IPA_CMD_SETVMAC); if (rc == 0) { dev_info(&card->gdev->dev, "MAC address %pM successfully registered\n", mac); } else { switch (rc) { case -EADDRINUSE: dev_warn(&card->gdev->dev, "MAC address %pM already exists\n", mac); break; case -EADDRNOTAVAIL: dev_warn(&card->gdev->dev, "MAC address %pM is not authorized\n", mac); break; } } return rc; } static int qeth_l2_write_mac(struct qeth_card *card, u8 *mac) { enum qeth_ipa_cmds cmd = is_multicast_ether_addr(mac) ? IPA_CMD_SETGMAC : IPA_CMD_SETVMAC; int rc; QETH_CARD_TEXT(card, 2, "L2Wmac"); rc = qeth_l2_send_setdelmac(card, mac, cmd); if (rc == -EADDRINUSE) QETH_DBF_MESSAGE(2, "MAC address %012llx is already registered on device %x\n", ether_addr_to_u64(mac), CARD_DEVID(card)); else if (rc) QETH_DBF_MESSAGE(2, "Failed to register MAC address %012llx on device %x: %d\n", ether_addr_to_u64(mac), CARD_DEVID(card), rc); return rc; } static int qeth_l2_remove_mac(struct qeth_card *card, u8 *mac) { enum qeth_ipa_cmds cmd = is_multicast_ether_addr(mac) ? IPA_CMD_DELGMAC : IPA_CMD_DELVMAC; int rc; QETH_CARD_TEXT(card, 2, "L2Rmac"); rc = qeth_l2_send_setdelmac(card, mac, cmd); if (rc) QETH_DBF_MESSAGE(2, "Failed to delete MAC address %012llx on device %x: %d\n", ether_addr_to_u64(mac), CARD_DEVID(card), rc); return rc; } static void qeth_l2_drain_rx_mode_cache(struct qeth_card *card) { struct qeth_mac *mac; struct hlist_node *tmp; int i; hash_for_each_safe(card->rx_mode_addrs, i, tmp, mac, hnode) { hash_del(&mac->hnode); kfree(mac); } } static void qeth_l2_fill_header(struct qeth_qdio_out_q *queue, struct qeth_hdr *hdr, struct sk_buff *skb, __be16 proto, unsigned int data_len) { int cast_type = qeth_get_ether_cast_type(skb); struct vlan_ethhdr *veth = vlan_eth_hdr(skb); hdr->hdr.l2.pkt_length = data_len; if (skb_is_gso(skb)) { hdr->hdr.l2.id = QETH_HEADER_TYPE_L2_TSO; } else { hdr->hdr.l2.id = QETH_HEADER_TYPE_LAYER2; if (skb->ip_summed == CHECKSUM_PARTIAL) qeth_tx_csum(skb, &hdr->hdr.l2.flags[1], proto); } /* set byte byte 3 to casting flags */ if (cast_type == RTN_MULTICAST) hdr->hdr.l2.flags[2] |= QETH_LAYER2_FLAG_MULTICAST; else if (cast_type == RTN_BROADCAST) hdr->hdr.l2.flags[2] |= QETH_LAYER2_FLAG_BROADCAST; else hdr->hdr.l2.flags[2] |= QETH_LAYER2_FLAG_UNICAST; /* VSWITCH relies on the VLAN * information to be present in * the QDIO header */ if (veth->h_vlan_proto == htons(ETH_P_8021Q)) { hdr->hdr.l2.flags[2] |= QETH_LAYER2_FLAG_VLAN; hdr->hdr.l2.vlan_id = ntohs(veth->h_vlan_TCI); } } static int qeth_l2_setdelvlan_makerc(struct qeth_card *card, u16 retcode) { if (retcode) QETH_CARD_TEXT_(card, 2, "err%04x", retcode); switch (retcode) { case IPA_RC_SUCCESS: return 0; case IPA_RC_L2_INVALID_VLAN_ID: return -EINVAL; case IPA_RC_L2_DUP_VLAN_ID: return -EEXIST; case IPA_RC_L2_VLAN_ID_NOT_FOUND: return -ENOENT; case IPA_RC_L2_VLAN_ID_NOT_ALLOWED: return -EPERM; default: return -EIO; } } static int qeth_l2_send_setdelvlan_cb(struct qeth_card *card, struct qeth_reply *reply, unsigned long data) { struct qeth_ipa_cmd *cmd = (struct qeth_ipa_cmd *) data; QETH_CARD_TEXT(card, 2, "L2sdvcb"); if (cmd->hdr.return_code) { QETH_DBF_MESSAGE(2, "Error in processing VLAN %u on device %x: %#x.\n", cmd->data.setdelvlan.vlan_id, CARD_DEVID(card), cmd->hdr.return_code); QETH_CARD_TEXT_(card, 2, "L2VL%4x", cmd->hdr.command); } return qeth_l2_setdelvlan_makerc(card, cmd->hdr.return_code); } static int qeth_l2_send_setdelvlan(struct qeth_card *card, __u16 i, enum qeth_ipa_cmds ipacmd) { struct qeth_ipa_cmd *cmd; struct qeth_cmd_buffer *iob; QETH_CARD_TEXT_(card, 4, "L2sdv%x", ipacmd); iob = qeth_ipa_alloc_cmd(card, ipacmd, QETH_PROT_IPV4, IPA_DATA_SIZEOF(setdelvlan)); if (!iob) return -ENOMEM; cmd = __ipa_cmd(iob); cmd->data.setdelvlan.vlan_id = i; return qeth_send_ipa_cmd(card, iob, qeth_l2_send_setdelvlan_cb, NULL); } static int qeth_l2_vlan_rx_add_vid(struct net_device *dev, __be16 proto, u16 vid) { struct qeth_card *card = dev->ml_priv; QETH_CARD_TEXT_(card, 4, "aid:%d", vid); if (!vid) return 0; return qeth_l2_send_setdelvlan(card, vid, IPA_CMD_SETVLAN); } static int qeth_l2_vlan_rx_kill_vid(struct net_device *dev, __be16 proto, u16 vid) { struct qeth_card *card = dev->ml_priv; QETH_CARD_TEXT_(card, 4, "kid:%d", vid); if (!vid) return 0; return qeth_l2_send_setdelvlan(card, vid, IPA_CMD_DELVLAN); } static void qeth_l2_set_pnso_mode(struct qeth_card *card, enum qeth_pnso_mode mode) { spin_lock_irq(get_ccwdev_lock(CARD_RDEV(card))); WRITE_ONCE(card->info.pnso_mode, mode); spin_unlock_irq(get_ccwdev_lock(CARD_RDEV(card))); if (mode == QETH_PNSO_NONE) drain_workqueue(card->event_wq); } static void qeth_l2_dev2br_fdb_flush(struct qeth_card *card) { struct switchdev_notifier_fdb_info info = {}; QETH_CARD_TEXT(card, 2, "fdbflush"); info.addr = NULL; /* flush all VLANs: */ info.vid = 0; info.added_by_user = false; info.offloaded = true; call_switchdev_notifiers(SWITCHDEV_FDB_FLUSH_TO_BRIDGE, card->dev, &info.info, NULL); } static int qeth_l2_request_initial_mac(struct qeth_card *card) { int rc = 0; QETH_CARD_TEXT(card, 2, "l2reqmac"); if (MACHINE_IS_VM) { rc = qeth_vm_request_mac(card); if (!rc) goto out; QETH_DBF_MESSAGE(2, "z/VM MAC Service failed on device %x: %#x\n", CARD_DEVID(card), rc); QETH_CARD_TEXT_(card, 2, "err%04x", rc); /* fall back to alternative mechanism: */ } rc = qeth_setadpparms_change_macaddr(card); if (!rc) goto out; QETH_DBF_MESSAGE(2, "READ_MAC Assist failed on device %x: %#x\n", CARD_DEVID(card), rc); QETH_CARD_TEXT_(card, 2, "1err%04x", rc); /* Fall back once more, but some devices don't support a custom MAC * address: */ if (IS_OSM(card) || IS_OSX(card)) return (rc) ? rc : -EADDRNOTAVAIL; eth_hw_addr_random(card->dev); out: QETH_CARD_HEX(card, 2, card->dev->dev_addr, card->dev->addr_len); return 0; } static void qeth_l2_register_dev_addr(struct qeth_card *card) { if (!is_valid_ether_addr(card->dev->dev_addr)) qeth_l2_request_initial_mac(card); if (!qeth_l2_send_setmac(card, card->dev->dev_addr)) card->info.dev_addr_is_registered = 1; else card->info.dev_addr_is_registered = 0; } static int qeth_l2_validate_addr(struct net_device *dev) { struct qeth_card *card = dev->ml_priv; if (card->info.dev_addr_is_registered) return eth_validate_addr(dev); QETH_CARD_TEXT(card, 4, "nomacadr"); return -EPERM; } static int qeth_l2_set_mac_address(struct net_device *dev, void *p) { struct sockaddr *addr = p; struct qeth_card *card = dev->ml_priv; u8 old_addr[ETH_ALEN]; int rc = 0; QETH_CARD_TEXT(card, 3, "setmac"); if (IS_OSM(card) || IS_OSX(card)) { QETH_CARD_TEXT(card, 3, "setmcTYP"); return -EOPNOTSUPP; } QETH_CARD_HEX(card, 3, addr->sa_data, ETH_ALEN); if (!is_valid_ether_addr(addr->sa_data)) return -EADDRNOTAVAIL; /* don't register the same address twice */ if (ether_addr_equal_64bits(dev->dev_addr, addr->sa_data) && card->info.dev_addr_is_registered) return 0; /* add the new address, switch over, drop the old */ rc = qeth_l2_send_setmac(card, addr->sa_data); if (rc) return rc; ether_addr_copy(old_addr, dev->dev_addr); eth_hw_addr_set(dev, addr->sa_data); if (card->info.dev_addr_is_registered) qeth_l2_remove_mac(card, old_addr); card->info.dev_addr_is_registered = 1; return 0; } static void qeth_l2_promisc_to_bridge(struct qeth_card *card, bool enable) { int role; int rc; QETH_CARD_TEXT(card, 3, "pmisc2br"); if (enable) { if (card->options.sbp.reflect_promisc_primary) role = QETH_SBP_ROLE_PRIMARY; else role = QETH_SBP_ROLE_SECONDARY; } else role = QETH_SBP_ROLE_NONE; rc = qeth_bridgeport_setrole(card, role); QETH_CARD_TEXT_(card, 2, "bpm%c%04x", enable ? '+' : '-', rc); if (!rc) { card->options.sbp.role = role; card->info.promisc_mode = enable; } } static void qeth_l2_set_promisc_mode(struct qeth_card *card) { bool enable = card->dev->flags & IFF_PROMISC; if (card->info.promisc_mode == enable) return; if (qeth_adp_supported(card, IPA_SETADP_SET_PROMISC_MODE)) { qeth_setadp_promisc_mode(card, enable); } else { mutex_lock(&card->sbp_lock); if (card->options.sbp.reflect_promisc) qeth_l2_promisc_to_bridge(card, enable); mutex_unlock(&card->sbp_lock); } } /* New MAC address is added to the hash table and marked to be written on card * only if there is not in the hash table storage already * */ static void qeth_l2_add_mac(struct qeth_card *card, struct netdev_hw_addr *ha) { u32 mac_hash = get_unaligned((u32 *)(&ha->addr[2])); struct qeth_mac *mac; hash_for_each_possible(card->rx_mode_addrs, mac, hnode, mac_hash) { if (ether_addr_equal_64bits(ha->addr, mac->mac_addr)) { mac->disp_flag = QETH_DISP_ADDR_DO_NOTHING; return; } } mac = kzalloc(sizeof(struct qeth_mac), GFP_ATOMIC); if (!mac) return; ether_addr_copy(mac->mac_addr, ha->addr); mac->disp_flag = QETH_DISP_ADDR_ADD; hash_add(card->rx_mode_addrs, &mac->hnode, mac_hash); } static void qeth_l2_rx_mode_work(struct work_struct *work) { struct qeth_card *card = container_of(work, struct qeth_card, rx_mode_work); struct net_device *dev = card->dev; struct netdev_hw_addr *ha; struct qeth_mac *mac; struct hlist_node *tmp; int i; int rc; QETH_CARD_TEXT(card, 3, "setmulti"); netif_addr_lock_bh(dev); netdev_for_each_mc_addr(ha, dev) qeth_l2_add_mac(card, ha); netdev_for_each_uc_addr(ha, dev) qeth_l2_add_mac(card, ha); netif_addr_unlock_bh(dev); hash_for_each_safe(card->rx_mode_addrs, i, tmp, mac, hnode) { switch (mac->disp_flag) { case QETH_DISP_ADDR_DELETE: qeth_l2_remove_mac(card, mac->mac_addr); hash_del(&mac->hnode); kfree(mac); break; case QETH_DISP_ADDR_ADD: rc = qeth_l2_write_mac(card, mac->mac_addr); if (rc) { hash_del(&mac->hnode); kfree(mac); break; } fallthrough; default: /* for next call to set_rx_mode(): */ mac->disp_flag = QETH_DISP_ADDR_DELETE; } } qeth_l2_set_promisc_mode(card); } static netdev_tx_t qeth_l2_hard_start_xmit(struct sk_buff *skb, struct net_device *dev) { struct qeth_card *card = dev->ml_priv; u16 txq = skb_get_queue_mapping(skb); struct qeth_qdio_out_q *queue; int rc; if (!skb_is_gso(skb)) qdisc_skb_cb(skb)->pkt_len = skb->len; if (IS_IQD(card)) txq = qeth_iqd_translate_txq(dev, txq); queue = card->qdio.out_qs[txq]; rc = qeth_xmit(card, skb, queue, vlan_get_protocol(skb), qeth_l2_fill_header); if (!rc) return NETDEV_TX_OK; QETH_TXQ_STAT_INC(queue, tx_dropped); kfree_skb(skb); return NETDEV_TX_OK; } static u16 qeth_l2_iqd_select_queue(struct net_device *dev, struct sk_buff *skb, struct net_device *sb_dev) { return qeth_iqd_select_queue(dev, skb, qeth_get_ether_cast_type(skb), sb_dev); } static void qeth_l2_set_rx_mode(struct net_device *dev) { struct qeth_card *card = dev->ml_priv; schedule_work(&card->rx_mode_work); } /** * qeth_l2_pnso() - perform network subchannel operation * @card: qeth_card structure pointer * @oc: Operation Code * @cnc: Boolean Change-Notification Control * @cb: Callback function will be executed for each element * of the address list * @priv: Pointer to pass to the callback function. * * Collects network information in a network address list and calls the * callback function for every entry in the list. If "change-notification- * control" is set, further changes in the address list will be reported * via the IPA command. */ static int qeth_l2_pnso(struct qeth_card *card, u8 oc, int cnc, void (*cb)(void *priv, struct chsc_pnso_naid_l2 *entry), void *priv) { struct ccw_device *ddev = CARD_DDEV(card); struct chsc_pnso_area *rr; u32 prev_instance = 0; int isfirstblock = 1; int i, size, elems; int rc; rr = (struct chsc_pnso_area *)get_zeroed_page(GFP_KERNEL); if (rr == NULL) return -ENOMEM; do { QETH_CARD_TEXT(card, 2, "PNSO"); /* on the first iteration, naihdr.resume_token will be zero */ rc = ccw_device_pnso(ddev, rr, oc, rr->naihdr.resume_token, cnc); if (rc) continue; if (cb == NULL) continue; size = rr->naihdr.naids; if (size != sizeof(struct chsc_pnso_naid_l2)) { WARN_ON_ONCE(1); continue; } elems = (rr->response.length - sizeof(struct chsc_header) - sizeof(struct chsc_pnso_naihdr)) / size; if (!isfirstblock && (rr->naihdr.instance != prev_instance)) { /* Inform the caller that they need to scrap */ /* the data that was already reported via cb */ rc = -EAGAIN; break; } isfirstblock = 0; prev_instance = rr->naihdr.instance; for (i = 0; i < elems; i++) (*cb)(priv, &rr->entries[i]); } while ((rc == -EBUSY) || (!rc && /* list stored */ /* resume token is non-zero => list incomplete */ (rr->naihdr.resume_token.t1 || rr->naihdr.resume_token.t2))); if (rc) QETH_CARD_TEXT_(card, 2, "PNrp%04x", rr->response.code); free_page((unsigned long)rr); return rc; } static bool qeth_is_my_net_if_token(struct qeth_card *card, struct net_if_token *token) { return ((card->info.ddev_devno == token->devnum) && (card->info.cssid == token->cssid) && (card->info.iid == token->iid) && (card->info.ssid == token->ssid) && (card->info.chpid == token->chpid) && (card->info.chid == token->chid)); } /** * qeth_l2_dev2br_fdb_notify() - update fdb of master bridge * @card: qeth_card structure pointer * @code: event bitmask: high order bit 0x80 set to * 1 - removal of an object * 0 - addition of an object * Object type(s): * 0x01 - VLAN, 0x02 - MAC, 0x03 - VLAN and MAC * @token: "network token" structure identifying 'physical' location * of the target * @addr_lnid: structure with MAC address and VLAN ID of the target */ static void qeth_l2_dev2br_fdb_notify(struct qeth_card *card, u8 code, struct net_if_token *token, struct mac_addr_lnid *addr_lnid) { struct switchdev_notifier_fdb_info info = {}; u8 ntfy_mac[ETH_ALEN]; ether_addr_copy(ntfy_mac, addr_lnid->mac); /* Ignore VLAN only changes */ if (!(code & IPA_ADDR_CHANGE_CODE_MACADDR)) return; /* Ignore mcast entries */ if (is_multicast_ether_addr(ntfy_mac)) return; /* Ignore my own addresses */ if (qeth_is_my_net_if_token(card, token)) return; info.addr = ntfy_mac; /* don't report VLAN IDs */ info.vid = 0; info.added_by_user = false; info.offloaded = true; if (code & IPA_ADDR_CHANGE_CODE_REMOVAL) { call_switchdev_notifiers(SWITCHDEV_FDB_DEL_TO_BRIDGE, card->dev, &info.info, NULL); QETH_CARD_TEXT(card, 4, "andelmac"); QETH_CARD_TEXT_(card, 4, "mc%012llx", ether_addr_to_u64(ntfy_mac)); } else { call_switchdev_notifiers(SWITCHDEV_FDB_ADD_TO_BRIDGE, card->dev, &info.info, NULL); QETH_CARD_TEXT(card, 4, "anaddmac"); QETH_CARD_TEXT_(card, 4, "mc%012llx", ether_addr_to_u64(ntfy_mac)); } } static void qeth_l2_dev2br_an_set_cb(void *priv, struct chsc_pnso_naid_l2 *entry) { u8 code = IPA_ADDR_CHANGE_CODE_MACADDR; struct qeth_card *card = priv; if (entry->addr_lnid.lnid < VLAN_N_VID) code |= IPA_ADDR_CHANGE_CODE_VLANID; qeth_l2_dev2br_fdb_notify(card, code, (struct net_if_token *)&entry->nit, (struct mac_addr_lnid *)&entry->addr_lnid); } /** * qeth_l2_dev2br_an_set() - * Enable or disable 'dev to bridge network address notification' * @card: qeth_card structure pointer * @enable: Enable or disable 'dev to bridge network address notification' * * Returns negative errno-compatible error indication or 0 on success. * * On enable, emits a series of address notifications for all * currently registered hosts. */ static int qeth_l2_dev2br_an_set(struct qeth_card *card, bool enable) { int rc; if (enable) { QETH_CARD_TEXT(card, 2, "anseton"); rc = qeth_l2_pnso(card, PNSO_OC_NET_ADDR_INFO, 1, qeth_l2_dev2br_an_set_cb, card); if (rc == -EAGAIN) /* address notification enabled, but inconsistent * addresses reported -> disable address notification */ qeth_l2_pnso(card, PNSO_OC_NET_ADDR_INFO, 0, NULL, NULL); } else { QETH_CARD_TEXT(card, 2, "ansetoff"); rc = qeth_l2_pnso(card, PNSO_OC_NET_ADDR_INFO, 0, NULL, NULL); } return rc; } struct qeth_l2_br2dev_event_work { struct work_struct work; struct net_device *br_dev; struct net_device *lsync_dev; struct net_device *dst_dev; unsigned long event; unsigned char addr[ETH_ALEN]; }; static const struct net_device_ops qeth_l2_iqd_netdev_ops; static const struct net_device_ops qeth_l2_osa_netdev_ops; static bool qeth_l2_must_learn(struct net_device *netdev, struct net_device *dstdev) { struct qeth_priv *priv; priv = netdev_priv(netdev); return (netdev != dstdev && (priv->brport_features & BR_LEARNING_SYNC) && !(br_port_flag_is_set(netdev, BR_ISOLATED) && br_port_flag_is_set(dstdev, BR_ISOLATED)) && (netdev->netdev_ops == &qeth_l2_iqd_netdev_ops || netdev->netdev_ops == &qeth_l2_osa_netdev_ops)); } /** * qeth_l2_br2dev_worker() - update local MACs * @work: bridge to device FDB update * * Update local MACs of a learning_sync bridgeport so it can receive * messages for a destination port. * In case of an isolated learning_sync port, also update its isolated * siblings. */ static void qeth_l2_br2dev_worker(struct work_struct *work) { struct qeth_l2_br2dev_event_work *br2dev_event_work = container_of(work, struct qeth_l2_br2dev_event_work, work); struct net_device *lsyncdev = br2dev_event_work->lsync_dev; struct net_device *dstdev = br2dev_event_work->dst_dev; struct net_device *brdev = br2dev_event_work->br_dev; unsigned long event = br2dev_event_work->event; unsigned char *addr = br2dev_event_work->addr; struct qeth_card *card = lsyncdev->ml_priv; struct net_device *lowerdev; struct list_head *iter; int err = 0; QETH_CARD_TEXT_(card, 4, "b2dw%04lx", event); QETH_CARD_TEXT_(card, 4, "ma%012llx", ether_addr_to_u64(addr)); rcu_read_lock(); /* Verify preconditions are still valid: */ if (!netif_is_bridge_port(lsyncdev) || brdev != netdev_master_upper_dev_get_rcu(lsyncdev)) goto unlock; if (!qeth_l2_must_learn(lsyncdev, dstdev)) goto unlock; if (br_port_flag_is_set(lsyncdev, BR_ISOLATED)) { /* Update lsyncdev and its isolated sibling(s): */ iter = &brdev->adj_list.lower; lowerdev = netdev_next_lower_dev_rcu(brdev, &iter); while (lowerdev) { if (br_port_flag_is_set(lowerdev, BR_ISOLATED)) { switch (event) { case SWITCHDEV_FDB_ADD_TO_DEVICE: err = dev_uc_add(lowerdev, addr); break; case SWITCHDEV_FDB_DEL_TO_DEVICE: err = dev_uc_del(lowerdev, addr); break; default: break; } if (err) { QETH_CARD_TEXT(card, 2, "b2derris"); QETH_CARD_TEXT_(card, 2, "err%02lx%03d", event, lowerdev->ifindex); } } lowerdev = netdev_next_lower_dev_rcu(brdev, &iter); } } else { switch (event) { case SWITCHDEV_FDB_ADD_TO_DEVICE: err = dev_uc_add(lsyncdev, addr); break; case SWITCHDEV_FDB_DEL_TO_DEVICE: err = dev_uc_del(lsyncdev, addr); break; default: break; } if (err) QETH_CARD_TEXT_(card, 2, "b2derr%02lx", event); } unlock: rcu_read_unlock(); dev_put(brdev); dev_put(lsyncdev); dev_put(dstdev); kfree(br2dev_event_work); } static int qeth_l2_br2dev_queue_work(struct net_device *brdev, struct net_device *lsyncdev, struct net_device *dstdev, unsigned long event, const unsigned char *addr) { struct qeth_l2_br2dev_event_work *worker_data; struct qeth_card *card; worker_data = kzalloc(sizeof(*worker_data), GFP_ATOMIC); if (!worker_data) return -ENOMEM; INIT_WORK(&worker_data->work, qeth_l2_br2dev_worker); worker_data->br_dev = brdev; worker_data->lsync_dev = lsyncdev; worker_data->dst_dev = dstdev; worker_data->event = event; ether_addr_copy(worker_data->addr, addr); card = lsyncdev->ml_priv; /* Take a reference on the sw port devices and the bridge */ dev_hold(brdev); dev_hold(lsyncdev); dev_hold(dstdev); queue_work(card->event_wq, &worker_data->work); return 0; } /* Called under rtnl_lock */ static int qeth_l2_switchdev_event(struct notifier_block *unused, unsigned long event, void *ptr) { struct net_device *dstdev, *brdev, *lowerdev; struct switchdev_notifier_fdb_info *fdb_info; struct switchdev_notifier_info *info = ptr; struct list_head *iter; struct qeth_card *card; int rc; if (!(event == SWITCHDEV_FDB_ADD_TO_DEVICE || event == SWITCHDEV_FDB_DEL_TO_DEVICE)) return NOTIFY_DONE; dstdev = switchdev_notifier_info_to_dev(info); brdev = netdev_master_upper_dev_get_rcu(dstdev); if (!brdev || !netif_is_bridge_master(brdev)) return NOTIFY_DONE; fdb_info = container_of(info, struct switchdev_notifier_fdb_info, info); iter = &brdev->adj_list.lower; lowerdev = netdev_next_lower_dev_rcu(brdev, &iter); while (lowerdev) { if (qeth_l2_must_learn(lowerdev, dstdev)) { card = lowerdev->ml_priv; QETH_CARD_TEXT_(card, 4, "b2dqw%03lx", event); rc = qeth_l2_br2dev_queue_work(brdev, lowerdev, dstdev, event, fdb_info->addr); if (rc) { QETH_CARD_TEXT(card, 2, "b2dqwerr"); return NOTIFY_BAD; } } lowerdev = netdev_next_lower_dev_rcu(brdev, &iter); } return NOTIFY_DONE; } static struct notifier_block qeth_l2_sw_notifier = { .notifier_call = qeth_l2_switchdev_event, }; static refcount_t qeth_l2_switchdev_notify_refcnt; /* Called under rtnl_lock */ static void qeth_l2_br2dev_get(void) { int rc; if (!refcount_inc_not_zero(&qeth_l2_switchdev_notify_refcnt)) { rc = register_switchdev_notifier(&qeth_l2_sw_notifier); if (rc) { QETH_DBF_MESSAGE(2, "failed to register qeth_l2_sw_notifier: %d\n", rc); } else { refcount_set(&qeth_l2_switchdev_notify_refcnt, 1); QETH_DBF_MESSAGE(2, "qeth_l2_sw_notifier registered\n"); } } QETH_DBF_TEXT_(SETUP, 2, "b2d+%04d", qeth_l2_switchdev_notify_refcnt.refs.counter); } /* Called under rtnl_lock */ static void qeth_l2_br2dev_put(void) { int rc; if (refcount_dec_and_test(&qeth_l2_switchdev_notify_refcnt)) { rc = unregister_switchdev_notifier(&qeth_l2_sw_notifier); if (rc) { QETH_DBF_MESSAGE(2, "failed to unregister qeth_l2_sw_notifier: %d\n", rc); } else { QETH_DBF_MESSAGE(2, "qeth_l2_sw_notifier unregistered\n"); } } QETH_DBF_TEXT_(SETUP, 2, "b2d-%04d", qeth_l2_switchdev_notify_refcnt.refs.counter); } static int qeth_l2_bridge_getlink(struct sk_buff *skb, u32 pid, u32 seq, struct net_device *dev, u32 filter_mask, int nlflags) { struct qeth_priv *priv = netdev_priv(dev); struct qeth_card *card = dev->ml_priv; u16 mode = BRIDGE_MODE_UNDEF; /* Do not even show qeth devs that cannot do bridge_setlink */ if (!priv->brport_hw_features || !netif_device_present(dev) || qeth_bridgeport_is_in_use(card)) return -EOPNOTSUPP; return ndo_dflt_bridge_getlink(skb, pid, seq, dev, mode, priv->brport_features, priv->brport_hw_features, nlflags, filter_mask, NULL); } static const struct nla_policy qeth_brport_policy[IFLA_BRPORT_MAX + 1] = { [IFLA_BRPORT_LEARNING_SYNC] = { .type = NLA_U8 }, }; /** * qeth_l2_bridge_setlink() - set bridgeport attributes * @dev: netdevice * @nlh: netlink message header * @flags: bridge flags (here: BRIDGE_FLAGS_SELF) * @extack: extended ACK report struct * * Called under rtnl_lock */ static int qeth_l2_bridge_setlink(struct net_device *dev, struct nlmsghdr *nlh, u16 flags, struct netlink_ext_ack *extack) { struct qeth_priv *priv = netdev_priv(dev); struct nlattr *bp_tb[IFLA_BRPORT_MAX + 1]; struct qeth_card *card = dev->ml_priv; struct nlattr *attr, *nested_attr; bool enable, has_protinfo = false; int rem1, rem2; int rc; if (!netif_device_present(dev)) return -ENODEV; nlmsg_for_each_attr(attr, nlh, sizeof(struct ifinfomsg), rem1) { if (nla_type(attr) == IFLA_PROTINFO) { rc = nla_parse_nested(bp_tb, IFLA_BRPORT_MAX, attr, qeth_brport_policy, extack); if (rc) return rc; has_protinfo = true; } else if (nla_type(attr) == IFLA_AF_SPEC) { nla_for_each_nested(nested_attr, attr, rem2) { if (nla_type(nested_attr) == IFLA_BRIDGE_FLAGS) continue; NL_SET_ERR_MSG_ATTR(extack, nested_attr, "Unsupported attribute"); return -EINVAL; } } else { NL_SET_ERR_MSG_ATTR(extack, attr, "Unsupported attribute"); return -EINVAL; } } if (!has_protinfo) return 0; if (!bp_tb[IFLA_BRPORT_LEARNING_SYNC]) return -EINVAL; if (!(priv->brport_hw_features & BR_LEARNING_SYNC)) { NL_SET_ERR_MSG_ATTR(extack, bp_tb[IFLA_BRPORT_LEARNING_SYNC], "Operation not supported by HW"); return -EOPNOTSUPP; } if (!IS_ENABLED(CONFIG_NET_SWITCHDEV)) { NL_SET_ERR_MSG_ATTR(extack, bp_tb[IFLA_BRPORT_LEARNING_SYNC], "Requires NET_SWITCHDEV"); return -EOPNOTSUPP; } enable = !!nla_get_u8(bp_tb[IFLA_BRPORT_LEARNING_SYNC]); if (enable == !!(priv->brport_features & BR_LEARNING_SYNC)) return 0; mutex_lock(&card->sbp_lock); /* do not change anything if BridgePort is enabled */ if (qeth_bridgeport_is_in_use(card)) { NL_SET_ERR_MSG(extack, "n/a (BridgePort)"); rc = -EBUSY; } else if (enable) { qeth_l2_set_pnso_mode(card, QETH_PNSO_ADDR_INFO); rc = qeth_l2_dev2br_an_set(card, true); if (rc) { qeth_l2_set_pnso_mode(card, QETH_PNSO_NONE); } else { priv->brport_features |= BR_LEARNING_SYNC; qeth_l2_br2dev_get(); } } else { rc = qeth_l2_dev2br_an_set(card, false); if (!rc) { qeth_l2_set_pnso_mode(card, QETH_PNSO_NONE); priv->brport_features ^= BR_LEARNING_SYNC; qeth_l2_dev2br_fdb_flush(card); qeth_l2_br2dev_put(); } } mutex_unlock(&card->sbp_lock); return rc; } static const struct net_device_ops qeth_l2_iqd_netdev_ops = { .ndo_open = qeth_open, .ndo_stop = qeth_stop, .ndo_get_stats64 = qeth_get_stats64, .ndo_start_xmit = qeth_l2_hard_start_xmit, .ndo_features_check = qeth_features_check, .ndo_select_queue = qeth_l2_iqd_select_queue, .ndo_validate_addr = qeth_l2_validate_addr, .ndo_set_rx_mode = qeth_l2_set_rx_mode, .ndo_eth_ioctl = qeth_do_ioctl, .ndo_siocdevprivate = qeth_siocdevprivate, .ndo_set_mac_address = qeth_l2_set_mac_address, .ndo_vlan_rx_add_vid = qeth_l2_vlan_rx_add_vid, .ndo_vlan_rx_kill_vid = qeth_l2_vlan_rx_kill_vid, .ndo_tx_timeout = qeth_tx_timeout, .ndo_fix_features = qeth_fix_features, .ndo_set_features = qeth_set_features, .ndo_bridge_getlink = qeth_l2_bridge_getlink, .ndo_bridge_setlink = qeth_l2_bridge_setlink, }; static const struct net_device_ops qeth_l2_osa_netdev_ops = { .ndo_open = qeth_open, .ndo_stop = qeth_stop, .ndo_get_stats64 = qeth_get_stats64, .ndo_start_xmit = qeth_l2_hard_start_xmit, .ndo_features_check = qeth_features_check, .ndo_select_queue = qeth_osa_select_queue, .ndo_validate_addr = qeth_l2_validate_addr, .ndo_set_rx_mode = qeth_l2_set_rx_mode, .ndo_eth_ioctl = qeth_do_ioctl, .ndo_siocdevprivate = qeth_siocdevprivate, .ndo_set_mac_address = qeth_l2_set_mac_address, .ndo_vlan_rx_add_vid = qeth_l2_vlan_rx_add_vid, .ndo_vlan_rx_kill_vid = qeth_l2_vlan_rx_kill_vid, .ndo_tx_timeout = qeth_tx_timeout, .ndo_fix_features = qeth_fix_features, .ndo_set_features = qeth_set_features, }; static int qeth_l2_setup_netdev(struct qeth_card *card) { card->dev->netdev_ops = IS_IQD(card) ? &qeth_l2_iqd_netdev_ops : &qeth_l2_osa_netdev_ops; card->dev->needed_headroom = sizeof(struct qeth_hdr); card->dev->priv_flags |= IFF_UNICAST_FLT; if (IS_OSM(card)) { card->dev->features |= NETIF_F_VLAN_CHALLENGED; } else { if (!IS_VM_NIC(card)) card->dev->hw_features |= NETIF_F_HW_VLAN_CTAG_FILTER; card->dev->features |= NETIF_F_HW_VLAN_CTAG_FILTER; } if (IS_OSD(card) && !IS_VM_NIC(card)) { card->dev->features |= NETIF_F_SG; /* OSA 3S and earlier has no RX/TX support */ if (qeth_is_supported(card, IPA_OUTBOUND_CHECKSUM)) { card->dev->hw_features |= NETIF_F_IP_CSUM; card->dev->vlan_features |= NETIF_F_IP_CSUM; } } if (qeth_is_supported6(card, IPA_OUTBOUND_CHECKSUM_V6)) { card->dev->hw_features |= NETIF_F_IPV6_CSUM; card->dev->vlan_features |= NETIF_F_IPV6_CSUM; } if (qeth_is_supported(card, IPA_INBOUND_CHECKSUM) || qeth_is_supported6(card, IPA_INBOUND_CHECKSUM_V6)) { card->dev->hw_features |= NETIF_F_RXCSUM; card->dev->vlan_features |= NETIF_F_RXCSUM; } if (qeth_is_supported(card, IPA_OUTBOUND_TSO)) { card->dev->hw_features |= NETIF_F_TSO; card->dev->vlan_features |= NETIF_F_TSO; } if (qeth_is_supported6(card, IPA_OUTBOUND_TSO)) { card->dev->hw_features |= NETIF_F_TSO6; card->dev->vlan_features |= NETIF_F_TSO6; } if (card->dev->hw_features & (NETIF_F_TSO | NETIF_F_TSO6)) { card->dev->needed_headroom = sizeof(struct qeth_hdr_tso); netif_keep_dst(card->dev); netif_set_tso_max_size(card->dev, PAGE_SIZE * (QDIO_MAX_ELEMENTS_PER_BUFFER - 1)); } netif_napi_add(card->dev, &card->napi, qeth_poll); return register_netdev(card->dev); } static void qeth_l2_trace_features(struct qeth_card *card) { /* Set BridgePort features */ QETH_CARD_TEXT(card, 2, "featuSBP"); QETH_CARD_HEX(card, 2, &card->options.sbp.supported_funcs, sizeof(card->options.sbp.supported_funcs)); /* VNIC Characteristics features */ QETH_CARD_TEXT(card, 2, "feaVNICC"); QETH_CARD_HEX(card, 2, &card->options.vnicc.sup_chars, sizeof(card->options.vnicc.sup_chars)); } static void qeth_l2_setup_bridgeport_attrs(struct qeth_card *card) { if (!card->options.sbp.reflect_promisc && card->options.sbp.role != QETH_SBP_ROLE_NONE) { /* Conditional to avoid spurious error messages */ qeth_bridgeport_setrole(card, card->options.sbp.role); /* Let the callback function refresh the stored role value. */ qeth_bridgeport_query_ports(card, &card->options.sbp.role, NULL); } if (card->options.sbp.hostnotification) { if (qeth_bridgeport_an_set(card, 1)) card->options.sbp.hostnotification = 0; } } /** * qeth_l2_detect_dev2br_support() - * Detect whether this card supports 'dev to bridge fdb network address * change notification' and thus can support the learning_sync bridgeport * attribute * @card: qeth_card structure pointer */ static void qeth_l2_detect_dev2br_support(struct qeth_card *card) { struct qeth_priv *priv = netdev_priv(card->dev); bool dev2br_supported; QETH_CARD_TEXT(card, 2, "d2brsup"); if (!IS_IQD(card)) return; /* dev2br requires valid cssid,iid,chid */ dev2br_supported = card->info.ids_valid && css_general_characteristics.enarf; QETH_CARD_TEXT_(card, 2, "D2Bsup%02x", dev2br_supported); if (dev2br_supported) priv->brport_hw_features |= BR_LEARNING_SYNC; else priv->brport_hw_features &= ~BR_LEARNING_SYNC; } static void qeth_l2_enable_brport_features(struct qeth_card *card) { struct qeth_priv *priv = netdev_priv(card->dev); int rc; if (priv->brport_features & BR_LEARNING_SYNC) { if (priv->brport_hw_features & BR_LEARNING_SYNC) { qeth_l2_set_pnso_mode(card, QETH_PNSO_ADDR_INFO); rc = qeth_l2_dev2br_an_set(card, true); if (rc == -EAGAIN) { /* Recoverable error, retry once */ qeth_l2_set_pnso_mode(card, QETH_PNSO_NONE); qeth_l2_dev2br_fdb_flush(card); qeth_l2_set_pnso_mode(card, QETH_PNSO_ADDR_INFO); rc = qeth_l2_dev2br_an_set(card, true); } if (rc) { netdev_err(card->dev, "failed to enable bridge learning_sync: %d\n", rc); qeth_l2_set_pnso_mode(card, QETH_PNSO_NONE); qeth_l2_dev2br_fdb_flush(card); priv->brport_features ^= BR_LEARNING_SYNC; } } else { dev_warn(&card->gdev->dev, "bridge learning_sync not supported\n"); priv->brport_features ^= BR_LEARNING_SYNC; } } } /* SETBRIDGEPORT support, async notifications */ enum qeth_an_event_type {anev_reg_unreg, anev_abort, anev_reset}; /** * qeth_bridge_emit_host_event() - bridgeport address change notification * @card: qeth_card structure pointer, for udev events. * @evtype: "normal" register/unregister, or abort, or reset. For abort * and reset token and addr_lnid are unused and may be NULL. * @code: event bitmask: high order bit 0x80 value 1 means removal of an * object, 0 - addition of an object. * 0x01 - VLAN, 0x02 - MAC, 0x03 - VLAN and MAC. * @token: "network token" structure identifying physical address of the port. * @addr_lnid: pointer to structure with MAC address and VLAN ID. * * This function is called when registrations and deregistrations are * reported by the hardware, and also when notifications are enabled - * for all currently registered addresses. */ static void qeth_bridge_emit_host_event(struct qeth_card *card, enum qeth_an_event_type evtype, u8 code, struct net_if_token *token, struct mac_addr_lnid *addr_lnid) { char str[7][32]; char *env[8]; int i = 0; switch (evtype) { case anev_reg_unreg: scnprintf(str[i], sizeof(str[i]), "BRIDGEDHOST=%s", (code & IPA_ADDR_CHANGE_CODE_REMOVAL) ? "deregister" : "register"); env[i] = str[i]; i++; if (code & IPA_ADDR_CHANGE_CODE_VLANID) { scnprintf(str[i], sizeof(str[i]), "VLAN=%d", addr_lnid->lnid); env[i] = str[i]; i++; } if (code & IPA_ADDR_CHANGE_CODE_MACADDR) { scnprintf(str[i], sizeof(str[i]), "MAC=%pM", addr_lnid->mac); env[i] = str[i]; i++; } scnprintf(str[i], sizeof(str[i]), "NTOK_BUSID=%x.%x.%04x", token->cssid, token->ssid, token->devnum); env[i] = str[i]; i++; scnprintf(str[i], sizeof(str[i]), "NTOK_IID=%02x", token->iid); env[i] = str[i]; i++; scnprintf(str[i], sizeof(str[i]), "NTOK_CHPID=%02x", token->chpid); env[i] = str[i]; i++; scnprintf(str[i], sizeof(str[i]), "NTOK_CHID=%04x", token->chid); env[i] = str[i]; i++; break; case anev_abort: scnprintf(str[i], sizeof(str[i]), "BRIDGEDHOST=abort"); env[i] = str[i]; i++; break; case anev_reset: scnprintf(str[i], sizeof(str[i]), "BRIDGEDHOST=reset"); env[i] = str[i]; i++; break; } env[i] = NULL; kobject_uevent_env(&card->gdev->dev.kobj, KOBJ_CHANGE, env); } struct qeth_bridge_state_data { struct work_struct worker; struct qeth_card *card; u8 role; u8 state; }; static void qeth_bridge_state_change_worker(struct work_struct *work) { struct qeth_bridge_state_data *data = container_of(work, struct qeth_bridge_state_data, worker); char env_locrem[32]; char env_role[32]; char env_state[32]; char *env[] = { env_locrem, env_role, env_state, NULL }; scnprintf(env_locrem, sizeof(env_locrem), "BRIDGEPORT=statechange"); scnprintf(env_role, sizeof(env_role), "ROLE=%s", (data->role == QETH_SBP_ROLE_NONE) ? "none" : (data->role == QETH_SBP_ROLE_PRIMARY) ? "primary" : (data->role == QETH_SBP_ROLE_SECONDARY) ? "secondary" : "<INVALID>"); scnprintf(env_state, sizeof(env_state), "STATE=%s", (data->state == QETH_SBP_STATE_INACTIVE) ? "inactive" : (data->state == QETH_SBP_STATE_STANDBY) ? "standby" : (data->state == QETH_SBP_STATE_ACTIVE) ? "active" : "<INVALID>"); kobject_uevent_env(&data->card->gdev->dev.kobj, KOBJ_CHANGE, env); kfree(data); } static void qeth_bridge_state_change(struct qeth_card *card, struct qeth_ipa_cmd *cmd) { struct qeth_sbp_port_data *qports = &cmd->data.sbp.data.port_data; struct qeth_bridge_state_data *data; QETH_CARD_TEXT(card, 2, "brstchng"); if (qports->num_entries == 0) { QETH_CARD_TEXT(card, 2, "BPempty"); return; } if (qports->entry_length != sizeof(struct qeth_sbp_port_entry)) { QETH_CARD_TEXT_(card, 2, "BPsz%04x", qports->entry_length); return; } data = kzalloc(sizeof(*data), GFP_ATOMIC); if (!data) { QETH_CARD_TEXT(card, 2, "BPSalloc"); return; } INIT_WORK(&data->worker, qeth_bridge_state_change_worker); data->card = card; /* Information for the local port: */ data->role = qports->entry[0].role; data->state = qports->entry[0].state; queue_work(card->event_wq, &data->worker); } struct qeth_addr_change_data { struct delayed_work dwork; struct qeth_card *card; struct qeth_ipacmd_addr_change ac_event; }; static void qeth_l2_dev2br_worker(struct work_struct *work) { struct delayed_work *dwork = to_delayed_work(work); struct qeth_addr_change_data *data; struct qeth_card *card; struct qeth_priv *priv; unsigned int i; int rc; data = container_of(dwork, struct qeth_addr_change_data, dwork); card = data->card; priv = netdev_priv(card->dev); QETH_CARD_TEXT(card, 4, "dev2brew"); if (READ_ONCE(card->info.pnso_mode) == QETH_PNSO_NONE) goto free; if (data->ac_event.lost_event_mask) { /* Potential re-config in progress, try again later: */ if (!rtnl_trylock()) { queue_delayed_work(card->event_wq, dwork, msecs_to_jiffies(100)); return; } if (!netif_device_present(card->dev)) { rtnl_unlock(); goto free; } QETH_DBF_MESSAGE(3, "Address change notification overflow on device %x\n", CARD_DEVID(card)); /* Card fdb and bridge fdb are out of sync, card has stopped * notifications (no need to drain_workqueue). Purge all * 'extern_learn' entries from the parent bridge and restart * the notifications. */ qeth_l2_dev2br_fdb_flush(card); rc = qeth_l2_dev2br_an_set(card, true); if (rc) { /* TODO: if we want to retry after -EAGAIN, be * aware there could be stale entries in the * workqueue now, that need to be drained. * For now we give up: */ netdev_err(card->dev, "bridge learning_sync failed to recover: %d\n", rc); WRITE_ONCE(card->info.pnso_mode, QETH_PNSO_NONE); /* To remove fdb entries reported by an_set: */ qeth_l2_dev2br_fdb_flush(card); priv->brport_features ^= BR_LEARNING_SYNC; } else { QETH_DBF_MESSAGE(3, "Address Notification resynced on device %x\n", CARD_DEVID(card)); } rtnl_unlock(); } else { for (i = 0; i < data->ac_event.num_entries; i++) { struct qeth_ipacmd_addr_change_entry *entry = &data->ac_event.entry[i]; qeth_l2_dev2br_fdb_notify(card, entry->change_code, &entry->token, &entry->addr_lnid); } } free: kfree(data); } static void qeth_addr_change_event_worker(struct work_struct *work) { struct delayed_work *dwork = to_delayed_work(work); struct qeth_addr_change_data *data; struct qeth_card *card; int i; data = container_of(dwork, struct qeth_addr_change_data, dwork); card = data->card; QETH_CARD_TEXT(data->card, 4, "adrchgew"); if (READ_ONCE(card->info.pnso_mode) == QETH_PNSO_NONE) goto free; if (data->ac_event.lost_event_mask) { /* Potential re-config in progress, try again later: */ if (!mutex_trylock(&card->sbp_lock)) { queue_delayed_work(card->event_wq, dwork, msecs_to_jiffies(100)); return; } dev_info(&data->card->gdev->dev, "Address change notification stopped on %s (%s)\n", netdev_name(card->dev), (data->ac_event.lost_event_mask == 0x01) ? "Overflow" : (data->ac_event.lost_event_mask == 0x02) ? "Bridge port state change" : "Unknown reason"); data->card->options.sbp.hostnotification = 0; card->info.pnso_mode = QETH_PNSO_NONE; mutex_unlock(&data->card->sbp_lock); qeth_bridge_emit_host_event(data->card, anev_abort, 0, NULL, NULL); } else for (i = 0; i < data->ac_event.num_entries; i++) { struct qeth_ipacmd_addr_change_entry *entry = &data->ac_event.entry[i]; qeth_bridge_emit_host_event(data->card, anev_reg_unreg, entry->change_code, &entry->token, &entry->addr_lnid); } free: kfree(data); } static void qeth_addr_change_event(struct qeth_card *card, struct qeth_ipa_cmd *cmd) { struct qeth_ipacmd_addr_change *hostevs = &cmd->data.addrchange; struct qeth_addr_change_data *data; int extrasize; if (card->info.pnso_mode == QETH_PNSO_NONE) return; QETH_CARD_TEXT(card, 4, "adrchgev"); if (cmd->hdr.return_code != 0x0000) { if (cmd->hdr.return_code == 0x0010) { if (hostevs->lost_event_mask == 0x00) hostevs->lost_event_mask = 0xff; } else { QETH_CARD_TEXT_(card, 2, "ACHN%04x", cmd->hdr.return_code); return; } } extrasize = sizeof(struct qeth_ipacmd_addr_change_entry) * hostevs->num_entries; data = kzalloc(sizeof(struct qeth_addr_change_data) + extrasize, GFP_ATOMIC); if (!data) { QETH_CARD_TEXT(card, 2, "ACNalloc"); return; } if (card->info.pnso_mode == QETH_PNSO_BRIDGEPORT) INIT_DELAYED_WORK(&data->dwork, qeth_addr_change_event_worker); else INIT_DELAYED_WORK(&data->dwork, qeth_l2_dev2br_worker); data->card = card; data->ac_event = *hostevs; memcpy(data->ac_event.entry, hostevs->entry, extrasize); queue_delayed_work(card->event_wq, &data->dwork, 0); } /* SETBRIDGEPORT support; sending commands */ struct _qeth_sbp_cbctl { union { u32 supported; struct { enum qeth_sbp_roles *role; enum qeth_sbp_states *state; } qports; } data; }; static int qeth_bridgeport_makerc(struct qeth_card *card, struct qeth_ipa_cmd *cmd) { struct qeth_ipacmd_setbridgeport *sbp = &cmd->data.sbp; enum qeth_ipa_sbp_cmd setcmd = sbp->hdr.command_code; u16 ipa_rc = cmd->hdr.return_code; u16 sbp_rc = sbp->hdr.return_code; int rc; if (ipa_rc == IPA_RC_SUCCESS && sbp_rc == IPA_RC_SUCCESS) return 0; if ((IS_IQD(card) && ipa_rc == IPA_RC_SUCCESS) || (!IS_IQD(card) && ipa_rc == sbp_rc)) { switch (sbp_rc) { case IPA_RC_SUCCESS: rc = 0; break; case IPA_RC_L2_UNSUPPORTED_CMD: case IPA_RC_UNSUPPORTED_COMMAND: rc = -EOPNOTSUPP; break; case IPA_RC_SBP_OSA_NOT_CONFIGURED: case IPA_RC_SBP_IQD_NOT_CONFIGURED: rc = -ENODEV; /* maybe not the best code here? */ dev_err(&card->gdev->dev, "The device is not configured as a Bridge Port\n"); break; case IPA_RC_SBP_OSA_OS_MISMATCH: case IPA_RC_SBP_IQD_OS_MISMATCH: rc = -EPERM; dev_err(&card->gdev->dev, "A Bridge Port is already configured by a different operating system\n"); break; case IPA_RC_SBP_OSA_ANO_DEV_PRIMARY: case IPA_RC_SBP_IQD_ANO_DEV_PRIMARY: switch (setcmd) { case IPA_SBP_SET_PRIMARY_BRIDGE_PORT: rc = -EEXIST; dev_err(&card->gdev->dev, "The LAN already has a primary Bridge Port\n"); break; case IPA_SBP_SET_SECONDARY_BRIDGE_PORT: rc = -EBUSY; dev_err(&card->gdev->dev, "The device is already a primary Bridge Port\n"); break; default: rc = -EIO; } break; case IPA_RC_SBP_OSA_CURRENT_SECOND: case IPA_RC_SBP_IQD_CURRENT_SECOND: rc = -EBUSY; dev_err(&card->gdev->dev, "The device is already a secondary Bridge Port\n"); break; case IPA_RC_SBP_OSA_LIMIT_SECOND: case IPA_RC_SBP_IQD_LIMIT_SECOND: rc = -EEXIST; dev_err(&card->gdev->dev, "The LAN cannot have more secondary Bridge Ports\n"); break; case IPA_RC_SBP_OSA_CURRENT_PRIMARY: case IPA_RC_SBP_IQD_CURRENT_PRIMARY: rc = -EBUSY; dev_err(&card->gdev->dev, "The device is already a primary Bridge Port\n"); break; case IPA_RC_SBP_OSA_NOT_AUTHD_BY_ZMAN: case IPA_RC_SBP_IQD_NOT_AUTHD_BY_ZMAN: rc = -EACCES; dev_err(&card->gdev->dev, "The device is not authorized to be a Bridge Port\n"); break; default: rc = -EIO; } } else { switch (ipa_rc) { case IPA_RC_NOTSUPP: rc = -EOPNOTSUPP; break; case IPA_RC_UNSUPPORTED_COMMAND: rc = -EOPNOTSUPP; break; default: rc = -EIO; } } if (rc) { QETH_CARD_TEXT_(card, 2, "SBPi%04x", ipa_rc); QETH_CARD_TEXT_(card, 2, "SBPc%04x", sbp_rc); } return rc; } static struct qeth_cmd_buffer *qeth_sbp_build_cmd(struct qeth_card *card, enum qeth_ipa_sbp_cmd sbp_cmd, unsigned int data_length) { enum qeth_ipa_cmds ipa_cmd = IS_IQD(card) ? IPA_CMD_SETBRIDGEPORT_IQD : IPA_CMD_SETBRIDGEPORT_OSA; struct qeth_ipacmd_sbp_hdr *hdr; struct qeth_cmd_buffer *iob; iob = qeth_ipa_alloc_cmd(card, ipa_cmd, QETH_PROT_NONE, data_length + offsetof(struct qeth_ipacmd_setbridgeport, data)); if (!iob) return iob; hdr = &__ipa_cmd(iob)->data.sbp.hdr; hdr->cmdlength = sizeof(*hdr) + data_length; hdr->command_code = sbp_cmd; hdr->used_total = 1; hdr->seq_no = 1; return iob; } static int qeth_bridgeport_query_support_cb(struct qeth_card *card, struct qeth_reply *reply, unsigned long data) { struct qeth_ipa_cmd *cmd = (struct qeth_ipa_cmd *) data; struct _qeth_sbp_cbctl *cbctl = (struct _qeth_sbp_cbctl *)reply->param; int rc; QETH_CARD_TEXT(card, 2, "brqsupcb"); rc = qeth_bridgeport_makerc(card, cmd); if (rc) return rc; cbctl->data.supported = cmd->data.sbp.data.query_cmds_supp.supported_cmds; return 0; } /** * qeth_bridgeport_query_support() - store bitmask of supported subfunctions. * @card: qeth_card structure pointer. * * Sets bitmask of supported setbridgeport subfunctions in the qeth_card * strucutre: card->options.sbp.supported_funcs. */ static void qeth_bridgeport_query_support(struct qeth_card *card) { struct qeth_cmd_buffer *iob; struct _qeth_sbp_cbctl cbctl; QETH_CARD_TEXT(card, 2, "brqsuppo"); iob = qeth_sbp_build_cmd(card, IPA_SBP_QUERY_COMMANDS_SUPPORTED, SBP_DATA_SIZEOF(query_cmds_supp)); if (!iob) return; if (qeth_send_ipa_cmd(card, iob, qeth_bridgeport_query_support_cb, &cbctl)) { card->options.sbp.role = QETH_SBP_ROLE_NONE; card->options.sbp.supported_funcs = 0; return; } card->options.sbp.supported_funcs = cbctl.data.supported; } static int qeth_bridgeport_query_ports_cb(struct qeth_card *card, struct qeth_reply *reply, unsigned long data) { struct qeth_ipa_cmd *cmd = (struct qeth_ipa_cmd *) data; struct _qeth_sbp_cbctl *cbctl = (struct _qeth_sbp_cbctl *)reply->param; struct qeth_sbp_port_data *qports; int rc; QETH_CARD_TEXT(card, 2, "brqprtcb"); rc = qeth_bridgeport_makerc(card, cmd); if (rc) return rc; qports = &cmd->data.sbp.data.port_data; if (qports->entry_length != sizeof(struct qeth_sbp_port_entry)) { QETH_CARD_TEXT_(card, 2, "SBPs%04x", qports->entry_length); return -EINVAL; } /* first entry contains the state of the local port */ if (qports->num_entries > 0) { if (cbctl->data.qports.role) *cbctl->data.qports.role = qports->entry[0].role; if (cbctl->data.qports.state) *cbctl->data.qports.state = qports->entry[0].state; } return 0; } /** * qeth_bridgeport_query_ports() - query local bridgeport status. * @card: qeth_card structure pointer. * @role: Role of the port: 0-none, 1-primary, 2-secondary. * @state: State of the port: 0-inactive, 1-standby, 2-active. * * Returns negative errno-compatible error indication or 0 on success. * * 'role' and 'state' are not updated in case of hardware operation failure. */ int qeth_bridgeport_query_ports(struct qeth_card *card, enum qeth_sbp_roles *role, enum qeth_sbp_states *state) { struct qeth_cmd_buffer *iob; struct _qeth_sbp_cbctl cbctl = { .data = { .qports = { .role = role, .state = state, }, }, }; QETH_CARD_TEXT(card, 2, "brqports"); if (!(card->options.sbp.supported_funcs & IPA_SBP_QUERY_BRIDGE_PORTS)) return -EOPNOTSUPP; iob = qeth_sbp_build_cmd(card, IPA_SBP_QUERY_BRIDGE_PORTS, 0); if (!iob) return -ENOMEM; return qeth_send_ipa_cmd(card, iob, qeth_bridgeport_query_ports_cb, &cbctl); } static int qeth_bridgeport_set_cb(struct qeth_card *card, struct qeth_reply *reply, unsigned long data) { struct qeth_ipa_cmd *cmd = (struct qeth_ipa_cmd *)data; QETH_CARD_TEXT(card, 2, "brsetrcb"); return qeth_bridgeport_makerc(card, cmd); } /** * qeth_bridgeport_setrole() - Assign primary role to the port. * @card: qeth_card structure pointer. * @role: Role to assign. * * Returns negative errno-compatible error indication or 0 on success. */ int qeth_bridgeport_setrole(struct qeth_card *card, enum qeth_sbp_roles role) { struct qeth_cmd_buffer *iob; enum qeth_ipa_sbp_cmd setcmd; unsigned int cmdlength = 0; QETH_CARD_TEXT(card, 2, "brsetrol"); switch (role) { case QETH_SBP_ROLE_NONE: setcmd = IPA_SBP_RESET_BRIDGE_PORT_ROLE; break; case QETH_SBP_ROLE_PRIMARY: setcmd = IPA_SBP_SET_PRIMARY_BRIDGE_PORT; cmdlength = SBP_DATA_SIZEOF(set_primary); break; case QETH_SBP_ROLE_SECONDARY: setcmd = IPA_SBP_SET_SECONDARY_BRIDGE_PORT; break; default: return -EINVAL; } if (!(card->options.sbp.supported_funcs & setcmd)) return -EOPNOTSUPP; iob = qeth_sbp_build_cmd(card, setcmd, cmdlength); if (!iob) return -ENOMEM; return qeth_send_ipa_cmd(card, iob, qeth_bridgeport_set_cb, NULL); } static void qeth_bridgeport_an_set_cb(void *priv, struct chsc_pnso_naid_l2 *entry) { struct qeth_card *card = (struct qeth_card *)priv; u8 code; code = IPA_ADDR_CHANGE_CODE_MACADDR; if (entry->addr_lnid.lnid < VLAN_N_VID) code |= IPA_ADDR_CHANGE_CODE_VLANID; qeth_bridge_emit_host_event(card, anev_reg_unreg, code, (struct net_if_token *)&entry->nit, (struct mac_addr_lnid *)&entry->addr_lnid); } /** * qeth_bridgeport_an_set() - Enable or disable bridgeport address notification * @card: qeth_card structure pointer. * @enable: 0 - disable, non-zero - enable notifications * * Returns negative errno-compatible error indication or 0 on success. * * On enable, emits a series of address notifications udev events for all * currently registered hosts. */ int qeth_bridgeport_an_set(struct qeth_card *card, int enable) { int rc; if (!card->options.sbp.supported_funcs) return -EOPNOTSUPP; if (enable) { qeth_bridge_emit_host_event(card, anev_reset, 0, NULL, NULL); qeth_l2_set_pnso_mode(card, QETH_PNSO_BRIDGEPORT); rc = qeth_l2_pnso(card, PNSO_OC_NET_BRIDGE_INFO, 1, qeth_bridgeport_an_set_cb, card); if (rc) qeth_l2_set_pnso_mode(card, QETH_PNSO_NONE); } else { rc = qeth_l2_pnso(card, PNSO_OC_NET_BRIDGE_INFO, 0, NULL, NULL); qeth_l2_set_pnso_mode(card, QETH_PNSO_NONE); } return rc; } /* VNIC Characteristics support */ /* handle VNICC IPA command return codes; convert to error codes */ static int qeth_l2_vnicc_makerc(struct qeth_card *card, u16 ipa_rc) { int rc; switch (ipa_rc) { case IPA_RC_SUCCESS: return ipa_rc; case IPA_RC_L2_UNSUPPORTED_CMD: case IPA_RC_NOTSUPP: rc = -EOPNOTSUPP; break; case IPA_RC_VNICC_OOSEQ: rc = -EALREADY; break; case IPA_RC_VNICC_VNICBP: rc = -EBUSY; break; case IPA_RC_L2_ADDR_TABLE_FULL: rc = -ENOSPC; break; case IPA_RC_L2_MAC_NOT_AUTH_BY_ADP: rc = -EACCES; break; default: rc = -EIO; } QETH_CARD_TEXT_(card, 2, "err%04x", ipa_rc); return rc; } /* generic VNICC request call back */ static int qeth_l2_vnicc_request_cb(struct qeth_card *card, struct qeth_reply *reply, unsigned long data) { struct qeth_ipa_cmd *cmd = (struct qeth_ipa_cmd *) data; struct qeth_ipacmd_vnicc *rep = &cmd->data.vnicc; u32 sub_cmd = cmd->data.vnicc.hdr.sub_command; QETH_CARD_TEXT(card, 2, "vniccrcb"); if (cmd->hdr.return_code) return qeth_l2_vnicc_makerc(card, cmd->hdr.return_code); /* return results to caller */ card->options.vnicc.sup_chars = rep->vnicc_cmds.supported; card->options.vnicc.cur_chars = rep->vnicc_cmds.enabled; if (sub_cmd == IPA_VNICC_QUERY_CMDS) *(u32 *)reply->param = rep->data.query_cmds.sup_cmds; else if (sub_cmd == IPA_VNICC_GET_TIMEOUT) *(u32 *)reply->param = rep->data.getset_timeout.timeout; return 0; } static struct qeth_cmd_buffer *qeth_l2_vnicc_build_cmd(struct qeth_card *card, u32 vnicc_cmd, unsigned int data_length) { struct qeth_ipacmd_vnicc_hdr *hdr; struct qeth_cmd_buffer *iob; iob = qeth_ipa_alloc_cmd(card, IPA_CMD_VNICC, QETH_PROT_NONE, data_length + offsetof(struct qeth_ipacmd_vnicc, data)); if (!iob) return NULL; hdr = &__ipa_cmd(iob)->data.vnicc.hdr; hdr->data_length = sizeof(*hdr) + data_length; hdr->sub_command = vnicc_cmd; return iob; } /* VNICC query VNIC characteristics request */ static int qeth_l2_vnicc_query_chars(struct qeth_card *card) { struct qeth_cmd_buffer *iob; QETH_CARD_TEXT(card, 2, "vniccqch"); iob = qeth_l2_vnicc_build_cmd(card, IPA_VNICC_QUERY_CHARS, 0); if (!iob) return -ENOMEM; return qeth_send_ipa_cmd(card, iob, qeth_l2_vnicc_request_cb, NULL); } /* VNICC query sub commands request */ static int qeth_l2_vnicc_query_cmds(struct qeth_card *card, u32 vnic_char, u32 *sup_cmds) { struct qeth_cmd_buffer *iob; QETH_CARD_TEXT(card, 2, "vniccqcm"); iob = qeth_l2_vnicc_build_cmd(card, IPA_VNICC_QUERY_CMDS, VNICC_DATA_SIZEOF(query_cmds)); if (!iob) return -ENOMEM; __ipa_cmd(iob)->data.vnicc.data.query_cmds.vnic_char = vnic_char; return qeth_send_ipa_cmd(card, iob, qeth_l2_vnicc_request_cb, sup_cmds); } /* VNICC enable/disable characteristic request */ static int qeth_l2_vnicc_set_char(struct qeth_card *card, u32 vnic_char, u32 cmd) { struct qeth_cmd_buffer *iob; QETH_CARD_TEXT(card, 2, "vniccedc"); iob = qeth_l2_vnicc_build_cmd(card, cmd, VNICC_DATA_SIZEOF(set_char)); if (!iob) return -ENOMEM; __ipa_cmd(iob)->data.vnicc.data.set_char.vnic_char = vnic_char; return qeth_send_ipa_cmd(card, iob, qeth_l2_vnicc_request_cb, NULL); } /* VNICC get/set timeout for characteristic request */ static int qeth_l2_vnicc_getset_timeout(struct qeth_card *card, u32 vnicc, u32 cmd, u32 *timeout) { struct qeth_vnicc_getset_timeout *getset_timeout; struct qeth_cmd_buffer *iob; QETH_CARD_TEXT(card, 2, "vniccgst"); iob = qeth_l2_vnicc_build_cmd(card, cmd, VNICC_DATA_SIZEOF(getset_timeout)); if (!iob) return -ENOMEM; getset_timeout = &__ipa_cmd(iob)->data.vnicc.data.getset_timeout; getset_timeout->vnic_char = vnicc; if (cmd == IPA_VNICC_SET_TIMEOUT) getset_timeout->timeout = *timeout; return qeth_send_ipa_cmd(card, iob, qeth_l2_vnicc_request_cb, timeout); } /* recover user timeout setting */ static bool qeth_l2_vnicc_recover_timeout(struct qeth_card *card, u32 vnicc, u32 *timeout) { if (card->options.vnicc.sup_chars & vnicc && card->options.vnicc.getset_timeout_sup & vnicc && !qeth_l2_vnicc_getset_timeout(card, vnicc, IPA_VNICC_SET_TIMEOUT, timeout)) return false; *timeout = QETH_VNICC_DEFAULT_TIMEOUT; return true; } /* set current VNICC flag state; called from sysfs store function */ int qeth_l2_vnicc_set_state(struct qeth_card *card, u32 vnicc, bool state) { int rc = 0; u32 cmd; QETH_CARD_TEXT(card, 2, "vniccsch"); /* check if characteristic and enable/disable are supported */ if (!(card->options.vnicc.sup_chars & vnicc) || !(card->options.vnicc.set_char_sup & vnicc)) return -EOPNOTSUPP; if (qeth_bridgeport_is_in_use(card)) return -EBUSY; /* set enable/disable command and store wanted characteristic */ if (state) { cmd = IPA_VNICC_ENABLE; card->options.vnicc.wanted_chars |= vnicc; } else { cmd = IPA_VNICC_DISABLE; card->options.vnicc.wanted_chars &= ~vnicc; } /* do we need to do anything? */ if (card->options.vnicc.cur_chars == card->options.vnicc.wanted_chars) return rc; /* if card is not ready, simply stop here */ if (!qeth_card_hw_is_reachable(card)) { if (state) card->options.vnicc.cur_chars |= vnicc; else card->options.vnicc.cur_chars &= ~vnicc; return rc; } rc = qeth_l2_vnicc_set_char(card, vnicc, cmd); if (rc) card->options.vnicc.wanted_chars = card->options.vnicc.cur_chars; else { /* successful online VNICC change; handle special cases */ if (state && vnicc == QETH_VNICC_RX_BCAST) card->options.vnicc.rx_bcast_enabled = true; if (!state && vnicc == QETH_VNICC_LEARNING) qeth_l2_vnicc_recover_timeout(card, vnicc, &card->options.vnicc.learning_timeout); } return rc; } /* get current VNICC flag state; called from sysfs show function */ int qeth_l2_vnicc_get_state(struct qeth_card *card, u32 vnicc, bool *state) { int rc = 0; QETH_CARD_TEXT(card, 2, "vniccgch"); /* check if characteristic is supported */ if (!(card->options.vnicc.sup_chars & vnicc)) return -EOPNOTSUPP; if (qeth_bridgeport_is_in_use(card)) return -EBUSY; /* if card is ready, query current VNICC state */ if (qeth_card_hw_is_reachable(card)) rc = qeth_l2_vnicc_query_chars(card); *state = (card->options.vnicc.cur_chars & vnicc) ? true : false; return rc; } /* set VNICC timeout; called from sysfs store function. Currently, only learning * supports timeout */ int qeth_l2_vnicc_set_timeout(struct qeth_card *card, u32 timeout) { int rc = 0; QETH_CARD_TEXT(card, 2, "vniccsto"); /* check if characteristic and set_timeout are supported */ if (!(card->options.vnicc.sup_chars & QETH_VNICC_LEARNING) || !(card->options.vnicc.getset_timeout_sup & QETH_VNICC_LEARNING)) return -EOPNOTSUPP; if (qeth_bridgeport_is_in_use(card)) return -EBUSY; /* do we need to do anything? */ if (card->options.vnicc.learning_timeout == timeout) return rc; /* if card is not ready, simply store the value internally and return */ if (!qeth_card_hw_is_reachable(card)) { card->options.vnicc.learning_timeout = timeout; return rc; } /* send timeout value to card; if successful, store value internally */ rc = qeth_l2_vnicc_getset_timeout(card, QETH_VNICC_LEARNING, IPA_VNICC_SET_TIMEOUT, &timeout); if (!rc) card->options.vnicc.learning_timeout = timeout; return rc; } /* get current VNICC timeout; called from sysfs show function. Currently, only * learning supports timeout */ int qeth_l2_vnicc_get_timeout(struct qeth_card *card, u32 *timeout) { int rc = 0; QETH_CARD_TEXT(card, 2, "vniccgto"); /* check if characteristic and get_timeout are supported */ if (!(card->options.vnicc.sup_chars & QETH_VNICC_LEARNING) || !(card->options.vnicc.getset_timeout_sup & QETH_VNICC_LEARNING)) return -EOPNOTSUPP; if (qeth_bridgeport_is_in_use(card)) return -EBUSY; /* if card is ready, get timeout. Otherwise, just return stored value */ *timeout = card->options.vnicc.learning_timeout; if (qeth_card_hw_is_reachable(card)) rc = qeth_l2_vnicc_getset_timeout(card, QETH_VNICC_LEARNING, IPA_VNICC_GET_TIMEOUT, timeout); return rc; } /* check if VNICC is currently enabled */ static bool _qeth_l2_vnicc_is_in_use(struct qeth_card *card) { if (!card->options.vnicc.sup_chars) return false; /* default values are only OK if rx_bcast was not enabled by user * or the card is offline. */ if (card->options.vnicc.cur_chars == QETH_VNICC_DEFAULT) { if (!card->options.vnicc.rx_bcast_enabled || !qeth_card_hw_is_reachable(card)) return false; } return true; } /** * qeth_bridgeport_allowed - are any qeth_bridgeport functions allowed? * @card: qeth_card structure pointer * * qeth_bridgeport functionality is mutually exclusive with usage of the * VNIC Characteristics and dev2br address notifications */ bool qeth_bridgeport_allowed(struct qeth_card *card) { struct qeth_priv *priv = netdev_priv(card->dev); return (!_qeth_l2_vnicc_is_in_use(card) && !(priv->brport_features & BR_LEARNING_SYNC)); } /* recover user characteristic setting */ static bool qeth_l2_vnicc_recover_char(struct qeth_card *card, u32 vnicc, bool enable) { u32 cmd = enable ? IPA_VNICC_ENABLE : IPA_VNICC_DISABLE; if (card->options.vnicc.sup_chars & vnicc && card->options.vnicc.set_char_sup & vnicc && !qeth_l2_vnicc_set_char(card, vnicc, cmd)) return false; card->options.vnicc.wanted_chars &= ~vnicc; card->options.vnicc.wanted_chars |= QETH_VNICC_DEFAULT & vnicc; return true; } /* (re-)initialize VNICC */ static void qeth_l2_vnicc_init(struct qeth_card *card) { u32 *timeout = &card->options.vnicc.learning_timeout; bool enable, error = false; unsigned int chars_len, i; unsigned long chars_tmp; u32 sup_cmds, vnicc; QETH_CARD_TEXT(card, 2, "vniccini"); /* reset rx_bcast */ card->options.vnicc.rx_bcast_enabled = 0; /* initial query and storage of VNIC characteristics */ if (qeth_l2_vnicc_query_chars(card)) { if (card->options.vnicc.wanted_chars != QETH_VNICC_DEFAULT || *timeout != QETH_VNICC_DEFAULT_TIMEOUT) dev_err(&card->gdev->dev, "Configuring the VNIC characteristics failed\n"); /* fail quietly if user didn't change the default config */ card->options.vnicc.sup_chars = 0; card->options.vnicc.cur_chars = 0; card->options.vnicc.wanted_chars = QETH_VNICC_DEFAULT; return; } /* get supported commands for each supported characteristic */ chars_tmp = card->options.vnicc.sup_chars; chars_len = sizeof(card->options.vnicc.sup_chars) * BITS_PER_BYTE; for_each_set_bit(i, &chars_tmp, chars_len) { vnicc = BIT(i); if (qeth_l2_vnicc_query_cmds(card, vnicc, &sup_cmds)) { sup_cmds = 0; error = true; } if ((sup_cmds & IPA_VNICC_SET_TIMEOUT) && (sup_cmds & IPA_VNICC_GET_TIMEOUT)) card->options.vnicc.getset_timeout_sup |= vnicc; else card->options.vnicc.getset_timeout_sup &= ~vnicc; if ((sup_cmds & IPA_VNICC_ENABLE) && (sup_cmds & IPA_VNICC_DISABLE)) card->options.vnicc.set_char_sup |= vnicc; else card->options.vnicc.set_char_sup &= ~vnicc; } /* enforce assumed default values and recover settings, if changed */ error |= qeth_l2_vnicc_recover_timeout(card, QETH_VNICC_LEARNING, timeout); /* Change chars, if necessary */ chars_tmp = card->options.vnicc.wanted_chars ^ card->options.vnicc.cur_chars; chars_len = sizeof(card->options.vnicc.wanted_chars) * BITS_PER_BYTE; for_each_set_bit(i, &chars_tmp, chars_len) { vnicc = BIT(i); enable = card->options.vnicc.wanted_chars & vnicc; error |= qeth_l2_vnicc_recover_char(card, vnicc, enable); } if (error) dev_err(&card->gdev->dev, "Configuring the VNIC characteristics failed\n"); } /* configure default values of VNIC characteristics */ static void qeth_l2_vnicc_set_defaults(struct qeth_card *card) { /* characteristics values */ card->options.vnicc.sup_chars = QETH_VNICC_ALL; card->options.vnicc.cur_chars = QETH_VNICC_DEFAULT; card->options.vnicc.learning_timeout = QETH_VNICC_DEFAULT_TIMEOUT; /* supported commands */ card->options.vnicc.set_char_sup = QETH_VNICC_ALL; card->options.vnicc.getset_timeout_sup = QETH_VNICC_LEARNING; /* settings wanted by users */ card->options.vnicc.wanted_chars = QETH_VNICC_DEFAULT; } static const struct device_type qeth_l2_devtype = { .name = "qeth_layer2", .groups = qeth_l2_attr_groups, }; static int qeth_l2_probe_device(struct ccwgroup_device *gdev) { struct qeth_card *card = dev_get_drvdata(&gdev->dev); int rc; qeth_l2_vnicc_set_defaults(card); mutex_init(&card->sbp_lock); if (gdev->dev.type) { rc = device_add_groups(&gdev->dev, qeth_l2_attr_groups); if (rc) return rc; } else { gdev->dev.type = &qeth_l2_devtype; } INIT_WORK(&card->rx_mode_work, qeth_l2_rx_mode_work); return 0; } static void qeth_l2_remove_device(struct ccwgroup_device *gdev) { struct qeth_card *card = dev_get_drvdata(&gdev->dev); struct qeth_priv *priv; if (gdev->dev.type != &qeth_l2_devtype) device_remove_groups(&gdev->dev, qeth_l2_attr_groups); qeth_set_allowed_threads(card, 0, 1); wait_event(card->wait_q, qeth_threads_running(card, 0xffffffff) == 0); if (gdev->state == CCWGROUP_ONLINE) qeth_set_offline(card, card->discipline, false); if (card->dev->reg_state == NETREG_REGISTERED) { priv = netdev_priv(card->dev); if (priv->brport_features & BR_LEARNING_SYNC) { rtnl_lock(); qeth_l2_br2dev_put(); rtnl_unlock(); } unregister_netdev(card->dev); } } static int qeth_l2_set_online(struct qeth_card *card, bool carrier_ok) { struct net_device *dev = card->dev; int rc = 0; qeth_l2_detect_dev2br_support(card); mutex_lock(&card->sbp_lock); qeth_bridgeport_query_support(card); if (card->options.sbp.supported_funcs) { qeth_l2_setup_bridgeport_attrs(card); dev_info(&card->gdev->dev, "The device represents a Bridge Capable Port\n"); } mutex_unlock(&card->sbp_lock); qeth_l2_register_dev_addr(card); /* for the rx_bcast characteristic, init VNICC after setmac */ qeth_l2_vnicc_init(card); qeth_l2_trace_features(card); /* softsetup */ QETH_CARD_TEXT(card, 2, "softsetp"); card->state = CARD_STATE_SOFTSETUP; qeth_set_allowed_threads(card, 0xffffffff, 0); if (dev->reg_state != NETREG_REGISTERED) { rc = qeth_l2_setup_netdev(card); if (rc) goto err_setup; if (carrier_ok) netif_carrier_on(dev); } else { rtnl_lock(); rc = qeth_set_real_num_tx_queues(card, qeth_tx_actual_queues(card)); if (rc) { rtnl_unlock(); goto err_set_queues; } if (carrier_ok) netif_carrier_on(dev); else netif_carrier_off(dev); netif_device_attach(dev); qeth_enable_hw_features(dev); qeth_l2_enable_brport_features(card); if (netif_running(dev)) { local_bh_disable(); napi_schedule(&card->napi); /* kick-start the NAPI softirq: */ local_bh_enable(); qeth_l2_set_rx_mode(dev); } rtnl_unlock(); } return 0; err_set_queues: err_setup: qeth_set_allowed_threads(card, 0, 1); card->state = CARD_STATE_DOWN; return rc; } static void qeth_l2_set_offline(struct qeth_card *card) { struct qeth_priv *priv = netdev_priv(card->dev); qeth_set_allowed_threads(card, 0, 1); qeth_l2_drain_rx_mode_cache(card); if (card->state == CARD_STATE_SOFTSETUP) card->state = CARD_STATE_DOWN; qeth_l2_set_pnso_mode(card, QETH_PNSO_NONE); if (priv->brport_features & BR_LEARNING_SYNC) qeth_l2_dev2br_fdb_flush(card); } /* Returns zero if the command is successfully "consumed" */ static int qeth_l2_control_event(struct qeth_card *card, struct qeth_ipa_cmd *cmd) { switch (cmd->hdr.command) { case IPA_CMD_SETBRIDGEPORT_OSA: case IPA_CMD_SETBRIDGEPORT_IQD: if (cmd->data.sbp.hdr.command_code == IPA_SBP_BRIDGE_PORT_STATE_CHANGE) { qeth_bridge_state_change(card, cmd); return 0; } return 1; case IPA_CMD_ADDRESS_CHANGE_NOTIF: qeth_addr_change_event(card, cmd); return 0; default: return 1; } } const struct qeth_discipline qeth_l2_discipline = { .setup = qeth_l2_probe_device, .remove = qeth_l2_remove_device, .set_online = qeth_l2_set_online, .set_offline = qeth_l2_set_offline, .control_event_handler = qeth_l2_control_event, }; EXPORT_SYMBOL_GPL(qeth_l2_discipline); static int __init qeth_l2_init(void) { pr_info("register layer 2 discipline\n"); refcount_set(&qeth_l2_switchdev_notify_refcnt, 0); return 0; } static void __exit qeth_l2_exit(void) { pr_info("unregister layer 2 discipline\n"); } module_init(qeth_l2_init); module_exit(qeth_l2_exit); MODULE_AUTHOR("Frank Blaschka <frank.blaschka@de.ibm.com>"); MODULE_DESCRIPTION("qeth layer 2 discipline"); MODULE_LICENSE("GPL");
Information contained on this website is for historical information purposes only and does not indicate or represent copyright ownership.
Created with Cregit http://github.com/cregit/cregit
Version 2.0-RC1