Author | Tokens | Token Proportion | Commits | Commit Proportion |
---|---|---|---|---|
Martin Schwidefsky | 1299 | 43.61% | 9 | 12.86% |
Peter Oberparleiter | 460 | 15.44% | 14 | 20.00% |
Alexandra Winter | 331 | 11.11% | 3 | 4.29% |
Sebastian Ott | 300 | 10.07% | 10 | 14.29% |
Cornelia Huck | 254 | 8.53% | 10 | 14.29% |
Halil Pasic | 114 | 3.83% | 3 | 4.29% |
Andrew Morton | 88 | 2.95% | 4 | 5.71% |
Thomas Spatzier | 34 | 1.14% | 1 | 1.43% |
Linus Torvalds | 33 | 1.11% | 2 | 2.86% |
Linus Torvalds (pre-git) | 24 | 0.81% | 7 | 10.00% |
Peter Tiedemann | 22 | 0.74% | 1 | 1.43% |
Frank Blaschka | 12 | 0.40% | 1 | 1.43% |
Jan Glauber | 4 | 0.13% | 1 | 1.43% |
Greg Kroah-Hartman | 2 | 0.07% | 2 | 2.86% |
Jan Höppner | 1 | 0.03% | 1 | 1.43% |
Paul Gortmaker | 1 | 0.03% | 1 | 1.43% |
Total | 2979 | 70 |
// SPDX-License-Identifier: GPL-1.0+ /* * Copyright IBM Corp. 2002, 2009 * * Author(s): Martin Schwidefsky (schwidefsky@de.ibm.com) * Cornelia Huck (cornelia.huck@de.ibm.com) */ #include <linux/export.h> #include <linux/init.h> #include <linux/errno.h> #include <linux/slab.h> #include <linux/list.h> #include <linux/device.h> #include <linux/delay.h> #include <linux/completion.h> #include <asm/ccwdev.h> #include <asm/idals.h> #include <asm/chpid.h> #include <asm/fcx.h> #include "cio.h" #include "cio_debug.h" #include "css.h" #include "chsc.h" #include "device.h" #include "chp.h" /** * ccw_device_set_options_mask() - set some options and unset the rest * @cdev: device for which the options are to be set * @flags: options to be set * * All flags specified in @flags are set, all flags not specified in @flags * are cleared. * Returns: * %0 on success, -%EINVAL on an invalid flag combination. */ int ccw_device_set_options_mask(struct ccw_device *cdev, unsigned long flags) { /* * The flag usage is mutal exclusive ... */ if ((flags & CCWDEV_EARLY_NOTIFICATION) && (flags & CCWDEV_REPORT_ALL)) return -EINVAL; cdev->private->options.fast = (flags & CCWDEV_EARLY_NOTIFICATION) != 0; cdev->private->options.repall = (flags & CCWDEV_REPORT_ALL) != 0; cdev->private->options.pgroup = (flags & CCWDEV_DO_PATHGROUP) != 0; cdev->private->options.force = (flags & CCWDEV_ALLOW_FORCE) != 0; cdev->private->options.mpath = (flags & CCWDEV_DO_MULTIPATH) != 0; return 0; } /** * ccw_device_set_options() - set some options * @cdev: device for which the options are to be set * @flags: options to be set * * All flags specified in @flags are set, the remainder is left untouched. * Returns: * %0 on success, -%EINVAL if an invalid flag combination would ensue. */ int ccw_device_set_options(struct ccw_device *cdev, unsigned long flags) { /* * The flag usage is mutal exclusive ... */ if (((flags & CCWDEV_EARLY_NOTIFICATION) && (flags & CCWDEV_REPORT_ALL)) || ((flags & CCWDEV_EARLY_NOTIFICATION) && cdev->private->options.repall) || ((flags & CCWDEV_REPORT_ALL) && cdev->private->options.fast)) return -EINVAL; cdev->private->options.fast |= (flags & CCWDEV_EARLY_NOTIFICATION) != 0; cdev->private->options.repall |= (flags & CCWDEV_REPORT_ALL) != 0; cdev->private->options.pgroup |= (flags & CCWDEV_DO_PATHGROUP) != 0; cdev->private->options.force |= (flags & CCWDEV_ALLOW_FORCE) != 0; cdev->private->options.mpath |= (flags & CCWDEV_DO_MULTIPATH) != 0; return 0; } /** * ccw_device_clear_options() - clear some options * @cdev: device for which the options are to be cleared * @flags: options to be cleared * * All flags specified in @flags are cleared, the remainder is left untouched. */ void ccw_device_clear_options(struct ccw_device *cdev, unsigned long flags) { cdev->private->options.fast &= (flags & CCWDEV_EARLY_NOTIFICATION) == 0; cdev->private->options.repall &= (flags & CCWDEV_REPORT_ALL) == 0; cdev->private->options.pgroup &= (flags & CCWDEV_DO_PATHGROUP) == 0; cdev->private->options.force &= (flags & CCWDEV_ALLOW_FORCE) == 0; cdev->private->options.mpath &= (flags & CCWDEV_DO_MULTIPATH) == 0; } /** * ccw_device_is_pathgroup() - determine if paths to this device are grouped * @cdev: ccw device * * Return non-zero if there is a path group, zero otherwise. */ int ccw_device_is_pathgroup(struct ccw_device *cdev) { return cdev->private->flags.pgroup; } EXPORT_SYMBOL(ccw_device_is_pathgroup); /** * ccw_device_is_multipath() - determine if device is operating in multipath mode * @cdev: ccw device * * Return non-zero if device is operating in multipath mode, zero otherwise. */ int ccw_device_is_multipath(struct ccw_device *cdev) { return cdev->private->flags.mpath; } EXPORT_SYMBOL(ccw_device_is_multipath); /** * ccw_device_clear() - terminate I/O request processing * @cdev: target ccw device * @intparm: interruption parameter to be returned upon conclusion of csch * * ccw_device_clear() calls csch on @cdev's subchannel. * Returns: * %0 on success, * -%ENODEV on device not operational, * -%EINVAL on invalid device state. * Context: * Interrupts disabled, ccw device lock held */ int ccw_device_clear(struct ccw_device *cdev, unsigned long intparm) { struct subchannel *sch; int ret; if (!cdev || !cdev->dev.parent) return -ENODEV; sch = to_subchannel(cdev->dev.parent); if (!sch->schib.pmcw.ena) return -EINVAL; if (cdev->private->state == DEV_STATE_NOT_OPER) return -ENODEV; if (cdev->private->state != DEV_STATE_ONLINE && cdev->private->state != DEV_STATE_W4SENSE) return -EINVAL; ret = cio_clear(sch); if (ret == 0) cdev->private->intparm = intparm; return ret; } /** * ccw_device_start_timeout_key() - start a s390 channel program with timeout and key * @cdev: target ccw device * @cpa: logical start address of channel program * @intparm: user specific interruption parameter; will be presented back to * @cdev's interrupt handler. Allows a device driver to associate * the interrupt with a particular I/O request. * @lpm: defines the channel path to be used for a specific I/O request. A * value of 0 will make cio use the opm. * @key: storage key to be used for the I/O * @flags: additional flags; defines the action to be performed for I/O * processing. * @expires: timeout value in jiffies * * Start a S/390 channel program. When the interrupt arrives, the * IRQ handler is called, either immediately, delayed (dev-end missing, * or sense required) or never (no IRQ handler registered). * This function notifies the device driver if the channel program has not * completed during the time specified by @expires. If a timeout occurs, the * channel program is terminated via xsch, hsch or csch, and the device's * interrupt handler will be called with an irb containing ERR_PTR(-%ETIMEDOUT). * The interruption handler will echo back the @intparm specified here, unless * another interruption parameter is specified by a subsequent invocation of * ccw_device_halt() or ccw_device_clear(). * Returns: * %0, if the operation was successful; * -%EBUSY, if the device is busy, or status pending; * -%EACCES, if no path specified in @lpm is operational; * -%ENODEV, if the device is not operational. * Context: * Interrupts disabled, ccw device lock held */ int ccw_device_start_timeout_key(struct ccw_device *cdev, struct ccw1 *cpa, unsigned long intparm, __u8 lpm, __u8 key, unsigned long flags, int expires) { struct subchannel *sch; int ret; if (!cdev || !cdev->dev.parent) return -ENODEV; sch = to_subchannel(cdev->dev.parent); if (!sch->schib.pmcw.ena) return -EINVAL; if (cdev->private->state == DEV_STATE_NOT_OPER) return -ENODEV; if (cdev->private->state == DEV_STATE_VERIFY || cdev->private->flags.doverify) { /* Remember to fake irb when finished. */ if (!cdev->private->flags.fake_irb) { cdev->private->flags.fake_irb = FAKE_CMD_IRB; cdev->private->intparm = intparm; CIO_MSG_EVENT(2, "fakeirb: queue device 0.%x.%04x intparm %lx type=%d\n", cdev->private->dev_id.ssid, cdev->private->dev_id.devno, intparm, cdev->private->flags.fake_irb); return 0; } else /* There's already a fake I/O around. */ return -EBUSY; } if (cdev->private->state != DEV_STATE_ONLINE || ((sch->schib.scsw.cmd.stctl & SCSW_STCTL_PRIM_STATUS) && !(sch->schib.scsw.cmd.stctl & SCSW_STCTL_SEC_STATUS))) return -EBUSY; ret = cio_set_options (sch, flags); if (ret) return ret; /* Adjust requested path mask to exclude unusable paths. */ if (lpm) { lpm &= sch->lpm; if (lpm == 0) return -EACCES; } ret = cio_start_key (sch, cpa, lpm, key); switch (ret) { case 0: cdev->private->intparm = intparm; if (expires) ccw_device_set_timeout(cdev, expires); break; case -EACCES: case -ENODEV: dev_fsm_event(cdev, DEV_EVENT_VERIFY); break; } return ret; } /** * ccw_device_start_key() - start a s390 channel program with key * @cdev: target ccw device * @cpa: logical start address of channel program * @intparm: user specific interruption parameter; will be presented back to * @cdev's interrupt handler. Allows a device driver to associate * the interrupt with a particular I/O request. * @lpm: defines the channel path to be used for a specific I/O request. A * value of 0 will make cio use the opm. * @key: storage key to be used for the I/O * @flags: additional flags; defines the action to be performed for I/O * processing. * * Start a S/390 channel program. When the interrupt arrives, the * IRQ handler is called, either immediately, delayed (dev-end missing, * or sense required) or never (no IRQ handler registered). * The interruption handler will echo back the @intparm specified here, unless * another interruption parameter is specified by a subsequent invocation of * ccw_device_halt() or ccw_device_clear(). * Returns: * %0, if the operation was successful; * -%EBUSY, if the device is busy, or status pending; * -%EACCES, if no path specified in @lpm is operational; * -%ENODEV, if the device is not operational. * Context: * Interrupts disabled, ccw device lock held */ int ccw_device_start_key(struct ccw_device *cdev, struct ccw1 *cpa, unsigned long intparm, __u8 lpm, __u8 key, unsigned long flags) { return ccw_device_start_timeout_key(cdev, cpa, intparm, lpm, key, flags, 0); } /** * ccw_device_start() - start a s390 channel program * @cdev: target ccw device * @cpa: logical start address of channel program * @intparm: user specific interruption parameter; will be presented back to * @cdev's interrupt handler. Allows a device driver to associate * the interrupt with a particular I/O request. * @lpm: defines the channel path to be used for a specific I/O request. A * value of 0 will make cio use the opm. * @flags: additional flags; defines the action to be performed for I/O * processing. * * Start a S/390 channel program. When the interrupt arrives, the * IRQ handler is called, either immediately, delayed (dev-end missing, * or sense required) or never (no IRQ handler registered). * The interruption handler will echo back the @intparm specified here, unless * another interruption parameter is specified by a subsequent invocation of * ccw_device_halt() or ccw_device_clear(). * Returns: * %0, if the operation was successful; * -%EBUSY, if the device is busy, or status pending; * -%EACCES, if no path specified in @lpm is operational; * -%ENODEV, if the device is not operational. * Context: * Interrupts disabled, ccw device lock held */ int ccw_device_start(struct ccw_device *cdev, struct ccw1 *cpa, unsigned long intparm, __u8 lpm, unsigned long flags) { return ccw_device_start_key(cdev, cpa, intparm, lpm, PAGE_DEFAULT_KEY, flags); } /** * ccw_device_start_timeout() - start a s390 channel program with timeout * @cdev: target ccw device * @cpa: logical start address of channel program * @intparm: user specific interruption parameter; will be presented back to * @cdev's interrupt handler. Allows a device driver to associate * the interrupt with a particular I/O request. * @lpm: defines the channel path to be used for a specific I/O request. A * value of 0 will make cio use the opm. * @flags: additional flags; defines the action to be performed for I/O * processing. * @expires: timeout value in jiffies * * Start a S/390 channel program. When the interrupt arrives, the * IRQ handler is called, either immediately, delayed (dev-end missing, * or sense required) or never (no IRQ handler registered). * This function notifies the device driver if the channel program has not * completed during the time specified by @expires. If a timeout occurs, the * channel program is terminated via xsch, hsch or csch, and the device's * interrupt handler will be called with an irb containing ERR_PTR(-%ETIMEDOUT). * The interruption handler will echo back the @intparm specified here, unless * another interruption parameter is specified by a subsequent invocation of * ccw_device_halt() or ccw_device_clear(). * Returns: * %0, if the operation was successful; * -%EBUSY, if the device is busy, or status pending; * -%EACCES, if no path specified in @lpm is operational; * -%ENODEV, if the device is not operational. * Context: * Interrupts disabled, ccw device lock held */ int ccw_device_start_timeout(struct ccw_device *cdev, struct ccw1 *cpa, unsigned long intparm, __u8 lpm, unsigned long flags, int expires) { return ccw_device_start_timeout_key(cdev, cpa, intparm, lpm, PAGE_DEFAULT_KEY, flags, expires); } /** * ccw_device_halt() - halt I/O request processing * @cdev: target ccw device * @intparm: interruption parameter to be returned upon conclusion of hsch * * ccw_device_halt() calls hsch on @cdev's subchannel. * The interruption handler will echo back the @intparm specified here, unless * another interruption parameter is specified by a subsequent invocation of * ccw_device_clear(). * Returns: * %0 on success, * -%ENODEV on device not operational, * -%EINVAL on invalid device state, * -%EBUSY on device busy or interrupt pending. * Context: * Interrupts disabled, ccw device lock held */ int ccw_device_halt(struct ccw_device *cdev, unsigned long intparm) { struct subchannel *sch; int ret; if (!cdev || !cdev->dev.parent) return -ENODEV; sch = to_subchannel(cdev->dev.parent); if (!sch->schib.pmcw.ena) return -EINVAL; if (cdev->private->state == DEV_STATE_NOT_OPER) return -ENODEV; if (cdev->private->state != DEV_STATE_ONLINE && cdev->private->state != DEV_STATE_W4SENSE) return -EINVAL; ret = cio_halt(sch); if (ret == 0) cdev->private->intparm = intparm; return ret; } /** * ccw_device_resume() - resume channel program execution * @cdev: target ccw device * * ccw_device_resume() calls rsch on @cdev's subchannel. * Returns: * %0 on success, * -%ENODEV on device not operational, * -%EINVAL on invalid device state, * -%EBUSY on device busy or interrupt pending. * Context: * Interrupts disabled, ccw device lock held */ int ccw_device_resume(struct ccw_device *cdev) { struct subchannel *sch; if (!cdev || !cdev->dev.parent) return -ENODEV; sch = to_subchannel(cdev->dev.parent); if (!sch->schib.pmcw.ena) return -EINVAL; if (cdev->private->state == DEV_STATE_NOT_OPER) return -ENODEV; if (cdev->private->state != DEV_STATE_ONLINE || !(sch->schib.scsw.cmd.actl & SCSW_ACTL_SUSPENDED)) return -EINVAL; return cio_resume(sch); } /** * ccw_device_get_ciw() - Search for CIW command in extended sense data. * @cdev: ccw device to inspect * @ct: command type to look for * * During SenseID, command information words (CIWs) describing special * commands available to the device may have been stored in the extended * sense data. This function searches for CIWs of a specified command * type in the extended sense data. * Returns: * %NULL if no extended sense data has been stored or if no CIW of the * specified command type could be found, * else a pointer to the CIW of the specified command type. */ struct ciw *ccw_device_get_ciw(struct ccw_device *cdev, __u32 ct) { int ciw_cnt; if (cdev->private->flags.esid == 0) return NULL; for (ciw_cnt = 0; ciw_cnt < MAX_CIWS; ciw_cnt++) if (cdev->private->dma_area->senseid.ciw[ciw_cnt].ct == ct) return cdev->private->dma_area->senseid.ciw + ciw_cnt; return NULL; } /** * ccw_device_get_path_mask() - get currently available paths * @cdev: ccw device to be queried * Returns: * %0 if no subchannel for the device is available, * else the mask of currently available paths for the ccw device's subchannel. */ __u8 ccw_device_get_path_mask(struct ccw_device *cdev) { struct subchannel *sch; if (!cdev->dev.parent) return 0; sch = to_subchannel(cdev->dev.parent); return sch->lpm; } /** * ccw_device_get_chp_desc() - return newly allocated channel-path descriptor * @cdev: device to obtain the descriptor for * @chp_idx: index of the channel path * * On success return a newly allocated copy of the channel-path description * data associated with the given channel path. Return %NULL on error. */ struct channel_path_desc_fmt0 *ccw_device_get_chp_desc(struct ccw_device *cdev, int chp_idx) { struct subchannel *sch; struct chp_id chpid; sch = to_subchannel(cdev->dev.parent); chp_id_init(&chpid); chpid.id = sch->schib.pmcw.chpid[chp_idx]; return chp_get_chp_desc(chpid); } /** * ccw_device_get_util_str() - return newly allocated utility strings * @cdev: device to obtain the utility strings for * @chp_idx: index of the channel path * * On success return a newly allocated copy of the utility strings * associated with the given channel path. Return %NULL on error. */ u8 *ccw_device_get_util_str(struct ccw_device *cdev, int chp_idx) { struct subchannel *sch = to_subchannel(cdev->dev.parent); struct channel_path *chp; struct chp_id chpid; u8 *util_str; chp_id_init(&chpid); chpid.id = sch->schib.pmcw.chpid[chp_idx]; chp = chpid_to_chp(chpid); util_str = kmalloc(sizeof(chp->desc_fmt3.util_str), GFP_KERNEL); if (!util_str) return NULL; mutex_lock(&chp->lock); memcpy(util_str, chp->desc_fmt3.util_str, sizeof(chp->desc_fmt3.util_str)); mutex_unlock(&chp->lock); return util_str; } /** * ccw_device_get_id() - obtain a ccw device id * @cdev: device to obtain the id for * @dev_id: where to fill in the values */ void ccw_device_get_id(struct ccw_device *cdev, struct ccw_dev_id *dev_id) { *dev_id = cdev->private->dev_id; } EXPORT_SYMBOL(ccw_device_get_id); /** * ccw_device_tm_start_timeout_key() - perform start function * @cdev: ccw device on which to perform the start function * @tcw: transport-command word to be started * @intparm: user defined parameter to be passed to the interrupt handler * @lpm: mask of paths to use * @key: storage key to use for storage access * @expires: time span in jiffies after which to abort request * * Start the tcw on the given ccw device. Return zero on success, non-zero * otherwise. */ int ccw_device_tm_start_timeout_key(struct ccw_device *cdev, struct tcw *tcw, unsigned long intparm, u8 lpm, u8 key, int expires) { struct subchannel *sch; int rc; sch = to_subchannel(cdev->dev.parent); if (!sch->schib.pmcw.ena) return -EINVAL; if (cdev->private->state == DEV_STATE_VERIFY) { /* Remember to fake irb when finished. */ if (!cdev->private->flags.fake_irb) { cdev->private->flags.fake_irb = FAKE_TM_IRB; cdev->private->intparm = intparm; CIO_MSG_EVENT(2, "fakeirb: queue device 0.%x.%04x intparm %lx type=%d\n", cdev->private->dev_id.ssid, cdev->private->dev_id.devno, intparm, cdev->private->flags.fake_irb); return 0; } else /* There's already a fake I/O around. */ return -EBUSY; } if (cdev->private->state != DEV_STATE_ONLINE) return -EIO; /* Adjust requested path mask to exclude unusable paths. */ if (lpm) { lpm &= sch->lpm; if (lpm == 0) return -EACCES; } rc = cio_tm_start_key(sch, tcw, lpm, key); if (rc == 0) { cdev->private->intparm = intparm; if (expires) ccw_device_set_timeout(cdev, expires); } return rc; } EXPORT_SYMBOL(ccw_device_tm_start_timeout_key); /** * ccw_device_tm_start_key() - perform start function * @cdev: ccw device on which to perform the start function * @tcw: transport-command word to be started * @intparm: user defined parameter to be passed to the interrupt handler * @lpm: mask of paths to use * @key: storage key to use for storage access * * Start the tcw on the given ccw device. Return zero on success, non-zero * otherwise. */ int ccw_device_tm_start_key(struct ccw_device *cdev, struct tcw *tcw, unsigned long intparm, u8 lpm, u8 key) { return ccw_device_tm_start_timeout_key(cdev, tcw, intparm, lpm, key, 0); } EXPORT_SYMBOL(ccw_device_tm_start_key); /** * ccw_device_tm_start() - perform start function * @cdev: ccw device on which to perform the start function * @tcw: transport-command word to be started * @intparm: user defined parameter to be passed to the interrupt handler * @lpm: mask of paths to use * * Start the tcw on the given ccw device. Return zero on success, non-zero * otherwise. */ int ccw_device_tm_start(struct ccw_device *cdev, struct tcw *tcw, unsigned long intparm, u8 lpm) { return ccw_device_tm_start_key(cdev, tcw, intparm, lpm, PAGE_DEFAULT_KEY); } EXPORT_SYMBOL(ccw_device_tm_start); /** * ccw_device_tm_start_timeout() - perform start function * @cdev: ccw device on which to perform the start function * @tcw: transport-command word to be started * @intparm: user defined parameter to be passed to the interrupt handler * @lpm: mask of paths to use * @expires: time span in jiffies after which to abort request * * Start the tcw on the given ccw device. Return zero on success, non-zero * otherwise. */ int ccw_device_tm_start_timeout(struct ccw_device *cdev, struct tcw *tcw, unsigned long intparm, u8 lpm, int expires) { return ccw_device_tm_start_timeout_key(cdev, tcw, intparm, lpm, PAGE_DEFAULT_KEY, expires); } EXPORT_SYMBOL(ccw_device_tm_start_timeout); /** * ccw_device_get_mdc() - accumulate max data count * @cdev: ccw device for which the max data count is accumulated * @mask: mask of paths to use * * Return the number of 64K-bytes blocks all paths at least support * for a transport command. Return value 0 indicates failure. */ int ccw_device_get_mdc(struct ccw_device *cdev, u8 mask) { struct subchannel *sch = to_subchannel(cdev->dev.parent); struct channel_path *chp; struct chp_id chpid; int mdc = 0, i; /* Adjust requested path mask to excluded varied off paths. */ if (mask) mask &= sch->lpm; else mask = sch->lpm; chp_id_init(&chpid); for (i = 0; i < 8; i++) { if (!(mask & (0x80 >> i))) continue; chpid.id = sch->schib.pmcw.chpid[i]; chp = chpid_to_chp(chpid); if (!chp) continue; mutex_lock(&chp->lock); if (!chp->desc_fmt1.f) { mutex_unlock(&chp->lock); return 0; } if (!chp->desc_fmt1.r) mdc = 1; mdc = mdc ? min_t(int, mdc, chp->desc_fmt1.mdc) : chp->desc_fmt1.mdc; mutex_unlock(&chp->lock); } return mdc; } EXPORT_SYMBOL(ccw_device_get_mdc); /** * ccw_device_tm_intrg() - perform interrogate function * @cdev: ccw device on which to perform the interrogate function * * Perform an interrogate function on the given ccw device. Return zero on * success, non-zero otherwise. */ int ccw_device_tm_intrg(struct ccw_device *cdev) { struct subchannel *sch = to_subchannel(cdev->dev.parent); if (!sch->schib.pmcw.ena) return -EINVAL; if (cdev->private->state != DEV_STATE_ONLINE) return -EIO; if (!scsw_is_tm(&sch->schib.scsw) || !(scsw_actl(&sch->schib.scsw) & SCSW_ACTL_START_PEND)) return -EINVAL; return cio_tm_intrg(sch); } EXPORT_SYMBOL(ccw_device_tm_intrg); /** * ccw_device_get_schid() - obtain a subchannel id * @cdev: device to obtain the id for * @schid: where to fill in the values */ void ccw_device_get_schid(struct ccw_device *cdev, struct subchannel_id *schid) { struct subchannel *sch = to_subchannel(cdev->dev.parent); *schid = sch->schid; } EXPORT_SYMBOL_GPL(ccw_device_get_schid); /** * ccw_device_pnso() - Perform Network-Subchannel Operation * @cdev: device on which PNSO is performed * @pnso_area: request and response block for the operation * @oc: Operation Code * @resume_token: resume token for multiblock response * @cnc: Boolean change-notification control * * pnso_area must be allocated by the caller with get_zeroed_page(GFP_KERNEL) * * Returns 0 on success. */ int ccw_device_pnso(struct ccw_device *cdev, struct chsc_pnso_area *pnso_area, u8 oc, struct chsc_pnso_resume_token resume_token, int cnc) { struct subchannel_id schid; ccw_device_get_schid(cdev, &schid); return chsc_pnso(schid, pnso_area, oc, resume_token, cnc); } EXPORT_SYMBOL_GPL(ccw_device_pnso); /** * ccw_device_get_cssid() - obtain Channel Subsystem ID * @cdev: device to obtain the CSSID for * @cssid: The resulting Channel Subsystem ID */ int ccw_device_get_cssid(struct ccw_device *cdev, u8 *cssid) { struct device *sch_dev = cdev->dev.parent; struct channel_subsystem *css = to_css(sch_dev->parent); if (css->id_valid) *cssid = css->cssid; return css->id_valid ? 0 : -ENODEV; } EXPORT_SYMBOL_GPL(ccw_device_get_cssid); /** * ccw_device_get_iid() - obtain MIF-image ID * @cdev: device to obtain the MIF-image ID for * @iid: The resulting MIF-image ID */ int ccw_device_get_iid(struct ccw_device *cdev, u8 *iid) { struct device *sch_dev = cdev->dev.parent; struct channel_subsystem *css = to_css(sch_dev->parent); if (css->id_valid) *iid = css->iid; return css->id_valid ? 0 : -ENODEV; } EXPORT_SYMBOL_GPL(ccw_device_get_iid); /** * ccw_device_get_chpid() - obtain Channel Path ID * @cdev: device to obtain the Channel Path ID for * @chp_idx: Index of the channel path * @chpid: The resulting Channel Path ID */ int ccw_device_get_chpid(struct ccw_device *cdev, int chp_idx, u8 *chpid) { struct subchannel *sch = to_subchannel(cdev->dev.parent); int mask; if ((chp_idx < 0) || (chp_idx > 7)) return -EINVAL; mask = 0x80 >> chp_idx; if (!(sch->schib.pmcw.pim & mask)) return -ENODEV; *chpid = sch->schib.pmcw.chpid[chp_idx]; return 0; } EXPORT_SYMBOL_GPL(ccw_device_get_chpid); /** * ccw_device_get_chid() - obtain Channel ID associated with specified CHPID * @cdev: device to obtain the Channel ID for * @chp_idx: Index of the channel path * @chid: The resulting Channel ID */ int ccw_device_get_chid(struct ccw_device *cdev, int chp_idx, u16 *chid) { struct chp_id cssid_chpid; struct channel_path *chp; int rc; chp_id_init(&cssid_chpid); rc = ccw_device_get_chpid(cdev, chp_idx, &cssid_chpid.id); if (rc) return rc; chp = chpid_to_chp(cssid_chpid); if (!chp) return -ENODEV; mutex_lock(&chp->lock); if (chp->desc_fmt1.flags & 0x10) *chid = chp->desc_fmt1.chid; else rc = -ENODEV; mutex_unlock(&chp->lock); return rc; } EXPORT_SYMBOL_GPL(ccw_device_get_chid); /* * Allocate zeroed dma coherent 31 bit addressable memory using * the subchannels dma pool. Maximal size of allocation supported * is PAGE_SIZE. */ void *ccw_device_dma_zalloc(struct ccw_device *cdev, size_t size, dma32_t *dma_handle) { void *addr; if (!get_device(&cdev->dev)) return NULL; addr = __cio_gp_dma_zalloc(cdev->private->dma_pool, &cdev->dev, size, dma_handle); if (IS_ERR_OR_NULL(addr)) put_device(&cdev->dev); return addr; } EXPORT_SYMBOL(ccw_device_dma_zalloc); void ccw_device_dma_free(struct ccw_device *cdev, void *cpu_addr, size_t size) { if (!cpu_addr) return; cio_gp_dma_free(cdev->private->dma_pool, cpu_addr, size); put_device(&cdev->dev); } EXPORT_SYMBOL(ccw_device_dma_free); EXPORT_SYMBOL(ccw_device_set_options_mask); EXPORT_SYMBOL(ccw_device_set_options); EXPORT_SYMBOL(ccw_device_clear_options); EXPORT_SYMBOL(ccw_device_clear); EXPORT_SYMBOL(ccw_device_halt); EXPORT_SYMBOL(ccw_device_resume); EXPORT_SYMBOL(ccw_device_start_timeout); EXPORT_SYMBOL(ccw_device_start); EXPORT_SYMBOL(ccw_device_start_timeout_key); EXPORT_SYMBOL(ccw_device_start_key); EXPORT_SYMBOL(ccw_device_get_ciw); EXPORT_SYMBOL(ccw_device_get_path_mask); EXPORT_SYMBOL_GPL(ccw_device_get_chp_desc); EXPORT_SYMBOL_GPL(ccw_device_get_util_str);
Information contained on this website is for historical information purposes only and does not indicate or represent copyright ownership.
Created with Cregit http://github.com/cregit/cregit
Version 2.0-RC1