Author | Tokens | Token Proportion | Commits | Commit Proportion |
---|---|---|---|---|
Liu Jinsong | 1490 | 89.28% | 3 | 11.11% |
Jan Beulich | 47 | 2.82% | 2 | 7.41% |
Dan Carpenter | 36 | 2.16% | 1 | 3.70% |
Jeremy Fitzhardinge | 33 | 1.98% | 5 | 18.52% |
Andi Kleen | 22 | 1.32% | 4 | 14.81% |
Joe Perches | 17 | 1.02% | 1 | 3.70% |
Juergen Gross | 5 | 0.30% | 1 | 3.70% |
Hidetoshi Seto | 4 | 0.24% | 1 | 3.70% |
Huang Ying | 3 | 0.18% | 1 | 3.70% |
Dave Jones | 2 | 0.12% | 1 | 3.70% |
Linus Torvalds | 2 | 0.12% | 1 | 3.70% |
Sheng Yang | 2 | 0.12% | 1 | 3.70% |
Rafael J. Wysocki | 2 | 0.12% | 1 | 3.70% |
Kay Sievers | 1 | 0.06% | 1 | 3.70% |
Dmitriy Zavin | 1 | 0.06% | 1 | 3.70% |
Al Viro | 1 | 0.06% | 1 | 3.70% |
Ingo Molnar | 1 | 0.06% | 1 | 3.70% |
Total | 1669 | 27 |
/****************************************************************************** * mcelog.c * Driver for receiving and transferring machine check error infomation * * Copyright (c) 2012 Intel Corporation * Author: Liu, Jinsong <jinsong.liu@intel.com> * Author: Jiang, Yunhong <yunhong.jiang@intel.com> * Author: Ke, Liping <liping.ke@intel.com> * * This program is free software; you can redistribute it and/or * modify it under the terms of the GNU General Public License version 2 * as published by the Free Software Foundation; or, when distributed * separately from the Linux kernel or incorporated into other * software packages, subject to the following license: * * Permission is hereby granted, free of charge, to any person obtaining a copy * of this source file (the "Software"), to deal in the Software without * restriction, including without limitation the rights to use, copy, modify, * merge, publish, distribute, sublicense, and/or sell copies of the Software, * and to permit persons to whom the Software is furnished to do so, subject to * the following conditions: * * The above copyright notice and this permission notice shall be included in * all copies or substantial portions of the Software. * * THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR * IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY, * FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT. IN NO EVENT SHALL THE * AUTHORS OR COPYRIGHT HOLDERS BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER * LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING * FROM, OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER DEALINGS * IN THE SOFTWARE. */ #define pr_fmt(fmt) "xen_mcelog: " fmt #include <linux/init.h> #include <linux/types.h> #include <linux/kernel.h> #include <linux/slab.h> #include <linux/fs.h> #include <linux/device.h> #include <linux/miscdevice.h> #include <linux/uaccess.h> #include <linux/capability.h> #include <linux/poll.h> #include <linux/sched.h> #include <xen/interface/xen.h> #include <xen/events.h> #include <xen/interface/vcpu.h> #include <xen/xen.h> #include <asm/xen/hypercall.h> #include <asm/xen/hypervisor.h> static struct mc_info g_mi; static struct mcinfo_logical_cpu *g_physinfo; static uint32_t ncpus; static DEFINE_MUTEX(mcelog_lock); static struct xen_mce_log xen_mcelog = { .signature = XEN_MCE_LOG_SIGNATURE, .len = XEN_MCE_LOG_LEN, .recordlen = sizeof(struct xen_mce), }; static DEFINE_SPINLOCK(xen_mce_chrdev_state_lock); static int xen_mce_chrdev_open_count; /* #times opened */ static int xen_mce_chrdev_open_exclu; /* already open exclusive? */ static DECLARE_WAIT_QUEUE_HEAD(xen_mce_chrdev_wait); static int xen_mce_chrdev_open(struct inode *inode, struct file *file) { spin_lock(&xen_mce_chrdev_state_lock); if (xen_mce_chrdev_open_exclu || (xen_mce_chrdev_open_count && (file->f_flags & O_EXCL))) { spin_unlock(&xen_mce_chrdev_state_lock); return -EBUSY; } if (file->f_flags & O_EXCL) xen_mce_chrdev_open_exclu = 1; xen_mce_chrdev_open_count++; spin_unlock(&xen_mce_chrdev_state_lock); return nonseekable_open(inode, file); } static int xen_mce_chrdev_release(struct inode *inode, struct file *file) { spin_lock(&xen_mce_chrdev_state_lock); xen_mce_chrdev_open_count--; xen_mce_chrdev_open_exclu = 0; spin_unlock(&xen_mce_chrdev_state_lock); return 0; } static ssize_t xen_mce_chrdev_read(struct file *filp, char __user *ubuf, size_t usize, loff_t *off) { char __user *buf = ubuf; unsigned num; int i, err; mutex_lock(&mcelog_lock); num = xen_mcelog.next; /* Only supports full reads right now */ err = -EINVAL; if (*off != 0 || usize < XEN_MCE_LOG_LEN*sizeof(struct xen_mce)) goto out; err = 0; for (i = 0; i < num; i++) { struct xen_mce *m = &xen_mcelog.entry[i]; err |= copy_to_user(buf, m, sizeof(*m)); buf += sizeof(*m); } memset(xen_mcelog.entry, 0, num * sizeof(struct xen_mce)); xen_mcelog.next = 0; if (err) err = -EFAULT; out: mutex_unlock(&mcelog_lock); return err ? err : buf - ubuf; } static __poll_t xen_mce_chrdev_poll(struct file *file, poll_table *wait) { poll_wait(file, &xen_mce_chrdev_wait, wait); if (xen_mcelog.next) return EPOLLIN | EPOLLRDNORM; return 0; } static long xen_mce_chrdev_ioctl(struct file *f, unsigned int cmd, unsigned long arg) { int __user *p = (int __user *)arg; if (!capable(CAP_SYS_ADMIN)) return -EPERM; switch (cmd) { case MCE_GET_RECORD_LEN: return put_user(sizeof(struct xen_mce), p); case MCE_GET_LOG_LEN: return put_user(XEN_MCE_LOG_LEN, p); case MCE_GETCLEAR_FLAGS: { unsigned flags; do { flags = xen_mcelog.flags; } while (cmpxchg(&xen_mcelog.flags, flags, 0) != flags); return put_user(flags, p); } default: return -ENOTTY; } } static const struct file_operations xen_mce_chrdev_ops = { .open = xen_mce_chrdev_open, .release = xen_mce_chrdev_release, .read = xen_mce_chrdev_read, .poll = xen_mce_chrdev_poll, .unlocked_ioctl = xen_mce_chrdev_ioctl, .llseek = no_llseek, }; static struct miscdevice xen_mce_chrdev_device = { MISC_MCELOG_MINOR, "mcelog", &xen_mce_chrdev_ops, }; /* * Caller should hold the mcelog_lock */ static void xen_mce_log(struct xen_mce *mce) { unsigned entry; entry = xen_mcelog.next; /* * When the buffer fills up discard new entries. * Assume that the earlier errors are the more * interesting ones: */ if (entry >= XEN_MCE_LOG_LEN) { set_bit(XEN_MCE_OVERFLOW, (unsigned long *)&xen_mcelog.flags); return; } memcpy(xen_mcelog.entry + entry, mce, sizeof(struct xen_mce)); xen_mcelog.next++; } static int convert_log(struct mc_info *mi) { struct mcinfo_common *mic; struct mcinfo_global *mc_global; struct mcinfo_bank *mc_bank; struct xen_mce m; unsigned int i, j; mic = NULL; x86_mcinfo_lookup(&mic, mi, MC_TYPE_GLOBAL); if (unlikely(!mic)) { pr_warn("Failed to find global error info\n"); return -ENODEV; } memset(&m, 0, sizeof(struct xen_mce)); mc_global = (struct mcinfo_global *)mic; m.mcgstatus = mc_global->mc_gstatus; m.apicid = mc_global->mc_apicid; for (i = 0; i < ncpus; i++) if (g_physinfo[i].mc_apicid == m.apicid) break; if (unlikely(i == ncpus)) { pr_warn("Failed to match cpu with apicid %d\n", m.apicid); return -ENODEV; } m.socketid = g_physinfo[i].mc_chipid; m.cpu = m.extcpu = g_physinfo[i].mc_cpunr; m.cpuvendor = (__u8)g_physinfo[i].mc_vendor; for (j = 0; j < g_physinfo[i].mc_nmsrvals; ++j) switch (g_physinfo[i].mc_msrvalues[j].reg) { case MSR_IA32_MCG_CAP: m.mcgcap = g_physinfo[i].mc_msrvalues[j].value; break; case MSR_PPIN: case MSR_AMD_PPIN: m.ppin = g_physinfo[i].mc_msrvalues[j].value; break; } mic = NULL; x86_mcinfo_lookup(&mic, mi, MC_TYPE_BANK); if (unlikely(!mic)) { pr_warn("Fail to find bank error info\n"); return -ENODEV; } do { if ((!mic) || (mic->size == 0) || (mic->type != MC_TYPE_GLOBAL && mic->type != MC_TYPE_BANK && mic->type != MC_TYPE_EXTENDED && mic->type != MC_TYPE_RECOVERY)) break; if (mic->type == MC_TYPE_BANK) { mc_bank = (struct mcinfo_bank *)mic; m.misc = mc_bank->mc_misc; m.status = mc_bank->mc_status; m.addr = mc_bank->mc_addr; m.tsc = mc_bank->mc_tsc; m.bank = mc_bank->mc_bank; m.finished = 1; /*log this record*/ xen_mce_log(&m); } mic = x86_mcinfo_next(mic); } while (1); return 0; } static int mc_queue_handle(uint32_t flags) { struct xen_mc mc_op; int ret = 0; mc_op.cmd = XEN_MC_fetch; set_xen_guest_handle(mc_op.u.mc_fetch.data, &g_mi); do { mc_op.u.mc_fetch.flags = flags; ret = HYPERVISOR_mca(&mc_op); if (ret) { pr_err("Failed to fetch %surgent error log\n", flags == XEN_MC_URGENT ? "" : "non"); break; } if (mc_op.u.mc_fetch.flags & XEN_MC_NODATA || mc_op.u.mc_fetch.flags & XEN_MC_FETCHFAILED) break; else { ret = convert_log(&g_mi); if (ret) pr_warn("Failed to convert this error log, continue acking it anyway\n"); mc_op.u.mc_fetch.flags = flags | XEN_MC_ACK; ret = HYPERVISOR_mca(&mc_op); if (ret) { pr_err("Failed to ack previous error log\n"); break; } } } while (1); return ret; } /* virq handler for machine check error info*/ static void xen_mce_work_fn(struct work_struct *work) { int err; mutex_lock(&mcelog_lock); /* urgent mc_info */ err = mc_queue_handle(XEN_MC_URGENT); if (err) pr_err("Failed to handle urgent mc_info queue, continue handling nonurgent mc_info queue anyway\n"); /* nonurgent mc_info */ err = mc_queue_handle(XEN_MC_NONURGENT); if (err) pr_err("Failed to handle nonurgent mc_info queue\n"); /* wake processes polling /dev/mcelog */ wake_up_interruptible(&xen_mce_chrdev_wait); mutex_unlock(&mcelog_lock); } static DECLARE_WORK(xen_mce_work, xen_mce_work_fn); static irqreturn_t xen_mce_interrupt(int irq, void *dev_id) { schedule_work(&xen_mce_work); return IRQ_HANDLED; } static int bind_virq_for_mce(void) { int ret; struct xen_mc mc_op; memset(&mc_op, 0, sizeof(struct xen_mc)); /* Fetch physical CPU Numbers */ mc_op.cmd = XEN_MC_physcpuinfo; set_xen_guest_handle(mc_op.u.mc_physcpuinfo.info, g_physinfo); ret = HYPERVISOR_mca(&mc_op); if (ret) { pr_err("Failed to get CPU numbers\n"); return ret; } /* Fetch each CPU Physical Info for later reference*/ ncpus = mc_op.u.mc_physcpuinfo.ncpus; g_physinfo = kcalloc(ncpus, sizeof(struct mcinfo_logical_cpu), GFP_KERNEL); if (!g_physinfo) return -ENOMEM; set_xen_guest_handle(mc_op.u.mc_physcpuinfo.info, g_physinfo); ret = HYPERVISOR_mca(&mc_op); if (ret) { pr_err("Failed to get CPU info\n"); kfree(g_physinfo); return ret; } ret = bind_virq_to_irqhandler(VIRQ_MCA, 0, xen_mce_interrupt, 0, "mce", NULL); if (ret < 0) { pr_err("Failed to bind virq\n"); kfree(g_physinfo); return ret; } return 0; } static int __init xen_late_init_mcelog(void) { int ret; /* Only DOM0 is responsible for MCE logging */ if (!xen_initial_domain()) return -ENODEV; /* register character device /dev/mcelog for xen mcelog */ ret = misc_register(&xen_mce_chrdev_device); if (ret) return ret; ret = bind_virq_for_mce(); if (ret) goto deregister; pr_info("/dev/mcelog registered by Xen\n"); return 0; deregister: misc_deregister(&xen_mce_chrdev_device); return ret; } device_initcall(xen_late_init_mcelog);
Information contained on this website is for historical information purposes only and does not indicate or represent copyright ownership.
Created with Cregit http://github.com/cregit/cregit
Version 2.0-RC1