Author | Tokens | Token Proportion | Commits | Commit Proportion |
---|---|---|---|---|
Juergen Gross | 1228 | 62.88% | 3 | 27.27% |
Jeremy Fitzhardinge | 673 | 34.46% | 2 | 18.18% |
Daniel De Graaf | 19 | 0.97% | 1 | 9.09% |
Joe Perches | 13 | 0.67% | 1 | 9.09% |
Joao Martins | 10 | 0.51% | 1 | 9.09% |
Olaf Hering | 6 | 0.31% | 1 | 9.09% |
Michael S. Tsirkin | 3 | 0.15% | 1 | 9.09% |
Stefano Stabellini | 1 | 0.05% | 1 | 9.09% |
Total | 1953 | 11 |
/****************************************************************************** * xenbus_comms.c * * Low level code to talks to Xen Store: ringbuffer and event channel. * * Copyright (C) 2005 Rusty Russell, IBM Corporation * * This program is free software; you can redistribute it and/or * modify it under the terms of the GNU General Public License version 2 * as published by the Free Software Foundation; or, when distributed * separately from the Linux kernel or incorporated into other * software packages, subject to the following license: * * Permission is hereby granted, free of charge, to any person obtaining a copy * of this source file (the "Software"), to deal in the Software without * restriction, including without limitation the rights to use, copy, modify, * merge, publish, distribute, sublicense, and/or sell copies of the Software, * and to permit persons to whom the Software is furnished to do so, subject to * the following conditions: * * The above copyright notice and this permission notice shall be included in * all copies or substantial portions of the Software. * * THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR * IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY, * FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT. IN NO EVENT SHALL THE * AUTHORS OR COPYRIGHT HOLDERS BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER * LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING * FROM, OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER DEALINGS * IN THE SOFTWARE. */ #define pr_fmt(fmt) KBUILD_MODNAME ": " fmt #include <linux/wait.h> #include <linux/interrupt.h> #include <linux/kthread.h> #include <linux/sched.h> #include <linux/err.h> #include <xen/xenbus.h> #include <asm/xen/hypervisor.h> #include <xen/events.h> #include <xen/page.h> #include "xenbus.h" /* A list of replies. Currently only one will ever be outstanding. */ LIST_HEAD(xs_reply_list); /* A list of write requests. */ LIST_HEAD(xb_write_list); DECLARE_WAIT_QUEUE_HEAD(xb_waitq); DEFINE_MUTEX(xb_write_mutex); /* Protect xenbus reader thread against save/restore. */ DEFINE_MUTEX(xs_response_mutex); static int xenbus_irq; static struct task_struct *xenbus_task; static DECLARE_WORK(probe_work, xenbus_probe); static irqreturn_t wake_waiting(int irq, void *unused) { if (unlikely(xenstored_ready == 0)) { xenstored_ready = 1; schedule_work(&probe_work); } wake_up(&xb_waitq); return IRQ_HANDLED; } static int check_indexes(XENSTORE_RING_IDX cons, XENSTORE_RING_IDX prod) { return ((prod - cons) <= XENSTORE_RING_SIZE); } static void *get_output_chunk(XENSTORE_RING_IDX cons, XENSTORE_RING_IDX prod, char *buf, uint32_t *len) { *len = XENSTORE_RING_SIZE - MASK_XENSTORE_IDX(prod); if ((XENSTORE_RING_SIZE - (prod - cons)) < *len) *len = XENSTORE_RING_SIZE - (prod - cons); return buf + MASK_XENSTORE_IDX(prod); } static const void *get_input_chunk(XENSTORE_RING_IDX cons, XENSTORE_RING_IDX prod, const char *buf, uint32_t *len) { *len = XENSTORE_RING_SIZE - MASK_XENSTORE_IDX(cons); if ((prod - cons) < *len) *len = prod - cons; return buf + MASK_XENSTORE_IDX(cons); } static int xb_data_to_write(void) { struct xenstore_domain_interface *intf = xen_store_interface; return (intf->req_prod - intf->req_cons) != XENSTORE_RING_SIZE && !list_empty(&xb_write_list); } /** * xb_write - low level write * @data: buffer to send * @len: length of buffer * * Returns number of bytes written or -err. */ static int xb_write(const void *data, unsigned int len) { struct xenstore_domain_interface *intf = xen_store_interface; XENSTORE_RING_IDX cons, prod; unsigned int bytes = 0; while (len != 0) { void *dst; unsigned int avail; /* Read indexes, then verify. */ cons = intf->req_cons; prod = intf->req_prod; if (!check_indexes(cons, prod)) { intf->req_cons = intf->req_prod = 0; return -EIO; } if (!xb_data_to_write()) return bytes; /* Must write data /after/ reading the consumer index. */ virt_mb(); dst = get_output_chunk(cons, prod, intf->req, &avail); if (avail == 0) continue; if (avail > len) avail = len; memcpy(dst, data, avail); data += avail; len -= avail; bytes += avail; /* Other side must not see new producer until data is there. */ virt_wmb(); intf->req_prod += avail; /* Implies mb(): other side will see the updated producer. */ if (prod <= intf->req_cons) notify_remote_via_evtchn(xen_store_evtchn); } return bytes; } static int xb_data_to_read(void) { struct xenstore_domain_interface *intf = xen_store_interface; return (intf->rsp_cons != intf->rsp_prod); } static int xb_read(void *data, unsigned int len) { struct xenstore_domain_interface *intf = xen_store_interface; XENSTORE_RING_IDX cons, prod; unsigned int bytes = 0; while (len != 0) { unsigned int avail; const char *src; /* Read indexes, then verify. */ cons = intf->rsp_cons; prod = intf->rsp_prod; if (cons == prod) return bytes; if (!check_indexes(cons, prod)) { intf->rsp_cons = intf->rsp_prod = 0; return -EIO; } src = get_input_chunk(cons, prod, intf->rsp, &avail); if (avail == 0) continue; if (avail > len) avail = len; /* Must read data /after/ reading the producer index. */ virt_rmb(); memcpy(data, src, avail); data += avail; len -= avail; bytes += avail; /* Other side must not see free space until we've copied out */ virt_mb(); intf->rsp_cons += avail; /* Implies mb(): other side will see the updated consumer. */ if (intf->rsp_prod - cons >= XENSTORE_RING_SIZE) notify_remote_via_evtchn(xen_store_evtchn); } return bytes; } static int process_msg(void) { static struct { struct xsd_sockmsg msg; char *body; union { void *alloc; struct xs_watch_event *watch; }; bool in_msg; bool in_hdr; unsigned int read; } state; struct xb_req_data *req; int err; unsigned int len; if (!state.in_msg) { state.in_msg = true; state.in_hdr = true; state.read = 0; /* * We must disallow save/restore while reading a message. * A partial read across s/r leaves us out of sync with * xenstored. * xs_response_mutex is locked as long as we are processing one * message. state.in_msg will be true as long as we are holding * the lock here. */ mutex_lock(&xs_response_mutex); if (!xb_data_to_read()) { /* We raced with save/restore: pending data 'gone'. */ mutex_unlock(&xs_response_mutex); state.in_msg = false; return 0; } } if (state.in_hdr) { if (state.read != sizeof(state.msg)) { err = xb_read((void *)&state.msg + state.read, sizeof(state.msg) - state.read); if (err < 0) goto out; state.read += err; if (state.read != sizeof(state.msg)) return 0; if (state.msg.len > XENSTORE_PAYLOAD_MAX) { err = -EINVAL; goto out; } } len = state.msg.len + 1; if (state.msg.type == XS_WATCH_EVENT) len += sizeof(*state.watch); state.alloc = kmalloc(len, GFP_NOIO | __GFP_HIGH); if (!state.alloc) return -ENOMEM; if (state.msg.type == XS_WATCH_EVENT) state.body = state.watch->body; else state.body = state.alloc; state.in_hdr = false; state.read = 0; } err = xb_read(state.body + state.read, state.msg.len - state.read); if (err < 0) goto out; state.read += err; if (state.read != state.msg.len) return 0; state.body[state.msg.len] = '\0'; if (state.msg.type == XS_WATCH_EVENT) { state.watch->len = state.msg.len; err = xs_watch_msg(state.watch); } else { err = -ENOENT; mutex_lock(&xb_write_mutex); list_for_each_entry(req, &xs_reply_list, list) { if (req->msg.req_id == state.msg.req_id) { list_del(&req->list); err = 0; break; } } mutex_unlock(&xb_write_mutex); if (err) goto out; if (req->state == xb_req_state_wait_reply) { req->msg.req_id = req->caller_req_id; req->msg.type = state.msg.type; req->msg.len = state.msg.len; req->body = state.body; req->state = xb_req_state_got_reply; req->cb(req); } else kfree(req); } mutex_unlock(&xs_response_mutex); state.in_msg = false; state.alloc = NULL; return err; out: mutex_unlock(&xs_response_mutex); state.in_msg = false; kfree(state.alloc); state.alloc = NULL; return err; } static int process_writes(void) { static struct { struct xb_req_data *req; int idx; unsigned int written; } state; void *base; unsigned int len; int err = 0; if (!xb_data_to_write()) return 0; mutex_lock(&xb_write_mutex); if (!state.req) { state.req = list_first_entry(&xb_write_list, struct xb_req_data, list); state.idx = -1; state.written = 0; } if (state.req->state == xb_req_state_aborted) goto out_err; while (state.idx < state.req->num_vecs) { if (state.idx < 0) { base = &state.req->msg; len = sizeof(state.req->msg); } else { base = state.req->vec[state.idx].iov_base; len = state.req->vec[state.idx].iov_len; } err = xb_write(base + state.written, len - state.written); if (err < 0) goto out_err; state.written += err; if (state.written != len) goto out; state.idx++; state.written = 0; } list_del(&state.req->list); state.req->state = xb_req_state_wait_reply; list_add_tail(&state.req->list, &xs_reply_list); state.req = NULL; out: mutex_unlock(&xb_write_mutex); return 0; out_err: state.req->msg.type = XS_ERROR; state.req->err = err; list_del(&state.req->list); if (state.req->state == xb_req_state_aborted) kfree(state.req); else { state.req->state = xb_req_state_got_reply; wake_up(&state.req->wq); } mutex_unlock(&xb_write_mutex); state.req = NULL; return err; } static int xb_thread_work(void) { return xb_data_to_read() || xb_data_to_write(); } static int xenbus_thread(void *unused) { int err; while (!kthread_should_stop()) { if (wait_event_interruptible(xb_waitq, xb_thread_work())) continue; err = process_msg(); if (err == -ENOMEM) schedule(); else if (err) pr_warn_ratelimited("error %d while reading message\n", err); err = process_writes(); if (err) pr_warn_ratelimited("error %d while writing message\n", err); } xenbus_task = NULL; return 0; } /** * xb_init_comms - Set up interrupt handler off store event channel. */ int xb_init_comms(void) { struct xenstore_domain_interface *intf = xen_store_interface; if (intf->req_prod != intf->req_cons) pr_err("request ring is not quiescent (%08x:%08x)!\n", intf->req_cons, intf->req_prod); if (intf->rsp_prod != intf->rsp_cons) { pr_warn("response ring is not quiescent (%08x:%08x): fixing up\n", intf->rsp_cons, intf->rsp_prod); /* breaks kdump */ if (!reset_devices) intf->rsp_cons = intf->rsp_prod; } if (xenbus_irq) { /* Already have an irq; assume we're resuming */ rebind_evtchn_irq(xen_store_evtchn, xenbus_irq); } else { int err; err = bind_evtchn_to_irqhandler(xen_store_evtchn, wake_waiting, 0, "xenbus", &xb_waitq); if (err < 0) { pr_err("request irq failed %i\n", err); return err; } xenbus_irq = err; if (!xenbus_task) { xenbus_task = kthread_run(xenbus_thread, NULL, "xenbus"); if (IS_ERR(xenbus_task)) return PTR_ERR(xenbus_task); } } return 0; } void xb_deinit_comms(void) { unbind_from_irqhandler(xenbus_irq, &xb_waitq); xenbus_irq = 0; }
Information contained on this website is for historical information purposes only and does not indicate or represent copyright ownership.
Created with Cregit http://github.com/cregit/cregit
Version 2.0-RC1