Author | Tokens | Token Proportion | Commits | Commit Proportion |
---|---|---|---|---|
K. Y. Srinivasan | 641 | 24.05% | 22 | 28.95% |
Hank Janssen | 455 | 17.07% | 5 | 6.58% |
Dexuan Cui | 435 | 16.32% | 5 | 6.58% |
Vitaly Kuznetsov | 309 | 11.59% | 5 | 6.58% |
Alex Ng | 229 | 8.59% | 5 | 6.58% |
Andres Beltran | 222 | 8.33% | 1 | 1.32% |
Vineeth Pillai | 119 | 4.47% | 2 | 2.63% |
Haiyang Zhang | 99 | 3.71% | 7 | 9.21% |
Vivek yadav | 71 | 2.66% | 1 | 1.32% |
Greg Kroah-Hartman | 30 | 1.13% | 10 | 13.16% |
Boqun Feng | 12 | 0.45% | 1 | 1.32% |
Arjan van de Ven | 10 | 0.38% | 1 | 1.32% |
Long Li | 7 | 0.26% | 1 | 1.32% |
Andrea Parri | 6 | 0.23% | 1 | 1.32% |
Jacob E Keller | 4 | 0.15% | 1 | 1.32% |
Himadri Pandya | 4 | 0.15% | 1 | 1.32% |
Linus Torvalds (pre-git) | 2 | 0.08% | 1 | 1.32% |
Bill Pemberton | 2 | 0.08% | 1 | 1.32% |
Yue haibing | 2 | 0.08% | 1 | 1.32% |
Thomas Gleixner | 2 | 0.08% | 1 | 1.32% |
Stephen Hemminger | 2 | 0.08% | 1 | 1.32% |
Linus Torvalds | 1 | 0.04% | 1 | 1.32% |
Colin Ian King | 1 | 0.04% | 1 | 1.32% |
Total | 2665 | 76 |
// SPDX-License-Identifier: GPL-2.0-only /* * Copyright (c) 2010, Microsoft Corporation. * * Authors: * Haiyang Zhang <haiyangz@microsoft.com> * Hank Janssen <hjanssen@microsoft.com> */ #define pr_fmt(fmt) KBUILD_MODNAME ": " fmt #include <linux/kernel.h> #include <linux/init.h> #include <linux/module.h> #include <linux/slab.h> #include <linux/sysctl.h> #include <linux/reboot.h> #include <linux/hyperv.h> #include <linux/clockchips.h> #include <linux/ptp_clock_kernel.h> #include <asm/mshyperv.h> #include "hyperv_vmbus.h" #define SD_MAJOR 3 #define SD_MINOR 0 #define SD_MINOR_1 1 #define SD_MINOR_2 2 #define SD_VERSION_3_1 (SD_MAJOR << 16 | SD_MINOR_1) #define SD_VERSION_3_2 (SD_MAJOR << 16 | SD_MINOR_2) #define SD_VERSION (SD_MAJOR << 16 | SD_MINOR) #define SD_MAJOR_1 1 #define SD_VERSION_1 (SD_MAJOR_1 << 16 | SD_MINOR) #define TS_MAJOR 4 #define TS_MINOR 0 #define TS_VERSION (TS_MAJOR << 16 | TS_MINOR) #define TS_MAJOR_1 1 #define TS_VERSION_1 (TS_MAJOR_1 << 16 | TS_MINOR) #define TS_MAJOR_3 3 #define TS_VERSION_3 (TS_MAJOR_3 << 16 | TS_MINOR) #define HB_MAJOR 3 #define HB_MINOR 0 #define HB_VERSION (HB_MAJOR << 16 | HB_MINOR) #define HB_MAJOR_1 1 #define HB_VERSION_1 (HB_MAJOR_1 << 16 | HB_MINOR) static int sd_srv_version; static int ts_srv_version; static int hb_srv_version; #define SD_VER_COUNT 4 static const int sd_versions[] = { SD_VERSION_3_2, SD_VERSION_3_1, SD_VERSION, SD_VERSION_1 }; #define TS_VER_COUNT 3 static const int ts_versions[] = { TS_VERSION, TS_VERSION_3, TS_VERSION_1 }; #define HB_VER_COUNT 2 static const int hb_versions[] = { HB_VERSION, HB_VERSION_1 }; #define FW_VER_COUNT 2 static const int fw_versions[] = { UTIL_FW_VERSION, UTIL_WS2K8_FW_VERSION }; /* * Send the "hibernate" udev event in a thread context. */ struct hibernate_work_context { struct work_struct work; struct hv_device *dev; }; static struct hibernate_work_context hibernate_context; static bool hibernation_supported; static void send_hibernate_uevent(struct work_struct *work) { char *uevent_env[2] = { "EVENT=hibernate", NULL }; struct hibernate_work_context *ctx; ctx = container_of(work, struct hibernate_work_context, work); kobject_uevent_env(&ctx->dev->device.kobj, KOBJ_CHANGE, uevent_env); pr_info("Sent hibernation uevent\n"); } static int hv_shutdown_init(struct hv_util_service *srv) { struct vmbus_channel *channel = srv->channel; INIT_WORK(&hibernate_context.work, send_hibernate_uevent); hibernate_context.dev = channel->device_obj; hibernation_supported = hv_is_hibernation_supported(); return 0; } static void shutdown_onchannelcallback(void *context); static struct hv_util_service util_shutdown = { .util_cb = shutdown_onchannelcallback, .util_init = hv_shutdown_init, }; static int hv_timesync_init(struct hv_util_service *srv); static int hv_timesync_pre_suspend(void); static void hv_timesync_deinit(void); static void timesync_onchannelcallback(void *context); static struct hv_util_service util_timesynch = { .util_cb = timesync_onchannelcallback, .util_init = hv_timesync_init, .util_pre_suspend = hv_timesync_pre_suspend, .util_deinit = hv_timesync_deinit, }; static void heartbeat_onchannelcallback(void *context); static struct hv_util_service util_heartbeat = { .util_cb = heartbeat_onchannelcallback, }; static struct hv_util_service util_kvp = { .util_cb = hv_kvp_onchannelcallback, .util_init = hv_kvp_init, .util_pre_suspend = hv_kvp_pre_suspend, .util_pre_resume = hv_kvp_pre_resume, .util_deinit = hv_kvp_deinit, }; static struct hv_util_service util_vss = { .util_cb = hv_vss_onchannelcallback, .util_init = hv_vss_init, .util_pre_suspend = hv_vss_pre_suspend, .util_pre_resume = hv_vss_pre_resume, .util_deinit = hv_vss_deinit, }; static struct hv_util_service util_fcopy = { .util_cb = hv_fcopy_onchannelcallback, .util_init = hv_fcopy_init, .util_pre_suspend = hv_fcopy_pre_suspend, .util_pre_resume = hv_fcopy_pre_resume, .util_deinit = hv_fcopy_deinit, }; static void perform_shutdown(struct work_struct *dummy) { orderly_poweroff(true); } static void perform_restart(struct work_struct *dummy) { orderly_reboot(); } /* * Perform the shutdown operation in a thread context. */ static DECLARE_WORK(shutdown_work, perform_shutdown); /* * Perform the restart operation in a thread context. */ static DECLARE_WORK(restart_work, perform_restart); static void shutdown_onchannelcallback(void *context) { struct vmbus_channel *channel = context; struct work_struct *work = NULL; u32 recvlen; u64 requestid; u8 *shut_txf_buf = util_shutdown.recv_buffer; struct shutdown_msg_data *shutdown_msg; struct icmsg_hdr *icmsghdrp; if (vmbus_recvpacket(channel, shut_txf_buf, HV_HYP_PAGE_SIZE, &recvlen, &requestid)) { pr_err_ratelimited("Shutdown request received. Could not read into shut txf buf\n"); return; } if (!recvlen) return; /* Ensure recvlen is big enough to read header data */ if (recvlen < ICMSG_HDR) { pr_err_ratelimited("Shutdown request received. Packet length too small: %d\n", recvlen); return; } icmsghdrp = (struct icmsg_hdr *)&shut_txf_buf[sizeof(struct vmbuspipe_hdr)]; if (icmsghdrp->icmsgtype == ICMSGTYPE_NEGOTIATE) { if (vmbus_prep_negotiate_resp(icmsghdrp, shut_txf_buf, recvlen, fw_versions, FW_VER_COUNT, sd_versions, SD_VER_COUNT, NULL, &sd_srv_version)) { pr_info("Shutdown IC version %d.%d\n", sd_srv_version >> 16, sd_srv_version & 0xFFFF); } } else if (icmsghdrp->icmsgtype == ICMSGTYPE_SHUTDOWN) { /* Ensure recvlen is big enough to contain shutdown_msg_data struct */ if (recvlen < ICMSG_HDR + sizeof(struct shutdown_msg_data)) { pr_err_ratelimited("Invalid shutdown msg data. Packet length too small: %u\n", recvlen); return; } shutdown_msg = (struct shutdown_msg_data *)&shut_txf_buf[ICMSG_HDR]; /* * shutdown_msg->flags can be 0(shut down), 2(reboot), * or 4(hibernate). It may bitwise-OR 1, which means * performing the request by force. Linux always tries * to perform the request by force. */ switch (shutdown_msg->flags) { case 0: case 1: icmsghdrp->status = HV_S_OK; work = &shutdown_work; pr_info("Shutdown request received - graceful shutdown initiated\n"); break; case 2: case 3: icmsghdrp->status = HV_S_OK; work = &restart_work; pr_info("Restart request received - graceful restart initiated\n"); break; case 4: case 5: pr_info("Hibernation request received\n"); icmsghdrp->status = hibernation_supported ? HV_S_OK : HV_E_FAIL; if (hibernation_supported) work = &hibernate_context.work; break; default: icmsghdrp->status = HV_E_FAIL; pr_info("Shutdown request received - Invalid request\n"); break; } } else { icmsghdrp->status = HV_E_FAIL; pr_err_ratelimited("Shutdown request received. Invalid msg type: %d\n", icmsghdrp->icmsgtype); } icmsghdrp->icflags = ICMSGHDRFLAG_TRANSACTION | ICMSGHDRFLAG_RESPONSE; vmbus_sendpacket(channel, shut_txf_buf, recvlen, requestid, VM_PKT_DATA_INBAND, 0); if (work) schedule_work(work); } /* * Set the host time in a process context. */ static struct work_struct adj_time_work; /* * The last time sample, received from the host. PTP device responds to * requests by using this data and the current partition-wide time reference * count. */ static struct { u64 host_time; u64 ref_time; spinlock_t lock; } host_ts; static inline u64 reftime_to_ns(u64 reftime) { return (reftime - WLTIMEDELTA) * 100; } /* * Hard coded threshold for host timesync delay: 600 seconds */ static const u64 HOST_TIMESYNC_DELAY_THRESH = 600 * (u64)NSEC_PER_SEC; static int hv_get_adj_host_time(struct timespec64 *ts) { u64 newtime, reftime, timediff_adj; unsigned long flags; int ret = 0; spin_lock_irqsave(&host_ts.lock, flags); reftime = hv_read_reference_counter(); /* * We need to let the caller know that last update from host * is older than the max allowable threshold. clock_gettime() * and PTP ioctl do not have a documented error that we could * return for this specific case. Use ESTALE to report this. */ timediff_adj = reftime - host_ts.ref_time; if (timediff_adj * 100 > HOST_TIMESYNC_DELAY_THRESH) { pr_warn_once("TIMESYNC IC: Stale time stamp, %llu nsecs old\n", (timediff_adj * 100)); ret = -ESTALE; } newtime = host_ts.host_time + timediff_adj; *ts = ns_to_timespec64(reftime_to_ns(newtime)); spin_unlock_irqrestore(&host_ts.lock, flags); return ret; } static void hv_set_host_time(struct work_struct *work) { struct timespec64 ts; if (!hv_get_adj_host_time(&ts)) do_settimeofday64(&ts); } /* * Synchronize time with host after reboot, restore, etc. * * ICTIMESYNCFLAG_SYNC flag bit indicates reboot, restore events of the VM. * After reboot the flag ICTIMESYNCFLAG_SYNC is included in the first time * message after the timesync channel is opened. Since the hv_utils module is * loaded after hv_vmbus, the first message is usually missed. This bit is * considered a hard request to discipline the clock. * * ICTIMESYNCFLAG_SAMPLE bit indicates a time sample from host. This is * typically used as a hint to the guest. The guest is under no obligation * to discipline the clock. */ static inline void adj_guesttime(u64 hosttime, u64 reftime, u8 adj_flags) { unsigned long flags; u64 cur_reftime; /* * Save the adjusted time sample from the host and the snapshot * of the current system time. */ spin_lock_irqsave(&host_ts.lock, flags); cur_reftime = hv_read_reference_counter(); host_ts.host_time = hosttime; host_ts.ref_time = cur_reftime; /* * TimeSync v4 messages contain reference time (guest's Hyper-V * clocksource read when the time sample was generated), we can * improve the precision by adding the delta between now and the * time of generation. For older protocols we set * reftime == cur_reftime on call. */ host_ts.host_time += (cur_reftime - reftime); spin_unlock_irqrestore(&host_ts.lock, flags); /* Schedule work to do do_settimeofday64() */ if (adj_flags & ICTIMESYNCFLAG_SYNC) schedule_work(&adj_time_work); } /* * Time Sync Channel message handler. */ static void timesync_onchannelcallback(void *context) { struct vmbus_channel *channel = context; u32 recvlen; u64 requestid; struct icmsg_hdr *icmsghdrp; struct ictimesync_data *timedatap; struct ictimesync_ref_data *refdata; u8 *time_txf_buf = util_timesynch.recv_buffer; /* * Drain the ring buffer and use the last packet to update * host_ts */ while (1) { int ret = vmbus_recvpacket(channel, time_txf_buf, HV_HYP_PAGE_SIZE, &recvlen, &requestid); if (ret) { pr_err_ratelimited("TimeSync IC pkt recv failed (Err: %d)\n", ret); break; } if (!recvlen) break; /* Ensure recvlen is big enough to read header data */ if (recvlen < ICMSG_HDR) { pr_err_ratelimited("Timesync request received. Packet length too small: %d\n", recvlen); break; } icmsghdrp = (struct icmsg_hdr *)&time_txf_buf[ sizeof(struct vmbuspipe_hdr)]; if (icmsghdrp->icmsgtype == ICMSGTYPE_NEGOTIATE) { if (vmbus_prep_negotiate_resp(icmsghdrp, time_txf_buf, recvlen, fw_versions, FW_VER_COUNT, ts_versions, TS_VER_COUNT, NULL, &ts_srv_version)) { pr_info("TimeSync IC version %d.%d\n", ts_srv_version >> 16, ts_srv_version & 0xFFFF); } } else if (icmsghdrp->icmsgtype == ICMSGTYPE_TIMESYNC) { if (ts_srv_version > TS_VERSION_3) { /* Ensure recvlen is big enough to read ictimesync_ref_data */ if (recvlen < ICMSG_HDR + sizeof(struct ictimesync_ref_data)) { pr_err_ratelimited("Invalid ictimesync ref data. Length too small: %u\n", recvlen); break; } refdata = (struct ictimesync_ref_data *)&time_txf_buf[ICMSG_HDR]; adj_guesttime(refdata->parenttime, refdata->vmreferencetime, refdata->flags); } else { /* Ensure recvlen is big enough to read ictimesync_data */ if (recvlen < ICMSG_HDR + sizeof(struct ictimesync_data)) { pr_err_ratelimited("Invalid ictimesync data. Length too small: %u\n", recvlen); break; } timedatap = (struct ictimesync_data *)&time_txf_buf[ICMSG_HDR]; adj_guesttime(timedatap->parenttime, hv_read_reference_counter(), timedatap->flags); } } else { icmsghdrp->status = HV_E_FAIL; pr_err_ratelimited("Timesync request received. Invalid msg type: %d\n", icmsghdrp->icmsgtype); } icmsghdrp->icflags = ICMSGHDRFLAG_TRANSACTION | ICMSGHDRFLAG_RESPONSE; vmbus_sendpacket(channel, time_txf_buf, recvlen, requestid, VM_PKT_DATA_INBAND, 0); } } /* * Heartbeat functionality. * Every two seconds, Hyper-V send us a heartbeat request message. * we respond to this message, and Hyper-V knows we are alive. */ static void heartbeat_onchannelcallback(void *context) { struct vmbus_channel *channel = context; u32 recvlen; u64 requestid; struct icmsg_hdr *icmsghdrp; struct heartbeat_msg_data *heartbeat_msg; u8 *hbeat_txf_buf = util_heartbeat.recv_buffer; while (1) { if (vmbus_recvpacket(channel, hbeat_txf_buf, HV_HYP_PAGE_SIZE, &recvlen, &requestid)) { pr_err_ratelimited("Heartbeat request received. Could not read into hbeat txf buf\n"); return; } if (!recvlen) break; /* Ensure recvlen is big enough to read header data */ if (recvlen < ICMSG_HDR) { pr_err_ratelimited("Heartbeat request received. Packet length too small: %d\n", recvlen); break; } icmsghdrp = (struct icmsg_hdr *)&hbeat_txf_buf[ sizeof(struct vmbuspipe_hdr)]; if (icmsghdrp->icmsgtype == ICMSGTYPE_NEGOTIATE) { if (vmbus_prep_negotiate_resp(icmsghdrp, hbeat_txf_buf, recvlen, fw_versions, FW_VER_COUNT, hb_versions, HB_VER_COUNT, NULL, &hb_srv_version)) { pr_info("Heartbeat IC version %d.%d\n", hb_srv_version >> 16, hb_srv_version & 0xFFFF); } } else if (icmsghdrp->icmsgtype == ICMSGTYPE_HEARTBEAT) { /* * Ensure recvlen is big enough to read seq_num. Reserved area is not * included in the check as the host may not fill it up entirely */ if (recvlen < ICMSG_HDR + sizeof(u64)) { pr_err_ratelimited("Invalid heartbeat msg data. Length too small: %u\n", recvlen); break; } heartbeat_msg = (struct heartbeat_msg_data *)&hbeat_txf_buf[ICMSG_HDR]; heartbeat_msg->seq_num += 1; } else { icmsghdrp->status = HV_E_FAIL; pr_err_ratelimited("Heartbeat request received. Invalid msg type: %d\n", icmsghdrp->icmsgtype); } icmsghdrp->icflags = ICMSGHDRFLAG_TRANSACTION | ICMSGHDRFLAG_RESPONSE; vmbus_sendpacket(channel, hbeat_txf_buf, recvlen, requestid, VM_PKT_DATA_INBAND, 0); } } #define HV_UTIL_RING_SEND_SIZE VMBUS_RING_SIZE(3 * HV_HYP_PAGE_SIZE) #define HV_UTIL_RING_RECV_SIZE VMBUS_RING_SIZE(3 * HV_HYP_PAGE_SIZE) static int util_probe(struct hv_device *dev, const struct hv_vmbus_device_id *dev_id) { struct hv_util_service *srv = (struct hv_util_service *)dev_id->driver_data; int ret; srv->recv_buffer = kmalloc(HV_HYP_PAGE_SIZE * 4, GFP_KERNEL); if (!srv->recv_buffer) return -ENOMEM; srv->channel = dev->channel; if (srv->util_init) { ret = srv->util_init(srv); if (ret) { ret = -ENODEV; goto error1; } } /* * The set of services managed by the util driver are not performance * critical and do not need batched reading. Furthermore, some services * such as KVP can only handle one message from the host at a time. * Turn off batched reading for all util drivers before we open the * channel. */ set_channel_read_mode(dev->channel, HV_CALL_DIRECT); hv_set_drvdata(dev, srv); ret = vmbus_open(dev->channel, HV_UTIL_RING_SEND_SIZE, HV_UTIL_RING_RECV_SIZE, NULL, 0, srv->util_cb, dev->channel); if (ret) goto error; return 0; error: if (srv->util_deinit) srv->util_deinit(); error1: kfree(srv->recv_buffer); return ret; } static int util_remove(struct hv_device *dev) { struct hv_util_service *srv = hv_get_drvdata(dev); if (srv->util_deinit) srv->util_deinit(); vmbus_close(dev->channel); kfree(srv->recv_buffer); return 0; } /* * When we're in util_suspend(), all the userspace processes have been frozen * (refer to hibernate() -> freeze_processes()). The userspace is thawed only * after the whole resume procedure, including util_resume(), finishes. */ static int util_suspend(struct hv_device *dev) { struct hv_util_service *srv = hv_get_drvdata(dev); int ret = 0; if (srv->util_pre_suspend) { ret = srv->util_pre_suspend(); if (ret) return ret; } vmbus_close(dev->channel); return 0; } static int util_resume(struct hv_device *dev) { struct hv_util_service *srv = hv_get_drvdata(dev); int ret = 0; if (srv->util_pre_resume) { ret = srv->util_pre_resume(); if (ret) return ret; } ret = vmbus_open(dev->channel, HV_UTIL_RING_SEND_SIZE, HV_UTIL_RING_RECV_SIZE, NULL, 0, srv->util_cb, dev->channel); return ret; } static const struct hv_vmbus_device_id id_table[] = { /* Shutdown guid */ { HV_SHUTDOWN_GUID, .driver_data = (unsigned long)&util_shutdown }, /* Time synch guid */ { HV_TS_GUID, .driver_data = (unsigned long)&util_timesynch }, /* Heartbeat guid */ { HV_HEART_BEAT_GUID, .driver_data = (unsigned long)&util_heartbeat }, /* KVP guid */ { HV_KVP_GUID, .driver_data = (unsigned long)&util_kvp }, /* VSS GUID */ { HV_VSS_GUID, .driver_data = (unsigned long)&util_vss }, /* File copy GUID */ { HV_FCOPY_GUID, .driver_data = (unsigned long)&util_fcopy }, { }, }; MODULE_DEVICE_TABLE(vmbus, id_table); /* The one and only one */ static struct hv_driver util_drv = { .name = "hv_utils", .id_table = id_table, .probe = util_probe, .remove = util_remove, .suspend = util_suspend, .resume = util_resume, .driver = { .probe_type = PROBE_PREFER_ASYNCHRONOUS, }, }; static int hv_ptp_enable(struct ptp_clock_info *info, struct ptp_clock_request *request, int on) { return -EOPNOTSUPP; } static int hv_ptp_settime(struct ptp_clock_info *p, const struct timespec64 *ts) { return -EOPNOTSUPP; } static int hv_ptp_adjfine(struct ptp_clock_info *ptp, long delta) { return -EOPNOTSUPP; } static int hv_ptp_adjtime(struct ptp_clock_info *ptp, s64 delta) { return -EOPNOTSUPP; } static int hv_ptp_gettime(struct ptp_clock_info *info, struct timespec64 *ts) { return hv_get_adj_host_time(ts); } static struct ptp_clock_info ptp_hyperv_info = { .name = "hyperv", .enable = hv_ptp_enable, .adjtime = hv_ptp_adjtime, .adjfine = hv_ptp_adjfine, .gettime64 = hv_ptp_gettime, .settime64 = hv_ptp_settime, .owner = THIS_MODULE, }; static struct ptp_clock *hv_ptp_clock; static int hv_timesync_init(struct hv_util_service *srv) { spin_lock_init(&host_ts.lock); INIT_WORK(&adj_time_work, hv_set_host_time); /* * ptp_clock_register() returns NULL when CONFIG_PTP_1588_CLOCK is * disabled but the driver is still useful without the PTP device * as it still handles the ICTIMESYNCFLAG_SYNC case. */ hv_ptp_clock = ptp_clock_register(&ptp_hyperv_info, NULL); if (IS_ERR_OR_NULL(hv_ptp_clock)) { pr_err("cannot register PTP clock: %d\n", PTR_ERR_OR_ZERO(hv_ptp_clock)); hv_ptp_clock = NULL; } return 0; } static void hv_timesync_cancel_work(void) { cancel_work_sync(&adj_time_work); } static int hv_timesync_pre_suspend(void) { hv_timesync_cancel_work(); return 0; } static void hv_timesync_deinit(void) { if (hv_ptp_clock) ptp_clock_unregister(hv_ptp_clock); hv_timesync_cancel_work(); } static int __init init_hyperv_utils(void) { pr_info("Registering HyperV Utility Driver\n"); return vmbus_driver_register(&util_drv); } static void exit_hyperv_utils(void) { pr_info("De-Registered HyperV Utility Driver\n"); vmbus_driver_unregister(&util_drv); } module_init(init_hyperv_utils); module_exit(exit_hyperv_utils); MODULE_DESCRIPTION("Hyper-V Utilities"); MODULE_LICENSE("GPL");
Information contained on this website is for historical information purposes only and does not indicate or represent copyright ownership.
Created with Cregit http://github.com/cregit/cregit
Version 2.0-RC1