Author | Tokens | Token Proportion | Commits | Commit Proportion |
---|---|---|---|---|
Aleksey Makarov | 1114 | 43.91% | 1 | 3.57% |
Yi Guo | 392 | 15.45% | 1 | 3.57% |
Sai Krishna | 271 | 10.68% | 1 | 3.57% |
Hariprasad Kelam | 234 | 9.22% | 2 | 7.14% |
Sunil Goutham | 216 | 8.51% | 11 | 39.29% |
Linu Cherian | 126 | 4.97% | 1 | 3.57% |
Naveen Mamindlapalli | 79 | 3.11% | 3 | 10.71% |
Christina Jacob | 36 | 1.42% | 2 | 7.14% |
Arnd Bergmann | 35 | 1.38% | 1 | 3.57% |
Geetha Sowjanya | 24 | 0.95% | 1 | 3.57% |
Tomasz Duszynski | 3 | 0.12% | 1 | 3.57% |
Jakub Kiciński | 3 | 0.12% | 1 | 3.57% |
Dan Carpenter | 2 | 0.08% | 1 | 3.57% |
Zheng Yongjun | 2 | 0.08% | 1 | 3.57% |
Total | 2537 | 28 |
// SPDX-License-Identifier: GPL-2.0 /* Marvell RVU Ethernet driver * * Copyright (C) 2020 Marvell. * */ #include <linux/module.h> #include "otx2_common.h" #include "otx2_ptp.h" static bool is_tstmp_atomic_update_supported(struct otx2_ptp *ptp) { struct ptp_get_cap_rsp *rsp; struct msg_req *req; int err; if (!ptp->nic) return false; mutex_lock(&ptp->nic->mbox.lock); req = otx2_mbox_alloc_msg_ptp_get_cap(&ptp->nic->mbox); if (!req) { mutex_unlock(&ptp->nic->mbox.lock); return false; } err = otx2_sync_mbox_msg(&ptp->nic->mbox); if (err) { mutex_unlock(&ptp->nic->mbox.lock); return false; } rsp = (struct ptp_get_cap_rsp *)otx2_mbox_get_rsp(&ptp->nic->mbox.mbox, 0, &req->hdr); mutex_unlock(&ptp->nic->mbox.lock); if (IS_ERR(rsp)) return false; if (rsp->cap & PTP_CAP_HW_ATOMIC_UPDATE) return true; return false; } static int otx2_ptp_hw_adjtime(struct ptp_clock_info *ptp_info, s64 delta) { struct otx2_ptp *ptp = container_of(ptp_info, struct otx2_ptp, ptp_info); struct otx2_nic *pfvf = ptp->nic; struct ptp_req *req; int rc; if (!ptp->nic) return -ENODEV; mutex_lock(&pfvf->mbox.lock); req = otx2_mbox_alloc_msg_ptp_op(&ptp->nic->mbox); if (!req) { mutex_unlock(&pfvf->mbox.lock); return -ENOMEM; } req->op = PTP_OP_ADJTIME; req->delta = delta; rc = otx2_sync_mbox_msg(&ptp->nic->mbox); mutex_unlock(&pfvf->mbox.lock); return rc; } static u64 otx2_ptp_get_clock(struct otx2_ptp *ptp) { struct ptp_req *req; struct ptp_rsp *rsp; int err; if (!ptp->nic) return 0; req = otx2_mbox_alloc_msg_ptp_op(&ptp->nic->mbox); if (!req) return 0; req->op = PTP_OP_GET_CLOCK; err = otx2_sync_mbox_msg(&ptp->nic->mbox); if (err) return 0; rsp = (struct ptp_rsp *)otx2_mbox_get_rsp(&ptp->nic->mbox.mbox, 0, &req->hdr); if (IS_ERR(rsp)) return 0; return rsp->clk; } static int otx2_ptp_hw_gettime(struct ptp_clock_info *ptp_info, struct timespec64 *ts) { struct otx2_ptp *ptp = container_of(ptp_info, struct otx2_ptp, ptp_info); u64 tstamp; tstamp = otx2_ptp_get_clock(ptp); *ts = ns_to_timespec64(tstamp); return 0; } static int otx2_ptp_hw_settime(struct ptp_clock_info *ptp_info, const struct timespec64 *ts) { struct otx2_ptp *ptp = container_of(ptp_info, struct otx2_ptp, ptp_info); struct otx2_nic *pfvf = ptp->nic; struct ptp_req *req; u64 nsec; int rc; if (!ptp->nic) return -ENODEV; nsec = timespec64_to_ns(ts); mutex_lock(&pfvf->mbox.lock); req = otx2_mbox_alloc_msg_ptp_op(&ptp->nic->mbox); if (!req) { mutex_unlock(&pfvf->mbox.lock); return -ENOMEM; } req->op = PTP_OP_SET_CLOCK; req->clk = nsec; rc = otx2_sync_mbox_msg(&ptp->nic->mbox); mutex_unlock(&pfvf->mbox.lock); return rc; } static int otx2_ptp_adjfine(struct ptp_clock_info *ptp_info, long scaled_ppm) { struct otx2_ptp *ptp = container_of(ptp_info, struct otx2_ptp, ptp_info); struct ptp_req *req; if (!ptp->nic) return -ENODEV; req = otx2_mbox_alloc_msg_ptp_op(&ptp->nic->mbox); if (!req) return -ENOMEM; req->op = PTP_OP_ADJFINE; req->scaled_ppm = scaled_ppm; return otx2_sync_mbox_msg(&ptp->nic->mbox); } static int ptp_set_thresh(struct otx2_ptp *ptp, u64 thresh) { struct ptp_req *req; if (!ptp->nic) return -ENODEV; req = otx2_mbox_alloc_msg_ptp_op(&ptp->nic->mbox); if (!req) return -ENOMEM; req->op = PTP_OP_SET_THRESH; req->thresh = thresh; return otx2_sync_mbox_msg(&ptp->nic->mbox); } static int ptp_pps_on(struct otx2_ptp *ptp, int on, u64 period) { struct ptp_req *req; if (!ptp->nic) return -ENODEV; req = otx2_mbox_alloc_msg_ptp_op(&ptp->nic->mbox); if (!req) return -ENOMEM; req->op = PTP_OP_PPS_ON; req->pps_on = on; req->period = period; return otx2_sync_mbox_msg(&ptp->nic->mbox); } static u64 ptp_cc_read(const struct cyclecounter *cc) { struct otx2_ptp *ptp = container_of(cc, struct otx2_ptp, cycle_counter); return otx2_ptp_get_clock(ptp); } static u64 ptp_tstmp_read(struct otx2_ptp *ptp) { struct ptp_req *req; struct ptp_rsp *rsp; int err; if (!ptp->nic) return 0; req = otx2_mbox_alloc_msg_ptp_op(&ptp->nic->mbox); if (!req) return 0; req->op = PTP_OP_GET_TSTMP; err = otx2_sync_mbox_msg(&ptp->nic->mbox); if (err) return 0; rsp = (struct ptp_rsp *)otx2_mbox_get_rsp(&ptp->nic->mbox.mbox, 0, &req->hdr); if (IS_ERR(rsp)) return 0; return rsp->clk; } static int otx2_ptp_tc_adjtime(struct ptp_clock_info *ptp_info, s64 delta) { struct otx2_ptp *ptp = container_of(ptp_info, struct otx2_ptp, ptp_info); struct otx2_nic *pfvf = ptp->nic; mutex_lock(&pfvf->mbox.lock); timecounter_adjtime(&ptp->time_counter, delta); mutex_unlock(&pfvf->mbox.lock); return 0; } static int otx2_ptp_tc_gettime(struct ptp_clock_info *ptp_info, struct timespec64 *ts) { struct otx2_ptp *ptp = container_of(ptp_info, struct otx2_ptp, ptp_info); u64 tstamp; mutex_lock(&ptp->nic->mbox.lock); tstamp = timecounter_read(&ptp->time_counter); mutex_unlock(&ptp->nic->mbox.lock); *ts = ns_to_timespec64(tstamp); return 0; } static int otx2_ptp_tc_settime(struct ptp_clock_info *ptp_info, const struct timespec64 *ts) { struct otx2_ptp *ptp = container_of(ptp_info, struct otx2_ptp, ptp_info); u64 nsec; nsec = timespec64_to_ns(ts); mutex_lock(&ptp->nic->mbox.lock); timecounter_init(&ptp->time_counter, &ptp->cycle_counter, nsec); mutex_unlock(&ptp->nic->mbox.lock); return 0; } static int otx2_ptp_verify_pin(struct ptp_clock_info *ptp, unsigned int pin, enum ptp_pin_function func, unsigned int chan) { switch (func) { case PTP_PF_NONE: case PTP_PF_EXTTS: case PTP_PF_PEROUT: break; case PTP_PF_PHYSYNC: return -1; } return 0; } static u64 otx2_ptp_hw_tstamp2time(const struct timecounter *time_counter, u64 tstamp) { /* On HW which supports atomic updates, timecounter is not initialized */ return tstamp; } static void otx2_ptp_extts_check(struct work_struct *work) { struct otx2_ptp *ptp = container_of(work, struct otx2_ptp, extts_work.work); struct ptp_clock_event event; u64 tstmp, new_thresh; mutex_lock(&ptp->nic->mbox.lock); tstmp = ptp_tstmp_read(ptp); mutex_unlock(&ptp->nic->mbox.lock); if (tstmp != ptp->last_extts) { event.type = PTP_CLOCK_EXTTS; event.index = 0; event.timestamp = ptp->ptp_tstamp2nsec(&ptp->time_counter, tstmp); ptp_clock_event(ptp->ptp_clock, &event); new_thresh = tstmp % 500000000; if (ptp->thresh != new_thresh) { mutex_lock(&ptp->nic->mbox.lock); ptp_set_thresh(ptp, new_thresh); mutex_unlock(&ptp->nic->mbox.lock); ptp->thresh = new_thresh; } ptp->last_extts = tstmp; } schedule_delayed_work(&ptp->extts_work, msecs_to_jiffies(200)); } static void otx2_sync_tstamp(struct work_struct *work) { struct otx2_ptp *ptp = container_of(work, struct otx2_ptp, synctstamp_work.work); struct otx2_nic *pfvf = ptp->nic; u64 tstamp; mutex_lock(&pfvf->mbox.lock); tstamp = otx2_ptp_get_clock(ptp); mutex_unlock(&pfvf->mbox.lock); ptp->tstamp = ptp->ptp_tstamp2nsec(&ptp->time_counter, tstamp); ptp->base_ns = tstamp % NSEC_PER_SEC; schedule_delayed_work(&ptp->synctstamp_work, msecs_to_jiffies(250)); } static int otx2_ptp_enable(struct ptp_clock_info *ptp_info, struct ptp_clock_request *rq, int on) { struct otx2_ptp *ptp = container_of(ptp_info, struct otx2_ptp, ptp_info); u64 period = 0; int pin; if (!ptp->nic) return -ENODEV; switch (rq->type) { case PTP_CLK_REQ_EXTTS: pin = ptp_find_pin(ptp->ptp_clock, PTP_PF_EXTTS, rq->extts.index); if (pin < 0) return -EBUSY; if (on) schedule_delayed_work(&ptp->extts_work, msecs_to_jiffies(200)); else cancel_delayed_work_sync(&ptp->extts_work); return 0; case PTP_CLK_REQ_PEROUT: if (rq->perout.flags) return -EOPNOTSUPP; if (rq->perout.index >= ptp_info->n_pins) return -EINVAL; if (on) { period = rq->perout.period.sec * NSEC_PER_SEC + rq->perout.period.nsec; ptp_pps_on(ptp, on, period); } else { ptp_pps_on(ptp, on, period); } return 0; default: break; } return -EOPNOTSUPP; } int otx2_ptp_init(struct otx2_nic *pfvf) { struct otx2_ptp *ptp_ptr; struct cyclecounter *cc; struct ptp_req *req; int err; if (is_otx2_lbkvf(pfvf->pdev)) { pfvf->ptp = NULL; return 0; } mutex_lock(&pfvf->mbox.lock); /* check if PTP block is available */ req = otx2_mbox_alloc_msg_ptp_op(&pfvf->mbox); if (!req) { mutex_unlock(&pfvf->mbox.lock); return -ENOMEM; } req->op = PTP_OP_GET_CLOCK; err = otx2_sync_mbox_msg(&pfvf->mbox); if (err) { mutex_unlock(&pfvf->mbox.lock); return err; } mutex_unlock(&pfvf->mbox.lock); ptp_ptr = kzalloc(sizeof(*ptp_ptr), GFP_KERNEL); if (!ptp_ptr) { err = -ENOMEM; goto error; } ptp_ptr->nic = pfvf; snprintf(ptp_ptr->extts_config.name, sizeof(ptp_ptr->extts_config.name), "TSTAMP"); ptp_ptr->extts_config.index = 0; ptp_ptr->extts_config.func = PTP_PF_NONE; ptp_ptr->ptp_info = (struct ptp_clock_info) { .owner = THIS_MODULE, .name = "OcteonTX2 PTP", .max_adj = 1000000000ull, .n_ext_ts = 1, .n_per_out = 1, .n_pins = 1, .pps = 0, .pin_config = &ptp_ptr->extts_config, .adjfine = otx2_ptp_adjfine, .enable = otx2_ptp_enable, .verify = otx2_ptp_verify_pin, }; /* Check whether hardware supports atomic updates to timestamp */ if (is_tstmp_atomic_update_supported(ptp_ptr)) { ptp_ptr->ptp_info.adjtime = otx2_ptp_hw_adjtime; ptp_ptr->ptp_info.gettime64 = otx2_ptp_hw_gettime; ptp_ptr->ptp_info.settime64 = otx2_ptp_hw_settime; ptp_ptr->ptp_tstamp2nsec = otx2_ptp_hw_tstamp2time; } else { ptp_ptr->ptp_info.adjtime = otx2_ptp_tc_adjtime; ptp_ptr->ptp_info.gettime64 = otx2_ptp_tc_gettime; ptp_ptr->ptp_info.settime64 = otx2_ptp_tc_settime; cc = &ptp_ptr->cycle_counter; cc->read = ptp_cc_read; cc->mask = CYCLECOUNTER_MASK(64); cc->mult = 1; cc->shift = 0; ptp_ptr->ptp_tstamp2nsec = timecounter_cyc2time; timecounter_init(&ptp_ptr->time_counter, &ptp_ptr->cycle_counter, ktime_to_ns(ktime_get_real())); } INIT_DELAYED_WORK(&ptp_ptr->extts_work, otx2_ptp_extts_check); ptp_ptr->ptp_clock = ptp_clock_register(&ptp_ptr->ptp_info, pfvf->dev); if (IS_ERR_OR_NULL(ptp_ptr->ptp_clock)) { err = ptp_ptr->ptp_clock ? PTR_ERR(ptp_ptr->ptp_clock) : -ENODEV; kfree(ptp_ptr); goto error; } if (is_dev_otx2(pfvf->pdev)) { ptp_ptr->convert_rx_ptp_tstmp = &otx2_ptp_convert_rx_timestamp; ptp_ptr->convert_tx_ptp_tstmp = &otx2_ptp_convert_tx_timestamp; } else { ptp_ptr->convert_rx_ptp_tstmp = &cn10k_ptp_convert_timestamp; ptp_ptr->convert_tx_ptp_tstmp = &cn10k_ptp_convert_timestamp; } INIT_DELAYED_WORK(&ptp_ptr->synctstamp_work, otx2_sync_tstamp); pfvf->ptp = ptp_ptr; error: return err; } EXPORT_SYMBOL_GPL(otx2_ptp_init); void otx2_ptp_destroy(struct otx2_nic *pfvf) { struct otx2_ptp *ptp = pfvf->ptp; if (!ptp) return; cancel_delayed_work(&pfvf->ptp->synctstamp_work); ptp_clock_unregister(ptp->ptp_clock); kfree(ptp); pfvf->ptp = NULL; } EXPORT_SYMBOL_GPL(otx2_ptp_destroy); int otx2_ptp_clock_index(struct otx2_nic *pfvf) { if (!pfvf->ptp) return -ENODEV; return ptp_clock_index(pfvf->ptp->ptp_clock); } EXPORT_SYMBOL_GPL(otx2_ptp_clock_index); int otx2_ptp_tstamp2time(struct otx2_nic *pfvf, u64 tstamp, u64 *tsns) { if (!pfvf->ptp) return -ENODEV; *tsns = pfvf->ptp->ptp_tstamp2nsec(&pfvf->ptp->time_counter, tstamp); return 0; } EXPORT_SYMBOL_GPL(otx2_ptp_tstamp2time); MODULE_AUTHOR("Sunil Goutham <sgoutham@marvell.com>"); MODULE_DESCRIPTION("Marvell RVU NIC PTP Driver"); MODULE_LICENSE("GPL v2");
Information contained on this website is for historical information purposes only and does not indicate or represent copyright ownership.
Created with Cregit http://github.com/cregit/cregit
Version 2.0-RC1