Author | Tokens | Token Proportion | Commits | Commit Proportion |
---|---|---|---|---|
Bob Pearson | 773 | 71.71% | 10 | 66.67% |
Moni Shoua | 274 | 25.42% | 1 | 6.67% |
Andrew Boyer | 26 | 2.41% | 1 | 6.67% |
Allen Pais | 3 | 0.28% | 1 | 6.67% |
Gustavo A. R. Silva | 1 | 0.09% | 1 | 6.67% |
Kamal Heib | 1 | 0.09% | 1 | 6.67% |
Total | 1078 | 15 |
// SPDX-License-Identifier: GPL-2.0 OR Linux-OpenIB /* * Copyright (c) 2016 Mellanox Technologies Ltd. All rights reserved. * Copyright (c) 2015 System Fabric Works, Inc. All rights reserved. */ #include "rxe.h" static struct workqueue_struct *rxe_wq; int rxe_alloc_wq(void) { rxe_wq = alloc_workqueue("rxe_wq", WQ_UNBOUND, WQ_MAX_ACTIVE); if (!rxe_wq) return -ENOMEM; return 0; } void rxe_destroy_wq(void) { destroy_workqueue(rxe_wq); } /* Check if task is idle i.e. not running, not scheduled in * work queue and not draining. If so move to busy to * reserve a slot in do_task() by setting to busy and taking * a qp reference to cover the gap from now until the task finishes. * state will move out of busy if task returns a non zero value * in do_task(). If state is already busy it is raised to armed * to indicate to do_task that additional pass should be made * over the task. * Context: caller should hold task->lock. * Returns: true if state transitioned from idle to busy else false. */ static bool __reserve_if_idle(struct rxe_task *task) { WARN_ON(rxe_read(task->qp) <= 0); if (task->state == TASK_STATE_IDLE) { rxe_get(task->qp); task->state = TASK_STATE_BUSY; task->num_sched++; return true; } if (task->state == TASK_STATE_BUSY) task->state = TASK_STATE_ARMED; return false; } /* check if task is idle or drained and not currently * scheduled in the work queue. This routine is * called by rxe_cleanup_task or rxe_disable_task to * see if the queue is empty. * Context: caller should hold task->lock. * Returns true if done else false. */ static bool __is_done(struct rxe_task *task) { if (work_pending(&task->work)) return false; if (task->state == TASK_STATE_IDLE || task->state == TASK_STATE_DRAINED) { return true; } return false; } /* a locked version of __is_done */ static bool is_done(struct rxe_task *task) { unsigned long flags; int done; spin_lock_irqsave(&task->lock, flags); done = __is_done(task); spin_unlock_irqrestore(&task->lock, flags); return done; } /* do_task is a wrapper for the three tasks (requester, * completer, responder) and calls them in a loop until * they return a non-zero value. It is called either * directly by rxe_run_task or indirectly if rxe_sched_task * schedules the task. They must call __reserve_if_idle to * move the task to busy before calling or scheduling. * The task can also be moved to drained or invalid * by calls to rxe_cleanup_task or rxe_disable_task. * In that case tasks which get here are not executed but * just flushed. The tasks are designed to look to see if * there is work to do and then do part of it before returning * here with a return value of zero until all the work * has been consumed then it returns a non-zero value. * The number of times the task can be run is limited by * max iterations so one task cannot hold the cpu forever. * If the limit is hit and work remains the task is rescheduled. */ static void do_task(struct rxe_task *task) { unsigned int iterations; unsigned long flags; int resched = 0; int cont; int ret; WARN_ON(rxe_read(task->qp) <= 0); spin_lock_irqsave(&task->lock, flags); if (task->state >= TASK_STATE_DRAINED) { rxe_put(task->qp); task->num_done++; spin_unlock_irqrestore(&task->lock, flags); return; } spin_unlock_irqrestore(&task->lock, flags); do { iterations = RXE_MAX_ITERATIONS; cont = 0; do { ret = task->func(task->qp); } while (ret == 0 && iterations-- > 0); spin_lock_irqsave(&task->lock, flags); /* we're not done yet but we ran out of iterations. * yield the cpu and reschedule the task */ if (!ret) { task->state = TASK_STATE_IDLE; resched = 1; goto exit; } switch (task->state) { case TASK_STATE_BUSY: task->state = TASK_STATE_IDLE; break; /* someone tried to schedule the task while we * were running, keep going */ case TASK_STATE_ARMED: task->state = TASK_STATE_BUSY; cont = 1; break; case TASK_STATE_DRAINING: task->state = TASK_STATE_DRAINED; break; default: WARN_ON(1); rxe_dbg_qp(task->qp, "unexpected task state = %d", task->state); task->state = TASK_STATE_IDLE; } exit: if (!cont) { task->num_done++; if (WARN_ON(task->num_done != task->num_sched)) rxe_dbg_qp( task->qp, "%ld tasks scheduled, %ld tasks done", task->num_sched, task->num_done); } spin_unlock_irqrestore(&task->lock, flags); } while (cont); task->ret = ret; if (resched) rxe_sched_task(task); rxe_put(task->qp); } /* wrapper around do_task to fix argument for work queue */ static void do_work(struct work_struct *work) { do_task(container_of(work, struct rxe_task, work)); } int rxe_init_task(struct rxe_task *task, struct rxe_qp *qp, int (*func)(struct rxe_qp *)) { WARN_ON(rxe_read(qp) <= 0); task->qp = qp; task->func = func; task->state = TASK_STATE_IDLE; spin_lock_init(&task->lock); INIT_WORK(&task->work, do_work); return 0; } /* rxe_cleanup_task is only called from rxe_do_qp_cleanup in * process context. The qp is already completed with no * remaining references. Once the queue is drained the * task is moved to invalid and returns. The qp cleanup * code then calls the task functions directly without * using the task struct to drain any late arriving packets * or work requests. */ void rxe_cleanup_task(struct rxe_task *task) { unsigned long flags; spin_lock_irqsave(&task->lock, flags); if (!__is_done(task) && task->state < TASK_STATE_DRAINED) { task->state = TASK_STATE_DRAINING; } else { task->state = TASK_STATE_INVALID; spin_unlock_irqrestore(&task->lock, flags); return; } spin_unlock_irqrestore(&task->lock, flags); /* now the task cannot be scheduled or run just wait * for the previously scheduled tasks to finish. */ while (!is_done(task)) cond_resched(); spin_lock_irqsave(&task->lock, flags); task->state = TASK_STATE_INVALID; spin_unlock_irqrestore(&task->lock, flags); } /* run the task inline if it is currently idle * cannot call do_task holding the lock */ void rxe_run_task(struct rxe_task *task) { unsigned long flags; bool run; WARN_ON(rxe_read(task->qp) <= 0); spin_lock_irqsave(&task->lock, flags); run = __reserve_if_idle(task); spin_unlock_irqrestore(&task->lock, flags); if (run) do_task(task); } /* schedule the task to run later as a work queue entry. * the queue_work call can be called holding * the lock. */ void rxe_sched_task(struct rxe_task *task) { unsigned long flags; WARN_ON(rxe_read(task->qp) <= 0); spin_lock_irqsave(&task->lock, flags); if (__reserve_if_idle(task)) queue_work(rxe_wq, &task->work); spin_unlock_irqrestore(&task->lock, flags); } /* rxe_disable/enable_task are only called from * rxe_modify_qp in process context. Task is moved * to the drained state by do_task. */ void rxe_disable_task(struct rxe_task *task) { unsigned long flags; WARN_ON(rxe_read(task->qp) <= 0); spin_lock_irqsave(&task->lock, flags); if (!__is_done(task) && task->state < TASK_STATE_DRAINED) { task->state = TASK_STATE_DRAINING; } else { task->state = TASK_STATE_DRAINED; spin_unlock_irqrestore(&task->lock, flags); return; } spin_unlock_irqrestore(&task->lock, flags); while (!is_done(task)) cond_resched(); spin_lock_irqsave(&task->lock, flags); task->state = TASK_STATE_DRAINED; spin_unlock_irqrestore(&task->lock, flags); } void rxe_enable_task(struct rxe_task *task) { unsigned long flags; WARN_ON(rxe_read(task->qp) <= 0); spin_lock_irqsave(&task->lock, flags); if (task->state == TASK_STATE_INVALID) { spin_unlock_irqrestore(&task->lock, flags); return; } task->state = TASK_STATE_IDLE; spin_unlock_irqrestore(&task->lock, flags); }
Information contained on this website is for historical information purposes only and does not indicate or represent copyright ownership.
Created with Cregit http://github.com/cregit/cregit
Version 2.0-RC1