cregit-Linux how code gets into the kernel

Release 4.7 drivers/staging/lustre/lnet/selftest/selftest.h

/*
 * GPL HEADER START
 *
 * DO NOT ALTER OR REMOVE COPYRIGHT NOTICES OR THIS FILE HEADER.
 *
 * This program is free software; you can redistribute it and/or modify
 * it under the terms of the GNU General Public License version 2 only,
 * as published by the Free Software Foundation.
 *
 * This program is distributed in the hope that it will be useful, but
 * WITHOUT ANY WARRANTY; without even the implied warranty of
 * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE.  See the GNU
 * General Public License version 2 for more details (a copy is included
 * in the LICENSE file that accompanied this code).
 *
 * You should have received a copy of the GNU General Public License
 * version 2 along with this program; If not, see
 * http://www.sun.com/software/products/lustre/docs/GPLv2.pdf
 * copy of GPLv2].
 *
 * Please contact Sun Microsystems, Inc., 4150 Network Circle, Santa Clara,
 * CA 95054 USA or visit www.sun.com if you need additional information or
 * have any questions.
 *
 * GPL HEADER END
 */
/*
 * Copyright (c) 2007, 2010, Oracle and/or its affiliates. All rights reserved.
 * Use is subject to license terms.
 *
 * Copyright (c) 2012, Intel Corporation.
 */
/*
 * This file is part of Lustre, http://www.lustre.org/
 * Lustre is a trademark of Sun Microsystems, Inc.
 *
 * lnet/selftest/selftest.h
 *
 * Author: Isaac Huang <isaac@clusterfs.com>
 */
#ifndef __SELFTEST_SELFTEST_H__

#define __SELFTEST_SELFTEST_H__


#define LNET_ONLY

#include "../../include/linux/libcfs/libcfs.h"
#include "../../include/linux/lnet/lnet.h"
#include "../../include/linux/lnet/lib-lnet.h"
#include "../../include/linux/lnet/lib-types.h"
#include "../../include/linux/lnet/lnetst.h"

#include "rpc.h"
#include "timer.h"

#ifndef MADE_WITHOUT_COMPROMISE

#define MADE_WITHOUT_COMPROMISE
#endif


#define SWI_STATE_NEWBORN		0

#define SWI_STATE_REPLY_SUBMITTED	1

#define SWI_STATE_REPLY_SENT		2

#define SWI_STATE_REQUEST_SUBMITTED	3

#define SWI_STATE_REQUEST_SENT		4

#define SWI_STATE_REPLY_RECEIVED	5

#define SWI_STATE_BULK_STARTED		6

#define SWI_STATE_DONE			10

/* forward refs */
struct srpc_service;
struct srpc_service_cd;
struct sfw_test_unit;
struct sfw_test_instance;

/* services below SRPC_FRAMEWORK_SERVICE_MAX_ID are framework
 * services, e.g. create/modify session.
 */

#define SRPC_SERVICE_DEBUG		0

#define SRPC_SERVICE_MAKE_SESSION	1

#define SRPC_SERVICE_REMOVE_SESSION	2

#define SRPC_SERVICE_BATCH		3

#define SRPC_SERVICE_TEST		4

#define SRPC_SERVICE_QUERY_STAT		5

#define SRPC_SERVICE_JOIN		6

#define SRPC_FRAMEWORK_SERVICE_MAX_ID	10
/* other services start from SRPC_FRAMEWORK_SERVICE_MAX_ID+1 */

#define SRPC_SERVICE_BRW		11

#define SRPC_SERVICE_PING		12

#define SRPC_SERVICE_MAX_ID		12


#define SRPC_REQUEST_PORTAL		50
/* a lazy portal for framework RPC requests */

#define SRPC_FRAMEWORK_REQUEST_PORTAL	51
/* all reply/bulk RDMAs go to this portal */

#define SRPC_RDMA_PORTAL		52


static inline enum srpc_msg_type srpc_service2request(int service) { switch (service) { default: LBUG(); case SRPC_SERVICE_DEBUG: return SRPC_MSG_DEBUG_REQST; case SRPC_SERVICE_MAKE_SESSION: return SRPC_MSG_MKSN_REQST; case SRPC_SERVICE_REMOVE_SESSION: return SRPC_MSG_RMSN_REQST; case SRPC_SERVICE_BATCH: return SRPC_MSG_BATCH_REQST; case SRPC_SERVICE_TEST: return SRPC_MSG_TEST_REQST; case SRPC_SERVICE_QUERY_STAT: return SRPC_MSG_STAT_REQST; case SRPC_SERVICE_BRW: return SRPC_MSG_BRW_REQST; case SRPC_SERVICE_PING: return SRPC_MSG_PING_REQST; case SRPC_SERVICE_JOIN: return SRPC_MSG_JOIN_REQST; } }

Contributors

PersonTokensPropCommitsCommitProp
peng taopeng tao7397.33%150.00%
james simmonsjames simmons22.67%150.00%
Total75100.00%2100.00%


static inline enum srpc_msg_type srpc_service2reply(int service) { return srpc_service2request(service) + 1; }

Contributors

PersonTokensPropCommitsCommitProp
peng taopeng tao1789.47%150.00%
james simmonsjames simmons210.53%150.00%
Total19100.00%2100.00%

enum srpc_event_type { SRPC_BULK_REQ_RCVD = 1, /* passive bulk request(PUT sink/GET source) * received */ SRPC_BULK_PUT_SENT = 2, /* active bulk PUT sent (source) */ SRPC_BULK_GET_RPLD = 3, /* active bulk GET replied (sink) */ SRPC_REPLY_RCVD = 4, /* incoming reply received */ SRPC_REPLY_SENT = 5, /* outgoing reply sent */ SRPC_REQUEST_RCVD = 6, /* incoming request received */ SRPC_REQUEST_SENT = 7, /* outgoing request sent */ }; /* RPC event */ struct srpc_event { enum srpc_event_type ev_type; /* what's up */ lnet_event_kind_t ev_lnet; /* LNet event type */ int ev_fired; /* LNet event fired? */ int ev_status; /* LNet event status */ void *ev_data; /* owning server/client RPC */ }; /* bulk descriptor */ struct srpc_bulk { int bk_len; /* len of bulk data */ lnet_handle_md_t bk_mdh; int bk_sink; /* sink/source */ int bk_niov; /* # iov in bk_iovs */ lnet_kiov_t bk_iovs[0]; }; /* message buffer descriptor */ struct srpc_buffer { struct list_head buf_list; /* chain on srpc_service::*_msgq */ struct srpc_msg buf_msg; lnet_handle_md_t buf_mdh; lnet_nid_t buf_self; lnet_process_id_t buf_peer; }; struct swi_workitem; typedef int (*swi_action_t) (struct swi_workitem *); struct swi_workitem { struct cfs_wi_sched *swi_sched; struct cfs_workitem swi_workitem; swi_action_t swi_action; int swi_state; }; /* server-side state of a RPC */ struct srpc_server_rpc { /* chain on srpc_service::*_rpcq */ struct list_head srpc_list; struct srpc_service_cd *srpc_scd; struct swi_workitem srpc_wi; struct srpc_event srpc_ev; /* bulk/reply event */ lnet_nid_t srpc_self; lnet_process_id_t srpc_peer; struct srpc_msg srpc_replymsg; lnet_handle_md_t srpc_replymdh; struct srpc_buffer *srpc_reqstbuf; struct srpc_bulk *srpc_bulk; unsigned int srpc_aborted; /* being given up */ int srpc_status; void (*srpc_done)(struct srpc_server_rpc *); }; /* client-side state of a RPC */ struct srpc_client_rpc { struct list_head crpc_list; /* chain on user's lists */ spinlock_t crpc_lock; /* serialize */ int crpc_service; atomic_t crpc_refcount; int crpc_timeout; /* # seconds to wait for reply */ struct stt_timer crpc_timer; struct swi_workitem crpc_wi; lnet_process_id_t crpc_dest; void (*crpc_done)(struct srpc_client_rpc *); void (*crpc_fini)(struct srpc_client_rpc *); int crpc_status; /* completion status */ void *crpc_priv; /* caller data */ /* state flags */ unsigned int crpc_aborted:1; /* being given up */ unsigned int crpc_closed:1; /* completed */ /* RPC events */ struct srpc_event crpc_bulkev; /* bulk event */ struct srpc_event crpc_reqstev; /* request event */ struct srpc_event crpc_replyev; /* reply event */ /* bulk, request(reqst), and reply exchanged on wire */ struct srpc_msg crpc_reqstmsg; struct srpc_msg crpc_replymsg; lnet_handle_md_t crpc_reqstmdh; lnet_handle_md_t crpc_replymdh; struct srpc_bulk crpc_bulk; }; #define srpc_client_rpc_size(rpc) \ offsetof(struct srpc_client_rpc, crpc_bulk.bk_iovs[(rpc)->crpc_bulk.bk_niov]) #define srpc_client_rpc_addref(rpc) \ do { \ CDEBUG(D_NET, "RPC[%p] -> %s (%d)++\n", \ (rpc), libcfs_id2str((rpc)->crpc_dest), \ atomic_read(&(rpc)->crpc_refcount)); \ LASSERT(atomic_read(&(rpc)->crpc_refcount) > 0); \ atomic_inc(&(rpc)->crpc_refcount); \ } while (0) #define srpc_client_rpc_decref(rpc) \ do { \ CDEBUG(D_NET, "RPC[%p] -> %s (%d)--\n", \ (rpc), libcfs_id2str((rpc)->crpc_dest), \ atomic_read(&(rpc)->crpc_refcount)); \ LASSERT(atomic_read(&(rpc)->crpc_refcount) > 0); \ if (atomic_dec_and_test(&(rpc)->crpc_refcount)) \ srpc_destroy_client_rpc(rpc); \ } while (0) #define srpc_event_pending(rpc) (!(rpc)->crpc_bulkev.ev_fired || \ !(rpc)->crpc_reqstev.ev_fired || \ !(rpc)->crpc_replyev.ev_fired) /* CPU partition data of srpc service */ struct srpc_service_cd { /** serialize */ spinlock_t scd_lock; /** backref to service */ struct srpc_service *scd_svc; /** event buffer */ struct srpc_event scd_ev; /** free RPC descriptors */ struct list_head scd_rpc_free; /** in-flight RPCs */ struct list_head scd_rpc_active; /** workitem for posting buffer */ struct swi_workitem scd_buf_wi; /** CPT id */ int scd_cpt; /** error code for scd_buf_wi */ int scd_buf_err; /** timestamp for scd_buf_err */ time64_t scd_buf_err_stamp; /** total # request buffers */ int scd_buf_total; /** # posted request buffers */ int scd_buf_nposted; /** in progress of buffer posting */ int scd_buf_posting; /** allocate more buffers if scd_buf_nposted < scd_buf_low */ int scd_buf_low; /** increase/decrease some buffers */ int scd_buf_adjust; /** posted message buffers */ struct list_head scd_buf_posted; /** blocked for RPC descriptor */ struct list_head scd_buf_blocked; }; /* number of server workitems (mini-thread) for testing service */ #define SFW_TEST_WI_MIN 256 #define SFW_TEST_WI_MAX 2048 /* extra buffers for tolerating buggy peers, or unbalanced number * of peers between partitions */ #define SFW_TEST_WI_EXTRA 64 /* number of server workitems (mini-thread) for framework service */ #define SFW_FRWK_WI_MIN 16 #define SFW_FRWK_WI_MAX 256 struct srpc_service { int sv_id; /* service id */ const char *sv_name; /* human readable name */ int sv_wi_total; /* total server workitems */ int sv_shuttingdown; int sv_ncpts; /* percpt data for srpc_service */ struct srpc_service_cd **sv_cpt_data; /* Service callbacks: * - sv_handler: process incoming RPC request * - sv_bulk_ready: notify bulk data */ int (*sv_handler)(struct srpc_server_rpc *); int (*sv_bulk_ready)(struct srpc_server_rpc *, int); }; struct sfw_session { struct list_head sn_list; /* chain on fw_zombie_sessions */ lst_sid_t sn_id; /* unique identifier */ unsigned int sn_timeout; /* # seconds' inactivity to expire */ int sn_timer_active; unsigned int sn_features; struct stt_timer sn_timer; struct list_head sn_batches; /* list of batches */ char sn_name[LST_NAME_SIZE]; atomic_t sn_refcount; atomic_t sn_brw_errors; atomic_t sn_ping_errors; unsigned long sn_started; }; #define sfw_sid_equal(sid0, sid1) ((sid0).ses_nid == (sid1).ses_nid && \ (sid0).ses_stamp == (sid1).ses_stamp) struct sfw_batch { struct list_head bat_list; /* chain on sn_batches */ lst_bid_t bat_id; /* batch id */ int bat_error; /* error code of batch */ struct sfw_session *bat_session; /* batch's session */ atomic_t bat_nactive; /* # of active tests */ struct list_head bat_tests; /* test instances */ }; struct sfw_test_client_ops { int (*tso_init)(struct sfw_test_instance *tsi); /* initialize test * client */ void (*tso_fini)(struct sfw_test_instance *tsi); /* finalize test * client */ int (*tso_prep_rpc)(struct sfw_test_unit *tsu, lnet_process_id_t dest, struct srpc_client_rpc **rpc); /* prep a tests rpc */ void (*tso_done_rpc)(struct sfw_test_unit *tsu, struct srpc_client_rpc *rpc); /* done a test rpc */ }; struct sfw_test_instance { struct list_head tsi_list; /* chain on batch */ int tsi_service; /* test type */ struct sfw_batch *tsi_batch; /* batch */ struct sfw_test_client_ops *tsi_ops; /* test client operation */ /* public parameter for all test units */ unsigned int tsi_is_client:1; /* is test client */ unsigned int tsi_stoptsu_onerr:1; /* stop tsu on error */ int tsi_concur; /* concurrency */ int tsi_loop; /* loop count */ /* status of test instance */ spinlock_t tsi_lock; /* serialize */ unsigned int tsi_stopping:1; /* test is stopping */ atomic_t tsi_nactive; /* # of active test * unit */ struct list_head tsi_units; /* test units */ struct list_head tsi_free_rpcs; /* free rpcs */ struct list_head tsi_active_rpcs; /* active rpcs */ union { struct test_ping_req ping; /* ping parameter */ struct test_bulk_req bulk_v0; /* bulk parameter */ struct test_bulk_req_v1 bulk_v1; /* bulk v1 parameter */ } tsi_u; }; /* XXX: trailing (PAGE_SIZE % sizeof(lnet_process_id_t)) bytes at the end of * pages are not used */ #define SFW_MAX_CONCUR LST_MAX_CONCUR #define SFW_ID_PER_PAGE (PAGE_SIZE / sizeof(lnet_process_id_packed_t)) #define SFW_MAX_NDESTS (LNET_MAX_IOV * SFW_ID_PER_PAGE) #define sfw_id_pages(n) (((n) + SFW_ID_PER_PAGE - 1) / SFW_ID_PER_PAGE) struct sfw_test_unit { struct list_head tsu_list; /* chain on lst_test_instance */ lnet_process_id_t tsu_dest; /* id of dest node */ int tsu_loop; /* loop count of the test */ struct sfw_test_instance *tsu_instance; /* pointer to test instance */ void *tsu_private; /* private data */ struct swi_workitem tsu_worker; /* workitem of the test unit */ }; struct sfw_test_case { struct list_head tsc_list; /* chain on fw_tests */ struct srpc_service *tsc_srv_service; /* test service */ struct sfw_test_client_ops *tsc_cli_ops; /* ops of test client */ }; struct srpc_client_rpc * sfw_create_rpc(lnet_process_id_t peer, int service, unsigned features, int nbulkiov, int bulklen, void (*done)(struct srpc_client_rpc *), void *priv); int sfw_create_test_rpc(struct sfw_test_unit *tsu, lnet_process_id_t peer, unsigned features, int nblk, int blklen, struct srpc_client_rpc **rpc); void sfw_abort_rpc(struct srpc_client_rpc *rpc); void sfw_post_rpc(struct srpc_client_rpc *rpc); void sfw_client_rpc_done(struct srpc_client_rpc *rpc); void sfw_unpack_message(struct srpc_msg *msg); void sfw_free_pages(struct srpc_server_rpc *rpc); void sfw_add_bulk_page(struct srpc_bulk *bk, struct page *pg, int i); int sfw_alloc_pages(struct srpc_server_rpc *rpc, int cpt, int npages, int len, int sink); int sfw_make_session(struct srpc_mksn_reqst *request, struct srpc_mksn_reply *reply); struct srpc_client_rpc * srpc_create_client_rpc(lnet_process_id_t peer, int service, int nbulkiov, int bulklen, void (*rpc_done)(struct srpc_client_rpc *), void (*rpc_fini)(struct srpc_client_rpc *), void *priv); void srpc_post_rpc(struct srpc_client_rpc *rpc); void srpc_abort_rpc(struct srpc_client_rpc *rpc, int why); void srpc_free_bulk(struct srpc_bulk *bk); struct srpc_bulk *srpc_alloc_bulk(int cpt, unsigned bulk_npg, unsigned bulk_len, int sink); int srpc_send_rpc(struct swi_workitem *wi); int srpc_send_reply(struct srpc_server_rpc *rpc); int srpc_add_service(struct srpc_service *sv); int srpc_remove_service(struct srpc_service *sv); void srpc_shutdown_service(struct srpc_service *sv); void srpc_abort_service(struct srpc_service *sv); int srpc_finish_service(struct srpc_service *sv); int srpc_service_add_buffers(struct srpc_service *sv, int nbuffer); void srpc_service_remove_buffers(struct srpc_service *sv, int nbuffer); void srpc_get_counters(srpc_counters_t *cnt); void srpc_set_counters(const srpc_counters_t *cnt); extern struct cfs_wi_sched *lst_sched_serial; extern struct cfs_wi_sched **lst_sched_test;
static inline int srpc_serv_is_framework(struct srpc_service *svc) { return svc->sv_id < SRPC_FRAMEWORK_SERVICE_MAX_ID; }

Contributors

PersonTokensPropCommitsCommitProp
peng taopeng tao19100.00%1100.00%
Total19100.00%1100.00%


static inline int swi_wi_action(struct cfs_workitem *wi) { struct swi_workitem *swi; swi = container_of(wi, struct swi_workitem, swi_workitem); return swi->swi_action(swi); }

Contributors

PersonTokensPropCommitsCommitProp
peng taopeng tao2772.97%133.33%
james simmonsjames simmons1027.03%266.67%
Total37100.00%3100.00%


static inline void swi_init_workitem(struct swi_workitem *swi, void *data, swi_action_t action, struct cfs_wi_sched *sched) { swi->swi_sched = sched; swi->swi_action = action; swi->swi_state = SWI_STATE_NEWBORN; cfs_wi_init(&swi->swi_workitem, data, swi_wi_action); }

Contributors

PersonTokensPropCommitsCommitProp
peng taopeng tao5296.30%150.00%
james simmonsjames simmons23.70%150.00%
Total54100.00%2100.00%


static inline void swi_schedule_workitem(struct swi_workitem *wi) { cfs_wi_schedule(wi->swi_sched, &wi->swi_workitem); }

Contributors

PersonTokensPropCommitsCommitProp
peng taopeng tao2291.67%150.00%
james simmonsjames simmons28.33%150.00%
Total24100.00%2100.00%


static inline void swi_exit_workitem(struct swi_workitem *swi) { cfs_wi_exit(swi->swi_sched, &swi->swi_workitem); }

Contributors

PersonTokensPropCommitsCommitProp
peng taopeng tao2291.67%150.00%
james simmonsjames simmons28.33%150.00%
Total24100.00%2100.00%


static inline int swi_deschedule_workitem(struct swi_workitem *swi) { return cfs_wi_deschedule(swi->swi_sched, &swi->swi_workitem); }

Contributors

PersonTokensPropCommitsCommitProp
peng taopeng tao2392.00%150.00%
james simmonsjames simmons28.00%150.00%
Total25100.00%2100.00%

int sfw_startup(void); int srpc_startup(void); void sfw_shutdown(void); void srpc_shutdown(void);
static inline void srpc_destroy_client_rpc(struct srpc_client_rpc *rpc) { LASSERT(rpc); LASSERT(!srpc_event_pending(rpc)); LASSERT(!atomic_read(&rpc->crpc_refcount)); if (!rpc->crpc_fini) LIBCFS_FREE(rpc, srpc_client_rpc_size(rpc)); else (*rpc->crpc_fini)(rpc); }

Contributors

PersonTokensPropCommitsCommitProp
peng taopeng tao6293.94%125.00%
james simmonsjames simmons46.06%375.00%
Total66100.00%4100.00%


static inline void srpc_init_client_rpc(struct srpc_client_rpc *rpc, lnet_process_id_t peer, int service, int nbulkiov, int bulklen, void (*rpc_done)(struct srpc_client_rpc *), void (*rpc_fini)(struct srpc_client_rpc *), void *priv) { LASSERT(nbulkiov <= LNET_MAX_IOV); memset(rpc, 0, offsetof(struct srpc_client_rpc, crpc_bulk.bk_iovs[nbulkiov])); INIT_LIST_HEAD(&rpc->crpc_list); swi_init_workitem(&rpc->crpc_wi, rpc, srpc_send_rpc, lst_sched_test[lnet_cpt_of_nid(peer.nid)]); spin_lock_init(&rpc->crpc_lock); atomic_set(&rpc->crpc_refcount, 1); /* 1 ref for caller */ rpc->crpc_dest = peer; rpc->crpc_priv = priv; rpc->crpc_service = service; rpc->crpc_bulk.bk_len = bulklen; rpc->crpc_bulk.bk_niov = nbulkiov; rpc->crpc_done = rpc_done; rpc->crpc_fini = rpc_fini; LNetInvalidateHandle(&rpc->crpc_reqstmdh); LNetInvalidateHandle(&rpc->crpc_replymdh); LNetInvalidateHandle(&rpc->crpc_bulk.bk_mdh); /* no event is expected at this point */ rpc->crpc_bulkev.ev_fired = 1; rpc->crpc_reqstev.ev_fired = 1; rpc->crpc_replyev.ev_fired = 1; rpc->crpc_reqstmsg.msg_magic = SRPC_MSG_MAGIC; rpc->crpc_reqstmsg.msg_version = SRPC_MSG_VERSION; rpc->crpc_reqstmsg.msg_type = srpc_service2request(service); }

Contributors

PersonTokensPropCommitsCommitProp
peng taopeng tao23895.20%133.33%
james simmonsjames simmons124.80%266.67%
Total250100.00%3100.00%


static inline const char * swi_state2str(int state) { #define STATE2STR(x) case x: return #x switch (state) { default: LBUG(); STATE2STR(SWI_STATE_NEWBORN); STATE2STR(SWI_STATE_REPLY_SUBMITTED); STATE2STR(SWI_STATE_REPLY_SENT); STATE2STR(SWI_STATE_REQUEST_SUBMITTED); STATE2STR(SWI_STATE_REQUEST_SENT); STATE2STR(SWI_STATE_REPLY_RECEIVED); STATE2STR(SWI_STATE_BULK_STARTED); STATE2STR(SWI_STATE_DONE); } #undef STATE2STR }

Contributors

PersonTokensPropCommitsCommitProp
peng taopeng tao72100.00%1100.00%
Total72100.00%1100.00%

#define selftest_wait_events() \ do { \ set_current_state(TASK_UNINTERRUPTIBLE); \ schedule_timeout(cfs_time_seconds(1) / 10); \ } while (0) #define lst_wait_until(cond, lock, fmt, ...) \ do { \ int __I = 2; \ while (!(cond)) { \ CDEBUG(is_power_of_2(++__I) ? D_WARNING : D_NET, \ fmt, ## __VA_ARGS__); \ spin_unlock(&(lock)); \ \ selftest_wait_events(); \ \ spin_lock(&(lock)); \ } \ } while (0)
static inline void srpc_wait_service_shutdown(struct srpc_service *sv) { int i = 2; LASSERT(sv->sv_shuttingdown); while (!srpc_finish_service(sv)) { i++; CDEBUG(((i & -i) == i) ? D_WARNING : D_NET, "Waiting for %s service to shutdown...\n", sv->sv_name); selftest_wait_events(); } }

Contributors

PersonTokensPropCommitsCommitProp
peng taopeng tao6195.31%133.33%
james simmonsjames simmons34.69%266.67%
Total64100.00%3100.00%

extern struct sfw_test_client_ops brw_test_client; void brw_init_test_client(void); extern struct srpc_service brw_test_service; void brw_init_test_service(void); extern struct sfw_test_client_ops ping_test_client; void ping_init_test_client(void); extern struct srpc_service ping_test_service; void ping_init_test_service(void); #endif /* __SELFTEST_SELFTEST_H__ */

Overall Contributors

PersonTokensPropCommitsCommitProp
peng taopeng tao195488.62%26.25%
james simmonsjames simmons1848.34%2062.50%
jeremiah mahlerjeremiah mahler361.63%13.12%
shraddha barkeshraddha barke110.50%13.12%
greg kroah-hartmangreg kroah-hartman70.32%26.25%
mike shueymike shuey50.23%13.12%
bhumika goyalbhumika goyal40.18%13.12%
kirill a. shutemovkirill a. shutemov20.09%26.25%
arnd bergmannarnd bergmann10.05%13.12%
aya mahfouzaya mahfouz10.05%13.12%
Total2205100.00%32100.00%
Information contained on this website is for historical information purposes only and does not indicate or represent copyright ownership.
{% endraw %}