cregit-Linux how code gets into the kernel

Release 4.12 include/linux/qed/qed_roce_if.h

/* QLogic qed NIC Driver
 * Copyright (c) 2015-2017  QLogic Corporation
 *
 * This software is available to you under a choice of one of two
 * licenses.  You may choose to be licensed under the terms of the GNU
 * General Public License (GPL) Version 2, available from the file
 * COPYING in the main directory of this source tree, or the
 * OpenIB.org BSD license below:
 *
 *     Redistribution and use in source and binary forms, with or
 *     without modification, are permitted provided that the following
 *     conditions are met:
 *
 *      - Redistributions of source code must retain the above
 *        copyright notice, this list of conditions and the following
 *        disclaimer.
 *
 *      - Redistributions in binary form must reproduce the above
 *        copyright notice, this list of conditions and the following
 *        disclaimer in the documentation and /or other materials
 *        provided with the distribution.
 *
 * THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND,
 * EXPRESS OR IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF
 * MERCHANTABILITY, FITNESS FOR A PARTICULAR PURPOSE AND
 * NONINFRINGEMENT. IN NO EVENT SHALL THE AUTHORS OR COPYRIGHT HOLDERS
 * BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER LIABILITY, WHETHER IN AN
 * ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING FROM, OUT OF OR IN
 * CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER DEALINGS IN THE
 * SOFTWARE.
 */
#ifndef _QED_ROCE_IF_H

#define _QED_ROCE_IF_H
#include <linux/types.h>
#include <linux/delay.h>
#include <linux/list.h>
#include <linux/mutex.h>
#include <linux/pci.h>
#include <linux/slab.h>
#include <linux/qed/qed_if.h>
#include <linux/qed/qed_ll2_if.h>
#include <linux/qed/rdma_common.h>


enum qed_roce_ll2_tx_dest {
	/* Light L2 TX Destination to the Network */
	
QED_ROCE_LL2_TX_DEST_NW,

	/* Light L2 TX Destination to the Loopback */
	
QED_ROCE_LL2_TX_DEST_LB,
	
QED_ROCE_LL2_TX_DEST_MAX
};


#define QED_RDMA_MAX_CNQ_SIZE               (0xFFFF)

/* rdma interface */


enum qed_roce_qp_state {
	
QED_ROCE_QP_STATE_RESET,
	
QED_ROCE_QP_STATE_INIT,
	
QED_ROCE_QP_STATE_RTR,
	
QED_ROCE_QP_STATE_RTS,
	
QED_ROCE_QP_STATE_SQD,
	
QED_ROCE_QP_STATE_ERR,
	
QED_ROCE_QP_STATE_SQE
};


enum qed_rdma_tid_type {
	
QED_RDMA_TID_REGISTERED_MR,
	
QED_RDMA_TID_FMR,
	
QED_RDMA_TID_MW_TYPE1,
	
QED_RDMA_TID_MW_TYPE2A
};


struct qed_rdma_events {
	
void *context;
	
void (*affiliated_event)(void *context, u8 fw_event_code,
				 void *fw_handle);
	
void (*unaffiliated_event)(void *context, u8 event_code);
};


struct qed_rdma_device {
	
u32 vendor_id;
	
u32 vendor_part_id;
	
u32 hw_ver;
	
u64 fw_ver;

	
u64 node_guid;
	
u64 sys_image_guid;

	
u8 max_cnq;
	
u8 max_sge;
	
u8 max_srq_sge;
	
u16 max_inline;
	
u32 max_wqe;
	
u32 max_srq_wqe;
	
u8 max_qp_resp_rd_atomic_resc;
	
u8 max_qp_req_rd_atomic_resc;
	
u64 max_dev_resp_rd_atomic_resc;
	
u32 max_cq;
	
u32 max_qp;
	
u32 max_srq;
	
u32 max_mr;
	
u64 max_mr_size;
	
u32 max_cqe;
	
u32 max_mw;
	
u32 max_fmr;
	
u32 max_mr_mw_fmr_pbl;
	
u64 max_mr_mw_fmr_size;
	
u32 max_pd;
	
u32 max_ah;
	
u8 max_pkey;
	
u16 max_srq_wr;
	
u8 max_stats_queues;
	
u32 dev_caps;

	/* Abilty to support RNR-NAK generation */


#define QED_RDMA_DEV_CAP_RNR_NAK_MASK                           0x1

#define QED_RDMA_DEV_CAP_RNR_NAK_SHIFT                  0
	/* Abilty to support shutdown port */

#define QED_RDMA_DEV_CAP_SHUTDOWN_PORT_MASK                     0x1

#define QED_RDMA_DEV_CAP_SHUTDOWN_PORT_SHIFT                    1
	/* Abilty to support port active event */

#define QED_RDMA_DEV_CAP_PORT_ACTIVE_EVENT_MASK         0x1

#define QED_RDMA_DEV_CAP_PORT_ACTIVE_EVENT_SHIFT                2
	/* Abilty to support port change event */

#define QED_RDMA_DEV_CAP_PORT_CHANGE_EVENT_MASK         0x1

#define QED_RDMA_DEV_CAP_PORT_CHANGE_EVENT_SHIFT                3
	/* Abilty to support system image GUID */

#define QED_RDMA_DEV_CAP_SYS_IMAGE_MASK                 0x1

#define QED_RDMA_DEV_CAP_SYS_IMAGE_SHIFT                        4
	/* Abilty to support bad P_Key counter support */

#define QED_RDMA_DEV_CAP_BAD_PKEY_CNT_MASK                      0x1

#define QED_RDMA_DEV_CAP_BAD_PKEY_CNT_SHIFT                     5
	/* Abilty to support atomic operations */

#define QED_RDMA_DEV_CAP_ATOMIC_OP_MASK                 0x1

#define QED_RDMA_DEV_CAP_ATOMIC_OP_SHIFT                        6

#define QED_RDMA_DEV_CAP_RESIZE_CQ_MASK                 0x1

#define QED_RDMA_DEV_CAP_RESIZE_CQ_SHIFT                        7
	/* Abilty to support modifying the maximum number of
         * outstanding work requests per QP
         */

#define QED_RDMA_DEV_CAP_RESIZE_MAX_WR_MASK                     0x1

#define QED_RDMA_DEV_CAP_RESIZE_MAX_WR_SHIFT                    8
	/* Abilty to support automatic path migration */

#define QED_RDMA_DEV_CAP_AUTO_PATH_MIG_MASK                     0x1

#define QED_RDMA_DEV_CAP_AUTO_PATH_MIG_SHIFT                    9
	/* Abilty to support the base memory management extensions */

#define QED_RDMA_DEV_CAP_BASE_MEMORY_EXT_MASK                   0x1

#define QED_RDMA_DEV_CAP_BASE_MEMORY_EXT_SHIFT          10

#define QED_RDMA_DEV_CAP_BASE_QUEUE_EXT_MASK                    0x1

#define QED_RDMA_DEV_CAP_BASE_QUEUE_EXT_SHIFT                   11
	/* Abilty to support multipile page sizes per memory region */

#define QED_RDMA_DEV_CAP_MULTI_PAGE_PER_MR_EXT_MASK             0x1

#define QED_RDMA_DEV_CAP_MULTI_PAGE_PER_MR_EXT_SHIFT            12
	/* Abilty to support block list physical buffer list */

#define QED_RDMA_DEV_CAP_BLOCK_MODE_MASK                        0x1

#define QED_RDMA_DEV_CAP_BLOCK_MODE_SHIFT                       13
	/* Abilty to support zero based virtual addresses */

#define QED_RDMA_DEV_CAP_ZBVA_MASK                              0x1

#define QED_RDMA_DEV_CAP_ZBVA_SHIFT                             14
	/* Abilty to support local invalidate fencing */

#define QED_RDMA_DEV_CAP_LOCAL_INV_FENCE_MASK                   0x1

#define QED_RDMA_DEV_CAP_LOCAL_INV_FENCE_SHIFT          15
	/* Abilty to support Loopback on QP */

#define QED_RDMA_DEV_CAP_LB_INDICATOR_MASK                      0x1

#define QED_RDMA_DEV_CAP_LB_INDICATOR_SHIFT                     16
	
u64 page_size_caps;
	
u8 dev_ack_delay;
	
u32 reserved_lkey;
	
u32 bad_pkey_counter;
	
struct qed_rdma_events events;
};


enum qed_port_state {
	
QED_RDMA_PORT_UP,
	
QED_RDMA_PORT_DOWN,
};


enum qed_roce_capability {
	
QED_ROCE_V1 = 1 << 0,
	
QED_ROCE_V2 = 1 << 1,
};


struct qed_rdma_port {
	
enum qed_port_state port_state;
	
int link_speed;
	
u64 max_msg_size;
	
u8 source_gid_table_len;
	
void *source_gid_table_ptr;
	
u8 pkey_table_len;
	
void *pkey_table_ptr;
	
u32 pkey_bad_counter;
	
enum qed_roce_capability capability;
};


struct qed_rdma_cnq_params {
	
u8 num_pbl_pages;
	
u64 pbl_ptr;
};

/* The CQ Mode affects the CQ doorbell transaction size.
 * 64/32 bit machines should configure to 32/16 bits respectively.
 */

enum qed_rdma_cq_mode {
	
QED_RDMA_CQ_MODE_16_BITS,
	
QED_RDMA_CQ_MODE_32_BITS,
};


struct qed_roce_dcqcn_params {
	
u8 notification_point;
	
u8 reaction_point;

	/* fields for notification point */
	
u32 cnp_send_timeout;

	/* fields for reaction point */
	
u32 rl_bc_rate;
	
u16 rl_max_rate;
	
u16 rl_r_ai;
	
u16 rl_r_hai;
	
u16 dcqcn_g;
	
u32 dcqcn_k_us;
	
u32 dcqcn_timeout_us;
};


struct qed_rdma_start_in_params {
	
struct qed_rdma_events *events;
	
struct qed_rdma_cnq_params cnq_pbl_list[128];
	
u8 desired_cnq;
	
enum qed_rdma_cq_mode cq_mode;
	
struct qed_roce_dcqcn_params dcqcn_params;
	
u16 max_mtu;
	
u8 mac_addr[ETH_ALEN];
	
u8 iwarp_flags;
};


struct qed_rdma_add_user_out_params {
	
u16 dpi;
	
u64 dpi_addr;
	
u64 dpi_phys_addr;
	
u32 dpi_size;
	
u16 wid_count;
};


enum roce_mode {
	
ROCE_V1,
	
ROCE_V2_IPV4,
	
ROCE_V2_IPV6,
	
MAX_ROCE_MODE
};


union qed_gid {
	
u8 bytes[16];
	
u16 words[8];
	
u32 dwords[4];
	
u64 qwords[2];
	
u32 ipv4_addr;
};


struct qed_rdma_register_tid_in_params {
	
u32 itid;
	
enum qed_rdma_tid_type tid_type;
	
u8 key;
	
u16 pd;
	
bool local_read;
	
bool local_write;
	
bool remote_read;
	
bool remote_write;
	
bool remote_atomic;
	
bool mw_bind;
	
u64 pbl_ptr;
	
bool pbl_two_level;
	
u8 pbl_page_size_log;
	
u8 page_size_log;
	
u32 fbo;
	
u64 length;
	
u64 vaddr;
	
bool zbva;
	
bool phy_mr;
	
bool dma_mr;

	
bool dif_enabled;
	
u64 dif_error_addr;
	
u64 dif_runt_addr;
};


struct qed_rdma_create_cq_in_params {
	
u32 cq_handle_lo;
	
u32 cq_handle_hi;
	
u32 cq_size;
	
u16 dpi;
	
bool pbl_two_level;
	
u64 pbl_ptr;
	
u16 pbl_num_pages;
	
u8 pbl_page_size_log;
	
u8 cnq_id;
	
u16 int_timeout;
};


struct qed_rdma_create_srq_in_params {
	
u64 pbl_base_addr;
	
u64 prod_pair_addr;
	
u16 num_pages;
	
u16 pd_id;
	
u16 page_size;
};


struct qed_rdma_destroy_cq_in_params {
	
u16 icid;
};


struct qed_rdma_destroy_cq_out_params {
	
u16 num_cq_notif;
};


struct qed_rdma_create_qp_in_params {
	
u32 qp_handle_lo;
	
u32 qp_handle_hi;
	
u32 qp_handle_async_lo;
	
u32 qp_handle_async_hi;
	
bool use_srq;
	
bool signal_all;
	
bool fmr_and_reserved_lkey;
	
u16 pd;
	
u16 dpi;
	
u16 sq_cq_id;
	
u16 sq_num_pages;
	
u64 sq_pbl_ptr;
	
u8 max_sq_sges;
	
u16 rq_cq_id;
	
u16 rq_num_pages;
	
u64 rq_pbl_ptr;
	
u16 srq_id;
	
u8 stats_queue;
};


struct qed_rdma_create_qp_out_params {
	
u32 qp_id;
	
u16 icid;
	
void *rq_pbl_virt;
	
dma_addr_t rq_pbl_phys;
	
void *sq_pbl_virt;
	
dma_addr_t sq_pbl_phys;
};


struct qed_rdma_modify_qp_in_params {
	
u32 modify_flags;

#define QED_RDMA_MODIFY_QP_VALID_NEW_STATE_MASK               0x1

#define QED_RDMA_MODIFY_QP_VALID_NEW_STATE_SHIFT              0

#define QED_ROCE_MODIFY_QP_VALID_PKEY_MASK                    0x1

#define QED_ROCE_MODIFY_QP_VALID_PKEY_SHIFT                   1

#define QED_RDMA_MODIFY_QP_VALID_RDMA_OPS_EN_MASK             0x1

#define QED_RDMA_MODIFY_QP_VALID_RDMA_OPS_EN_SHIFT            2

#define QED_ROCE_MODIFY_QP_VALID_DEST_QP_MASK                 0x1

#define QED_ROCE_MODIFY_QP_VALID_DEST_QP_SHIFT                3

#define QED_ROCE_MODIFY_QP_VALID_ADDRESS_VECTOR_MASK          0x1

#define QED_ROCE_MODIFY_QP_VALID_ADDRESS_VECTOR_SHIFT         4

#define QED_ROCE_MODIFY_QP_VALID_RQ_PSN_MASK                  0x1

#define QED_ROCE_MODIFY_QP_VALID_RQ_PSN_SHIFT                 5

#define QED_ROCE_MODIFY_QP_VALID_SQ_PSN_MASK                  0x1

#define QED_ROCE_MODIFY_QP_VALID_SQ_PSN_SHIFT                 6

#define QED_RDMA_MODIFY_QP_VALID_MAX_RD_ATOMIC_REQ_MASK       0x1

#define QED_RDMA_MODIFY_QP_VALID_MAX_RD_ATOMIC_REQ_SHIFT      7

#define QED_RDMA_MODIFY_QP_VALID_MAX_RD_ATOMIC_RESP_MASK      0x1

#define QED_RDMA_MODIFY_QP_VALID_MAX_RD_ATOMIC_RESP_SHIFT     8

#define QED_ROCE_MODIFY_QP_VALID_ACK_TIMEOUT_MASK             0x1

#define QED_ROCE_MODIFY_QP_VALID_ACK_TIMEOUT_SHIFT            9

#define QED_ROCE_MODIFY_QP_VALID_RETRY_CNT_MASK               0x1

#define QED_ROCE_MODIFY_QP_VALID_RETRY_CNT_SHIFT              10

#define QED_ROCE_MODIFY_QP_VALID_RNR_RETRY_CNT_MASK           0x1

#define QED_ROCE_MODIFY_QP_VALID_RNR_RETRY_CNT_SHIFT          11

#define QED_ROCE_MODIFY_QP_VALID_MIN_RNR_NAK_TIMER_MASK       0x1

#define QED_ROCE_MODIFY_QP_VALID_MIN_RNR_NAK_TIMER_SHIFT      12

#define QED_ROCE_MODIFY_QP_VALID_E2E_FLOW_CONTROL_EN_MASK     0x1

#define QED_ROCE_MODIFY_QP_VALID_E2E_FLOW_CONTROL_EN_SHIFT    13

#define QED_ROCE_MODIFY_QP_VALID_ROCE_MODE_MASK               0x1

#define QED_ROCE_MODIFY_QP_VALID_ROCE_MODE_SHIFT              14

	
enum qed_roce_qp_state new_state;
	
u16 pkey;
	
bool incoming_rdma_read_en;
	
bool incoming_rdma_write_en;
	
bool incoming_atomic_en;
	
bool e2e_flow_control_en;
	
u32 dest_qp;
	
bool lb_indication;
	
u16 mtu;
	
u8 traffic_class_tos;
	
u8 hop_limit_ttl;
	
u32 flow_label;
	
union qed_gid sgid;
	
union qed_gid dgid;
	
u16 udp_src_port;

	
u16 vlan_id;

	
u32 rq_psn;
	
u32 sq_psn;
	
u8 max_rd_atomic_resp;
	
u8 max_rd_atomic_req;
	
u32 ack_timeout;
	
u8 retry_cnt;
	
u8 rnr_retry_cnt;
	
u8 min_rnr_nak_timer;
	
bool sqd_async;
	
u8 remote_mac_addr[6];
	
u8 local_mac_addr[6];
	
bool use_local_mac;
	
enum roce_mode roce_mode;
};


struct qed_rdma_query_qp_out_params {
	
enum qed_roce_qp_state state;
	
u32 rq_psn;
	
u32 sq_psn;
	
bool draining;
	
u16 mtu;
	
u32 dest_qp;
	
bool incoming_rdma_read_en;
	
bool incoming_rdma_write_en;
	
bool incoming_atomic_en;
	
bool e2e_flow_control_en;
	
union qed_gid sgid;
	
union qed_gid dgid;
	
u32 flow_label;
	
u8 hop_limit_ttl;
	
u8 traffic_class_tos;
	
u32 timeout;
	
u8 rnr_retry;
	
u8 retry_cnt;
	
u8 min_rnr_nak_timer;
	
u16 pkey_index;
	
u8 max_rd_atomic;
	
u8 max_dest_rd_atomic;
	
bool sqd_async;
};


struct qed_rdma_create_srq_out_params {
	
u16 srq_id;
};


struct qed_rdma_destroy_srq_in_params {
	
u16 srq_id;
};


struct qed_rdma_modify_srq_in_params {
	
u32 wqe_limit;
	
u16 srq_id;
};


struct qed_rdma_stats_out_params {
	
u64 sent_bytes;
	
u64 sent_pkts;
	
u64 rcv_bytes;
	
u64 rcv_pkts;
};


struct qed_rdma_counters_out_params {
	
u64 pd_count;
	
u64 max_pd;
	
u64 dpi_count;
	
u64 max_dpi;
	
u64 cq_count;
	
u64 max_cq;
	
u64 qp_count;
	
u64 max_qp;
	
u64 tid_count;
	
u64 max_tid;
};


#define QED_ROCE_TX_HEAD_FAILURE        (1)

#define QED_ROCE_TX_FRAG_FAILURE        (2)


struct qed_roce_ll2_header {
	
void *vaddr;
	
dma_addr_t baddr;
	
size_t len;
};


struct qed_roce_ll2_buffer {
	
dma_addr_t baddr;
	
size_t len;
};


struct qed_roce_ll2_packet {
	
struct qed_roce_ll2_header header;
	
int n_seg;
	
struct qed_roce_ll2_buffer payload[RDMA_MAX_SGE_PER_SQ_WQE];
	
int roce_mode;
	
enum qed_roce_ll2_tx_dest tx_dest;
};


struct qed_roce_ll2_tx_params {
	
int reserved;
};


struct qed_roce_ll2_rx_params {
	
u16 vlan_id;
	
u8 smac[ETH_ALEN];
	
int rc;
};


struct qed_roce_ll2_cbs {
	
void (*tx_cb)(void *pdev, struct qed_roce_ll2_packet *pkt);

	
void (*rx_cb)(void *pdev, struct qed_roce_ll2_packet *pkt,
		      struct qed_roce_ll2_rx_params *params);
};


struct qed_roce_ll2_params {
	
u16 max_rx_buffers;
	
u16 max_tx_buffers;
	
u16 mtu;
	
u8 mac_address[ETH_ALEN];
	
struct qed_roce_ll2_cbs cbs;
	
void *cb_cookie;
};


struct qed_roce_ll2_info {
	
u8 handle;
	
struct qed_roce_ll2_cbs cbs;
	
u8 mac_address[ETH_ALEN];
	
void *cb_cookie;

	/* Lock to protect ll2 */
	
struct mutex lock;
};


enum qed_rdma_type {
	
QED_RDMA_TYPE_ROCE,
};


struct qed_dev_rdma_info {
	
struct qed_dev_info common;
	
enum qed_rdma_type rdma_type;
	
u8 user_dpm_enabled;
};


struct qed_rdma_ops {
	
const struct qed_common_ops *common;

	
int (*fill_dev_info)(struct qed_dev *cdev,
			     struct qed_dev_rdma_info *info);
	
void *(*rdma_get_rdma_ctx)(struct qed_dev *cdev);

	
int (*rdma_init)(struct qed_dev *dev,
			 struct qed_rdma_start_in_params *iparams);

	
int (*rdma_add_user)(void *rdma_cxt,
			     struct qed_rdma_add_user_out_params *oparams);

	
void (*rdma_remove_user)(void *rdma_cxt, u16 dpi);
	
int (*rdma_stop)(void *rdma_cxt);
	
struct qed_rdma_device* (*rdma_query_device)(void *rdma_cxt);
	
struct qed_rdma_port* (*rdma_query_port)(void *rdma_cxt);
	
int (*rdma_get_start_sb)(struct qed_dev *cdev);
	
int (*rdma_get_min_cnq_msix)(struct qed_dev *cdev);
	
void (*rdma_cnq_prod_update)(void *rdma_cxt, u8 cnq_index, u16 prod);
	
int (*rdma_get_rdma_int)(struct qed_dev *cdev,
				 struct qed_int_info *info);
	
int (*rdma_set_rdma_int)(struct qed_dev *cdev, u16 cnt);
	
int (*rdma_alloc_pd)(void *rdma_cxt, u16 *pd);
	
void (*rdma_dealloc_pd)(void *rdma_cxt, u16 pd);
	
int (*rdma_create_cq)(void *rdma_cxt,
			      struct qed_rdma_create_cq_in_params *params,
			      u16 *icid);
	
int (*rdma_destroy_cq)(void *rdma_cxt,
			       struct qed_rdma_destroy_cq_in_params *iparams,
			       struct qed_rdma_destroy_cq_out_params *oparams);
	struct qed_rdma_qp *
	
(*rdma_create_qp)(void *rdma_cxt,
			  struct qed_rdma_create_qp_in_params *iparams,
			  struct qed_rdma_create_qp_out_params *oparams);

	
int (*rdma_modify_qp)(void *roce_cxt, struct qed_rdma_qp *qp,
			      struct qed_rdma_modify_qp_in_params *iparams);

	
int (*rdma_query_qp)(void *rdma_cxt, struct qed_rdma_qp *qp,
			     struct qed_rdma_query_qp_out_params *oparams);
	
int (*rdma_destroy_qp)(void *rdma_cxt, struct qed_rdma_qp *qp);
	int
	
(*rdma_register_tid)(void *rdma_cxt,
			     struct qed_rdma_register_tid_in_params *iparams);
	
int (*rdma_deregister_tid)(void *rdma_cxt, u32 itid);
	
int (*rdma_alloc_tid)(void *rdma_cxt, u32 *itid);
	
void (*rdma_free_tid)(void *rdma_cxt, u32 itid);
	
int (*roce_ll2_start)(struct qed_dev *cdev,
			      struct qed_roce_ll2_params *params);
	
int (*roce_ll2_stop)(struct qed_dev *cdev);
	
int (*roce_ll2_tx)(struct qed_dev *cdev,
			   struct qed_roce_ll2_packet *packet,
			   struct qed_roce_ll2_tx_params *params);
	
int (*roce_ll2_post_rx_buffer)(struct qed_dev *cdev,
				       struct qed_roce_ll2_buffer *buf,
				       u64 cookie, u8 notify_fw);
	
int (*roce_ll2_set_mac_filter)(struct qed_dev *cdev,
				       u8 *old_mac_address,
				       u8 *new_mac_address);
	
int (*roce_ll2_stats)(struct qed_dev *cdev,
			      struct qed_ll2_stats *stats);
};

const struct qed_rdma_ops *qed_get_rdma_ops(void);

#endif

Overall Contributors

PersonTokensPropCommitsCommitProp
Ram Amrani195299.95%685.71%
Yuval Mintz10.05%114.29%
Total1953100.00%7100.00%
Information contained on this website is for historical information purposes only and does not indicate or represent copyright ownership.
Created with cregit.