[PATCH v8 04/16] lpfc: vmid: Add the datastructure for supporting VMID in lpfc
Hannes Reinecke
hare at suse.de
Thu Mar 4 12:29:42 GMT 2021
On 3/4/21 6:27 AM, Muneendra wrote:
> From: Gaurav Srivastava <gaurav.srivastava at broadcom.com>
>
> This patch adds the primary datastructures needed to implement VMID in lpfc
> driver. It maintains the capability, current state, hash table for the
> vmid/appid along with other information.
>
> Signed-off-by: Gaurav Srivastava <gaurav.srivastava at broadcom.com>
> Signed-off-by: James Smart <jsmart2021 at gmail.com>
>
> ---
> v8:
> modify structure member to uniform data type naming scheme
>
> v7:
> No change
>
> v6:
> No change
>
> v5:
> No Change
>
> v4:
> No change
>
> v3:
> No change
>
> v2:
> Ported the patch on top of 5.10/scsi-queue
> Removed unused variable.
> ---
> drivers/scsi/lpfc/lpfc.h | 97 ++++++++++++++++++++++++++++++++++++++++
> 1 file changed, 97 insertions(+)
>
> diff --git a/drivers/scsi/lpfc/lpfc.h b/drivers/scsi/lpfc/lpfc.h
> index 6ba5fa08c47a..dccebe3aae68 100644
> --- a/drivers/scsi/lpfc/lpfc.h
> +++ b/drivers/scsi/lpfc/lpfc.h
> @@ -303,6 +303,63 @@ struct lpfc_stats {
> struct lpfc_hba;
>
>
> +#define LPFC_VMID_TIMER 300 /* timer interval in seconds. */
> +
> +#define LPFC_MAX_VMID_SIZE 256
> +#define LPFC_COMPRESS_VMID_SIZE 16
> +
> +union lpfc_vmid_io_tag {
> + u32 app_id; /* App Id vmid */
> + u8 cs_ctl_vmid; /* Priority tag vmid */
> +};
> +
> +#define JIFFIES_PER_HR (HZ * 60 * 60)
> +
> +struct lpfc_vmid {
> + u8 flag;
> +#define LPFC_VMID_SLOT_FREE 0x0
> +#define LPFC_VMID_SLOT_USED 0x1
> +#define LPFC_VMID_REQ_REGISTER 0x2
> +#define LPFC_VMID_REGISTERED 0x4
> +#define LPFC_VMID_DE_REGISTER 0x8
> + char host_vmid[LPFC_MAX_VMID_SIZE];
> + union lpfc_vmid_io_tag un;
> + u64 io_rd_cnt;
> + u64 io_wr_cnt;
> + u8 vmid_len;
> + u8 delete_inactive; /* Delete if inactive flag 0 = no, 1 = yes */
Please use 'bool' here to indicate it just has 'on' or 'off' values.
> + u32 hash_index;
> + u64 __percpu *last_io_time;
> +};
> +
> +#define lpfc_vmid_is_type_priority_tag(vport)\
> + (vport->vmid_priority_tagging ? 1 : 0)
> +
> +#define LPFC_VMID_HASH_SIZE 256
> +#define LPFC_VMID_HASH_MASK 255
> +#define LPFC_VMID_HASH_SHIFT 6
> +
> +struct lpfc_vmid_context {
> + struct lpfc_vmid *vmp;
> + struct lpfc_nodelist *nlp;
> + u8 instantiated;
Can this also be made a 'bool' type?
> +};
> +
> +struct lpfc_vmid_priority_range {
> + u8 low;
> + u8 high;
> + u8 qos;
> +};
> +
> +struct lpfc_vmid_priority_info {
> + u32 num_descriptors;
> + struct lpfc_vmid_priority_range *vmid_range;
> +};
> +
> +#define QFPA_EVEN_ONLY 0x01
> +#define QFPA_ODD_ONLY 0x02
> +#define QFPA_EVEN_ODD 0x03
> +
> enum discovery_state {
> LPFC_VPORT_UNKNOWN = 0, /* vport state is unknown */
> LPFC_VPORT_FAILED = 1, /* vport has failed */
> @@ -442,6 +499,9 @@ struct lpfc_vport {
> #define WORKER_RAMP_DOWN_QUEUE 0x800 /* hba: Decrease Q depth */
> #define WORKER_RAMP_UP_QUEUE 0x1000 /* hba: Increase Q depth */
> #define WORKER_SERVICE_TXQ 0x2000 /* hba: IOCBs on the txq */
> +#define WORKER_CHECK_INACTIVE_VMID 0x4000 /* hba: check inactive vmids */
> +#define WORKER_CHECK_VMID_ISSUE_QFPA 0x8000 /* vport: Check if qfpa need */
> + /* to issue */
>
> struct timer_list els_tmofunc;
> struct timer_list delayed_disc_tmo;
> @@ -452,6 +512,8 @@ struct lpfc_vport {
> #define FC_LOADING 0x1 /* HBA in process of loading drvr */
> #define FC_UNLOADING 0x2 /* HBA in process of unloading drvr */
> #define FC_ALLOW_FDMI 0x4 /* port is ready for FDMI requests */
> +#define FC_ALLOW_VMID 0x8 /* Allow VMID IO's */
> +#define FC_DEREGISTER_ALL_APP_ID 0x10 /* Deregister all vmid's */
> /* Vport Config Parameters */
> uint32_t cfg_scan_down;
> uint32_t cfg_lun_queue_depth;
> @@ -470,9 +532,36 @@ struct lpfc_vport {
> uint32_t cfg_tgt_queue_depth;
> uint32_t cfg_first_burst_size;
> uint32_t dev_loss_tmo_changed;
> + /* VMID parameters */
> + u8 lpfc_vmid_host_uuid[LPFC_COMPRESS_VMID_SIZE];
> + u32 max_vmid; /* maximum VMIDs allowed per port */
> + u32 cur_vmid_cnt; /* Current VMID count */
> +#define LPFC_MIN_VMID 4
> +#define LPFC_MAX_VMID 255
> + u32 vmid_inactivity_timeout; /* Time after which the VMID */
> + /* deregisters from switch */
> + u32 vmid_priority_tagging;
> +#define LPFC_VMID_PRIO_TAG_DISABLE 0 /* Disable */
> +#define LPFC_VMID_PRIO_TAG_SUP_TARGETS 1 /* Allow supported targets only */
> +#define LPFC_VMID_PRIO_TAG_ALL_TARGETS 2 /* Allow all targets */
> + unsigned long *vmid_priority_range;
> +#define LPFC_VMID_MAX_PRIORITY_RANGE 256
> +#define LPFC_VMID_PRIORITY_BITMAP_SIZE 32
> + u8 vmid_flag;
> +#define LPFC_VMID_IN_USE 0x1
> +#define LPFC_VMID_ISSUE_QFPA 0x2
> +#define LPFC_VMID_QFPA_CMPL 0x4
> +#define LPFC_VMID_QOS_ENABLED 0x8
> +#define LPFC_VMID_TIMER_ENBLD 0x10
> + struct fc_qfpa_res *qfpa_res;
>
> struct fc_vport *fc_vport;
>
> + struct lpfc_vmid *vmid;
> + struct lpfc_vmid *hash_table[LPFC_VMID_HASH_SIZE];
> + rwlock_t vmid_lock;
> + struct lpfc_vmid_priority_info vmid_priority;
> +
> #ifdef CONFIG_SCSI_LPFC_DEBUG_FS
> struct dentry *debug_disc_trc;
> struct dentry *debug_nodelist;
> @@ -937,6 +1026,13 @@ struct lpfc_hba {
> struct nvmet_fc_target_port *targetport;
> lpfc_vpd_t vpd; /* vital product data */
>
> + u32 cfg_max_vmid; /* maximum VMIDs allowed per port */
> + u32 cfg_vmid_app_header;
> +#define LPFC_VMID_APP_HEADER_DISABLE 0
> +#define LPFC_VMID_APP_HEADER_ENABLE 1
> + u32 cfg_vmid_priority_tagging;
> + u32 cfg_vmid_inactivity_timeout; /* Time after which the VMID */
> + /* deregisters from switch */
> struct pci_dev *pcidev;
> struct list_head work_list;
> uint32_t work_ha; /* Host Attention Bits for WT */
> @@ -1177,6 +1273,7 @@ struct lpfc_hba {
> struct list_head ct_ev_waiters;
> struct unsol_rcv_ct_ctx ct_ctx[LPFC_CT_CTX_MAX];
> uint32_t ctx_idx;
> + struct timer_list inactive_vmid_poll;
>
> /* RAS Support */
> struct lpfc_ras_fwlog ras_fwlog;
>
Cheers,
Hannes
--
Dr. Hannes Reinecke Kernel Storage Architect
hare at suse.de +49 911 74053 688
SUSE Software Solutions GmbH, Maxfeldstr. 5, 90409 Nürnberg
HRB 36809 (AG Nürnberg), Geschäftsführer: Felix Imendörffer
More information about the Linux-nvme
mailing list