Message ID | 20210423233455.27243-12-jsmart2021@gmail.com |
---|---|
State | Superseded |
Headers | show |
Series | efct: Broadcom (Emulex) FC Target driver | expand |
On 4/24/21 1:34 AM, James Smart wrote: > This patch continues the libefc library population. > > This patch adds library interface definitions for: > - SLI and FC port (aka n_port_id) registration, allocation and > deallocation. > > Co-developed-by: Ram Vegesna <ram.vegesna@broadcom.com> > Signed-off-by: Ram Vegesna <ram.vegesna@broadcom.com> > Signed-off-by: James Smart <jsmart2021@gmail.com> > Reviewed-by: Daniel Wagner <dwagner@suse.de> > --- > drivers/scsi/elx/libefc/efc_nport.c | 792 ++++++++++++++++++++++++++++ > drivers/scsi/elx/libefc/efc_nport.h | 50 ++ > 2 files changed, 842 insertions(+) > create mode 100644 drivers/scsi/elx/libefc/efc_nport.c > create mode 100644 drivers/scsi/elx/libefc/efc_nport.h > > diff --git a/drivers/scsi/elx/libefc/efc_nport.c b/drivers/scsi/elx/libefc/efc_nport.c > new file mode 100644 > index 000000000000..bd50f0320631 > --- /dev/null > +++ b/drivers/scsi/elx/libefc/efc_nport.c > @@ -0,0 +1,792 @@ > +// SPDX-License-Identifier: GPL-2.0 > +/* > + * Copyright (C) 2021 Broadcom. All Rights Reserved. The term > + * “Broadcom” refers to Broadcom Inc. and/or its subsidiaries. > + */ > + > +/* > + * NPORT > + * > + * Port object for physical port and NPIV ports. > + */ > + > +/* > + * NPORT REFERENCE COUNTING > + * > + * A nport reference should be taken when: > + * - an nport is allocated > + * - a vport populates associated nport > + * - a remote node is allocated > + * - a unsolicited frame is processed > + * The reference should be dropped when: > + * - the unsolicited frame processesing is done > + * - the remote node is removed > + * - the vport is removed > + * - the nport is removed > + */ > + > +#include "efc.h" > + > +void > +efc_nport_cb(void *arg, int event, void *data) > +{ > + struct efc *efc = arg; > + struct efc_nport *nport = data; > + unsigned long flags = 0; > + > + efc_log_debug(efc, "nport event: %s\n", efc_sm_event_name(event)); > + > + spin_lock_irqsave(&efc->lock, flags); > + efc_sm_post_event(&nport->sm, event, NULL); > + spin_unlock_irqrestore(&efc->lock, flags); > +} > + > +static struct efc_nport * > +efc_nport_find_wwn(struct efc_domain *domain, uint64_t wwnn, uint64_t wwpn) > +{ > + struct efc_nport *nport = NULL; > + > + /* Find a nport, given the WWNN and WWPN */ > + list_for_each_entry(nport, &domain->nport_list, list_entry) { > + if (nport->wwnn == wwnn && nport->wwpn == wwpn) > + return nport; > + } > + return NULL; > +} > + > +static void > +_efc_nport_free(struct kref *arg) > +{ > + struct efc_nport *nport = container_of(arg, struct efc_nport, ref); > + > + kfree(nport); > +} > + > +struct efc_nport * > +efc_nport_alloc(struct efc_domain *domain, uint64_t wwpn, uint64_t wwnn, > + u32 fc_id, bool enable_ini, bool enable_tgt) > +{ > + struct efc_nport *nport; > + > + if (domain->efc->enable_ini) > + enable_ini = 0; > + > + /* Return a failure if this nport has already been allocated */ > + if ((wwpn != 0) || (wwnn != 0)) { > + nport = efc_nport_find_wwn(domain, wwnn, wwpn); > + if (nport) { > + efc_log_err(domain->efc, > + "Err: NPORT %016llX %016llX already allocated\n", > + wwnn, wwpn); > + return NULL; > + } > + } > + > + nport = kzalloc(sizeof(*nport), GFP_ATOMIC); > + if (!nport) > + return nport; > + > + /* initialize refcount */ > + kref_init(&nport->ref); > + nport->release = _efc_nport_free; > + > + nport->efc = domain->efc; > + snprintf(nport->display_name, sizeof(nport->display_name), "------"); > + nport->domain = domain; > + xa_init(&nport->lookup); > + nport->instance_index = domain->nport_count++; > + nport->sm.app = nport; > + nport->enable_ini = enable_ini; > + nport->enable_tgt = enable_tgt; > + nport->enable_rscn = (nport->enable_ini || > + (nport->enable_tgt && enable_target_rscn(nport->efc))); > + > + /* Copy service parameters from domain */ > + memcpy(nport->service_params, domain->service_params, > + sizeof(struct fc_els_flogi)); > + > + /* Update requested fc_id */ > + nport->fc_id = fc_id; > + > + /* Update the nport's service parameters for the new wwn's */ > + nport->wwpn = wwpn; > + nport->wwnn = wwnn; > + snprintf(nport->wwnn_str, sizeof(nport->wwnn_str), "%016llX", > + (unsigned long long)wwnn); > + > + /* > + * if this is the "first" nport of the domain, > + * then make it the "phys" nport > + */ > + if (list_empty(&domain->nport_list)) > + domain->nport = nport; > + > + INIT_LIST_HEAD(&nport->list_entry); > + list_add_tail(&nport->list_entry, &domain->nport_list); > + > + kref_get(&domain->ref); > + > + efc_log_debug(domain->efc, "New Nport [%s]\n", nport->display_name); > + > + return nport; > +} > + > +void > +efc_nport_free(struct efc_nport *nport) > +{ > + struct efc_domain *domain; > + > + if (!nport) > + return; > + > + domain = nport->domain; > + efc_log_debug(domain->efc, "[%s] free nport\n", nport->display_name); > + list_del(&nport->list_entry); > + /* > + * if this is the physical nport, > + * then clear it out of the domain > + */ > + if (nport == domain->nport) > + domain->nport = NULL; > + > + xa_destroy(&nport->lookup); > + xa_erase(&domain->lookup, nport->fc_id); > + > + if (list_empty(&domain->nport_list)) > + efc_domain_post_event(domain, EFC_EVT_ALL_CHILD_NODES_FREE, > + NULL); > + > + kref_put(&domain->ref, domain->release); > + kref_put(&nport->ref, nport->release); > + > +} > + > +struct efc_nport * > +efc_nport_find(struct efc_domain *domain, u32 d_id) > +{ > + struct efc_nport *nport; > + > + /* Find a nport object, given an FC_ID */ > + nport = xa_load(&domain->lookup, d_id); > + if (!nport || !kref_get_unless_zero(&nport->ref)) > + return NULL; > + > + return nport; > +} > + > +int > +efc_nport_attach(struct efc_nport *nport, u32 fc_id) > +{ > + int rc; > + struct efc_node *node; > + struct efc *efc = nport->efc; > + unsigned long index; > + > + /* Set our lookup */ > + rc = xa_err(xa_store(&nport->domain->lookup, fc_id, nport, GFP_ATOMIC)); > + if (rc) { > + efc_log_err(efc, "Sport lookup store failed: %d\n", rc); > + return rc; > + } > + > + /* Update our display_name */ > + efc_node_fcid_display(fc_id, nport->display_name, > + sizeof(nport->display_name)); > + > + xa_for_each(&nport->lookup, index, node) { > + efc_node_update_display_name(node); > + } > + > + efc_log_debug(nport->efc, "[%s] attach nport: fc_id x%06x\n", > + nport->display_name, fc_id); > + > + /* Register a nport, given an FC_ID */ > + rc = efc_cmd_nport_attach(efc, nport, fc_id); > + if (rc != EFC_HW_RTN_SUCCESS) { > + efc_log_err(nport->efc, > + "efc_hw_port_attach failed: %d\n", rc); > + return EFC_FAIL; > + } > + return EFC_SUCCESS; > +} > + > +static void > +efc_nport_shutdown(struct efc_nport *nport) > +{ > + struct efc *efc = nport->efc; > + struct efc_node *node; > + unsigned long index; > + > + xa_for_each(&nport->lookup, index, node) { > + if (!(node->rnode.fc_id == FC_FID_FLOGI && nport->is_vport)) { > + efc_node_post_event(node, EFC_EVT_SHUTDOWN, NULL); > + continue; > + } > + > + /* > + * If this is a vport, logout of the fabric > + * controller so that it deletes the vport > + * on the switch. > + */ > + /* if link is down, don't send logo */ > + if (efc->link_status == EFC_LINK_STATUS_DOWN) { > + efc_node_post_event(node, EFC_EVT_SHUTDOWN, NULL); > + continue; > + } > + > + efc_log_debug(efc, "[%s] nport shutdown vport, send logo\n", > + node->display_name); > + > + if (!efc_send_logo(node)) { > + /* sent LOGO, wait for response */ > + efc_node_transition(node, __efc_d_wait_logo_rsp, NULL); > + continue; > + } > + > + /* > + * failed to send LOGO, > + * go ahead and cleanup node anyways > + */ > + node_printf(node, "Failed to send LOGO\n"); > + efc_node_post_event(node, EFC_EVT_SHUTDOWN_EXPLICIT_LOGO, NULL); > + } > +} > + > +static void > +efc_vport_link_down(struct efc_nport *nport) > +{ > + struct efc *efc = nport->efc; > + struct efc_vport *vport; > + > + /* Clear the nport reference in the vport specification */ > + list_for_each_entry(vport, &efc->vport_list, list_entry) { > + if (vport->nport == nport) { > + kref_put(&nport->ref, nport->release); > + vport->nport = NULL; > + break; > + } > + } > +} > + > +static void > +__efc_nport_common(const char *funcname, struct efc_sm_ctx *ctx, > + enum efc_sm_event evt, void *arg) > +{ > + struct efc_nport *nport = ctx->app; > + struct efc_domain *domain = nport->domain; > + struct efc *efc = nport->efc; > + > + switch (evt) { > + case EFC_EVT_ENTER: > + case EFC_EVT_REENTER: > + case EFC_EVT_EXIT: > + case EFC_EVT_ALL_CHILD_NODES_FREE: > + break; > + case EFC_EVT_NPORT_ATTACH_OK: > + efc_sm_transition(ctx, __efc_nport_attached, NULL); > + break; > + case EFC_EVT_SHUTDOWN: > + /* Flag this nport as shutting down */ > + nport->shutting_down = true; > + > + if (nport->is_vport) > + efc_vport_link_down(nport); > + > + if (xa_empty(&nport->lookup)) { > + /* Remove the nport from the domain's lookup table */ > + xa_erase(&domain->lookup, nport->fc_id); > + efc_sm_transition(ctx, __efc_nport_wait_port_free, > + NULL); > + if (efc_cmd_nport_free(efc, nport)) { > + efc_log_debug(nport->efc, > + "efc_hw_port_free failed\n"); > + /* Not much we can do, free the nport anyways */ > + efc_nport_free(nport); > + } > + } else { > + /* sm: node list is not empty / shutdown nodes */ > + efc_sm_transition(ctx, > + __efc_nport_wait_shutdown, NULL); > + efc_nport_shutdown(nport); > + } > + break; > + default: > + efc_log_debug(nport->efc, "[%s] %-20s %-20s not handled\n", > + nport->display_name, funcname, > + efc_sm_event_name(evt)); > + } > +} > + > +void > +__efc_nport_allocated(struct efc_sm_ctx *ctx, > + enum efc_sm_event evt, void *arg) > +{ > + struct efc_nport *nport = ctx->app; > + struct efc_domain *domain = nport->domain; > + > + nport_sm_trace(nport); > + > + switch (evt) { > + /* the physical nport is attached */ > + case EFC_EVT_NPORT_ATTACH_OK: > + WARN_ON(nport != domain->nport); > + efc_sm_transition(ctx, __efc_nport_attached, NULL); > + break; > + > + case EFC_EVT_NPORT_ALLOC_OK: > + /* ignore */ > + break; > + default: > + __efc_nport_common(__func__, ctx, evt, arg); > + } > +} > + > +void > +__efc_nport_vport_init(struct efc_sm_ctx *ctx, > + enum efc_sm_event evt, void *arg) > +{ > + struct efc_nport *nport = ctx->app; > + struct efc *efc = nport->efc; > + > + nport_sm_trace(nport); > + > + switch (evt) { > + case EFC_EVT_ENTER: { > + __be64 be_wwpn = cpu_to_be64(nport->wwpn); > + > + if (nport->wwpn == 0) > + efc_log_debug(efc, "vport: letting f/w select WWN\n"); > + > + if (nport->fc_id != U32_MAX) { > + efc_log_debug(efc, "vport: hard coding port id: %x\n", > + nport->fc_id); > + } > + > + efc_sm_transition(ctx, __efc_nport_vport_wait_alloc, NULL); > + /* If wwpn is zero, then we'll let the f/w */ We'll let the f/w _what_? > + if (efc_cmd_nport_alloc(efc, nport, nport->domain, > + nport->wwpn == 0 ? NULL : > + (uint8_t *)&be_wwpn)) { > + efc_log_err(efc, "Can't allocate port\n"); > + break; > + } > + > + break; > + } > + default: > + __efc_nport_common(__func__, ctx, evt, arg); > + } > +} > + > +void > +__efc_nport_vport_wait_alloc(struct efc_sm_ctx *ctx, > + enum efc_sm_event evt, void *arg) > +{ > + struct efc_nport *nport = ctx->app; > + struct efc *efc = nport->efc; > + > + nport_sm_trace(nport); > + > + switch (evt) { > + case EFC_EVT_NPORT_ALLOC_OK: { > + struct fc_els_flogi *sp; > + > + sp = (struct fc_els_flogi *)nport->service_params; > + /* > + * If we let f/w assign wwn's, > + * then nport wwn's with those returned by hw this sentence no verb ... > + */ > + if (nport->wwnn == 0) { > + nport->wwnn = be64_to_cpu(nport->sli_wwnn); > + nport->wwpn = be64_to_cpu(nport->sli_wwpn); > + snprintf(nport->wwnn_str, sizeof(nport->wwnn_str), > + "%016llX", nport->wwpn); > + } > + > + /* Update the nport's service parameters */ > + sp->fl_wwpn = cpu_to_be64(nport->wwpn); > + sp->fl_wwnn = cpu_to_be64(nport->wwnn); > + > + /* > + * if nport->fc_id is uninitialized, > + * then request that the fabric node use FDISC > + * to find an fc_id. > + * Otherwise we're restoring vports, or we're in > + * fabric emulation mode, so attach the fc_id > + */ > + if (nport->fc_id == U32_MAX) { > + struct efc_node *fabric; > + > + fabric = efc_node_alloc(nport, FC_FID_FLOGI, false, > + false); > + if (!fabric) { > + efc_log_err(efc, "efc_node_alloc() failed\n"); > + return; > + } > + efc_node_transition(fabric, __efc_vport_fabric_init, > + NULL); > + } else { > + snprintf(nport->wwnn_str, sizeof(nport->wwnn_str), > + "%016llX", nport->wwpn); > + efc_nport_attach(nport, nport->fc_id); > + } > + efc_sm_transition(ctx, __efc_nport_vport_allocated, NULL); > + break; > + } > + default: > + __efc_nport_common(__func__, ctx, evt, arg); > + } > +} > + > +void > +__efc_nport_vport_allocated(struct efc_sm_ctx *ctx, > + enum efc_sm_event evt, void *arg) > +{ > + struct efc_nport *nport = ctx->app; > + struct efc *efc = nport->efc; > + > + nport_sm_trace(nport); > + > + /* > + * This state is entered after the nport is allocated; > + * it then waits for a fabric node > + * FDISC to complete, which requests a nport attach. > + * The nport attach complete is handled in this state. > + */ > + switch (evt) { > + case EFC_EVT_NPORT_ATTACH_OK: { > + struct efc_node *node; > + > + /* Find our fabric node, and forward this event */ > + node = efc_node_find(nport, FC_FID_FLOGI); > + if (!node) { > + efc_log_debug(efc, "can't find node %06x\n", > + FC_FID_FLOGI); > + break; > + } > + /* sm: / forward nport attach to fabric node */ > + efc_node_post_event(node, evt, NULL); > + efc_sm_transition(ctx, __efc_nport_attached, NULL); > + break; > + } > + default: > + __efc_nport_common(__func__, ctx, evt, arg); > + } > +} > + > +static void > +efc_vport_update_spec(struct efc_nport *nport) > +{ > + struct efc *efc = nport->efc; > + struct efc_vport *vport; > + unsigned long flags = 0; > + > + spin_lock_irqsave(&efc->vport_lock, flags); > + list_for_each_entry(vport, &efc->vport_list, list_entry) { > + if (vport->nport == nport) { > + vport->wwnn = nport->wwnn; > + vport->wwpn = nport->wwpn; > + vport->tgt_data = nport->tgt_data; > + vport->ini_data = nport->ini_data; > + break; > + } > + } > + spin_unlock_irqrestore(&efc->vport_lock, flags); > +} > + > +void > +__efc_nport_attached(struct efc_sm_ctx *ctx, > + enum efc_sm_event evt, void *arg) > +{ > + struct efc_nport *nport = ctx->app; > + struct efc *efc = nport->efc; > + > + nport_sm_trace(nport); > + > + switch (evt) { > + case EFC_EVT_ENTER: { > + struct efc_node *node; > + unsigned long index; > + > + efc_log_debug(efc, > + "[%s] NPORT attached WWPN %016llX WWNN %016llX\n", > + nport->display_name, > + nport->wwpn, nport->wwnn); > + > + xa_for_each(&nport->lookup, index, node) { > + efc_node_update_display_name(node); > + } > + Superfluous braces > + nport->tgt_id = nport->fc_id; > + > + efc->tt.new_nport(efc, nport); > + > + /* > + * Update the vport (if its not the physical nport) > + * parameters > + */ > + if (nport->is_vport) > + efc_vport_update_spec(nport); > + break; > + } > + > + case EFC_EVT_EXIT: > + efc_log_debug(efc, > + "[%s] NPORT deattached WWPN %016llX WWNN %016llX\n", > + nport->display_name, > + nport->wwpn, nport->wwnn); > + > + efc->tt.del_nport(efc, nport); > + break; > + default: > + __efc_nport_common(__func__, ctx, evt, arg); > + } > +} > + > + > +void > +__efc_nport_wait_shutdown(struct efc_sm_ctx *ctx, > + enum efc_sm_event evt, void *arg) > +{ > + struct efc_nport *nport = ctx->app; > + struct efc_domain *domain = nport->domain; > + struct efc *efc = nport->efc; > + > + nport_sm_trace(nport); > + > + switch (evt) { > + case EFC_EVT_NPORT_ALLOC_OK: > + case EFC_EVT_NPORT_ALLOC_FAIL: > + case EFC_EVT_NPORT_ATTACH_OK: > + case EFC_EVT_NPORT_ATTACH_FAIL: > + /* ignore these events - just wait for the all free event */ > + break; > + > + case EFC_EVT_ALL_CHILD_NODES_FREE: { > + /* > + * Remove the nport from the domain's > + * sparse vector lookup table > + */ > + xa_erase(&domain->lookup, nport->fc_id); > + efc_sm_transition(ctx, __efc_nport_wait_port_free, NULL); > + if (efc_cmd_nport_free(efc, nport)) { > + efc_log_err(nport->efc, "efc_hw_port_free failed\n"); > + /* Not much we can do, free the nport anyways */ > + efc_nport_free(nport); > + } > + break; > + } > + default: > + __efc_nport_common(__func__, ctx, evt, arg); > + } > +} > + > +void > +__efc_nport_wait_port_free(struct efc_sm_ctx *ctx, > + enum efc_sm_event evt, void *arg) > +{ > + struct efc_nport *nport = ctx->app; > + > + nport_sm_trace(nport); > + > + switch (evt) { > + case EFC_EVT_NPORT_ATTACH_OK: > + /* Ignore as we are waiting for the free CB */ > + break; > + case EFC_EVT_NPORT_FREE_OK: { > + /* All done, free myself */ > + efc_nport_free(nport); > + break; Set ctx->app to NULL? > + } > + default: > + __efc_nport_common(__func__, ctx, evt, arg); > + } > +} > + > +static int > +efc_vport_nport_alloc(struct efc_domain *domain, struct efc_vport *vport) > +{ > + struct efc_nport *nport; > + > + lockdep_assert_held(&domain->efc->lock); > + > + nport = efc_nport_alloc(domain, vport->wwpn, vport->wwnn, vport->fc_id, > + vport->enable_ini, vport->enable_tgt); > + vport->nport = nport; > + if (!nport) > + return EFC_FAIL; > + > + kref_get(&nport->ref); > + nport->is_vport = true; > + nport->tgt_data = vport->tgt_data; > + nport->ini_data = vport->ini_data; > + > + efc_sm_transition(&nport->sm, __efc_nport_vport_init, NULL); > + > + return EFC_SUCCESS; > +} > + > +int > +efc_vport_start(struct efc_domain *domain) > +{ > + struct efc *efc = domain->efc; > + struct efc_vport *vport; > + struct efc_vport *next; > + int rc = EFC_SUCCESS; > + unsigned long flags = 0; > + > + /* Use the vport spec to find the associated vports and start them */ > + spin_lock_irqsave(&efc->vport_lock, flags); > + list_for_each_entry_safe(vport, next, &efc->vport_list, list_entry) { > + if (!vport->nport) { > + if (efc_vport_nport_alloc(domain, vport)) > + rc = EFC_FAIL; > + } > + } > + spin_unlock_irqrestore(&efc->vport_lock, flags); > + > + return rc; > +} > + > +int > +efc_nport_vport_new(struct efc_domain *domain, uint64_t wwpn, uint64_t wwnn, > + u32 fc_id, bool ini, bool tgt, void *tgt_data, > + void *ini_data) > +{ > + struct efc *efc = domain->efc; > + struct efc_vport *vport; > + int rc = EFC_SUCCESS; > + unsigned long flags = 0; > + > + if (ini && domain->efc->enable_ini == 0) { > + efc_log_debug(efc, > + "driver initiator functionality not enabled\n"); > + return EFC_FAIL; > + } > + > + if (tgt && domain->efc->enable_tgt == 0) { > + efc_log_debug(efc, > + "driver target functionality not enabled\n"); > + return EFC_FAIL; > + } > + > + /* > + * Create a vport spec if we need to recreate > + * this vport after a link up event > + */ > + vport = efc_vport_create_spec(domain->efc, wwnn, wwpn, fc_id, ini, tgt, > + tgt_data, ini_data); > + if (!vport) { > + efc_log_err(efc, "failed to create vport object entry\n"); > + return EFC_FAIL; > + } > + > + spin_lock_irqsave(&efc->lock, flags); > + rc = efc_vport_nport_alloc(domain, vport); > + spin_unlock_irqrestore(&efc->lock, flags); > + > + return rc; > +} > + > +int > +efc_nport_vport_del(struct efc *efc, struct efc_domain *domain, > + u64 wwpn, uint64_t wwnn) > +{ > + struct efc_nport *nport; > + int found = 0; > + struct efc_vport *vport; > + struct efc_vport *next; > + unsigned long flags = 0; > + > + spin_lock_irqsave(&efc->vport_lock, flags); > + /* walk the efc_vport_list and remove from there */ > + list_for_each_entry_safe(vport, next, &efc->vport_list, list_entry) { > + if (vport->wwpn == wwpn && vport->wwnn == wwnn) { > + list_del(&vport->list_entry); > + kfree(vport); > + break; > + } > + } > + spin_unlock_irqrestore(&efc->vport_lock, flags); > + > + if (!domain) { > + /* No domain means no nport to look for */ > + return EFC_SUCCESS; > + } > + > + spin_lock_irqsave(&efc->lock, flags); > + list_for_each_entry(nport, &domain->nport_list, list_entry) { > + if (nport->wwpn == wwpn && nport->wwnn == wwnn) { > + found = 1; > + break; > + } > + } > + > + if (found) { > + kref_put(&nport->ref, nport->release); > + /* Shutdown this NPORT */ > + efc_sm_post_event(&nport->sm, EFC_EVT_SHUTDOWN, NULL); > + } Bah. Using the loop variable outside of the loop is bad style. Why not moving this directly within the loop and drop the 'found' variable? > + spin_unlock_irqrestore(&efc->lock, flags); > + return EFC_SUCCESS; > +} > + > +void > +efc_vport_del_all(struct efc *efc) > +{ > + struct efc_vport *vport; > + struct efc_vport *next; > + unsigned long flags = 0; > + > + spin_lock_irqsave(&efc->vport_lock, flags); > + list_for_each_entry_safe(vport, next, &efc->vport_list, list_entry) { > + list_del(&vport->list_entry); > + kfree(vport); > + } > + spin_unlock_irqrestore(&efc->vport_lock, flags); > +} > + > +struct efc_vport * > +efc_vport_create_spec(struct efc *efc, uint64_t wwnn, uint64_t wwpn, > + u32 fc_id, bool enable_ini, > + bool enable_tgt, void *tgt_data, void *ini_data) > +{ > + struct efc_vport *vport; > + unsigned long flags = 0; > + > + /* > + * walk the efc_vport_list and return failure > + * if a valid(vport with non zero WWPN and WWNN) vport entry > + * is already created > + */ > + spin_lock_irqsave(&efc->vport_lock, flags); > + list_for_each_entry(vport, &efc->vport_list, list_entry) { > + if ((wwpn && vport->wwpn == wwpn) && > + (wwnn && vport->wwnn == wwnn)) { > + efc_log_err(efc, > + "Failed: VPORT %016llX %016llX already allocated\n", > + wwnn, wwpn); > + spin_unlock_irqrestore(&efc->vport_lock, flags); > + return NULL; > + } > + } > + > + vport = kzalloc(sizeof(*vport), GFP_ATOMIC); > + if (!vport) { > + spin_unlock_irqrestore(&efc->vport_lock, flags); > + return NULL; > + } > + > + vport->wwnn = wwnn; > + vport->wwpn = wwpn; > + vport->fc_id = fc_id; > + vport->enable_tgt = enable_tgt; > + vport->enable_ini = enable_ini; > + vport->tgt_data = tgt_data; > + vport->ini_data = ini_data; > + > + INIT_LIST_HEAD(&vport->list_entry); > + list_add_tail(&vport->list_entry, &efc->vport_list); > + spin_unlock_irqrestore(&efc->vport_lock, flags); > + return vport; > +} > diff --git a/drivers/scsi/elx/libefc/efc_nport.h b/drivers/scsi/elx/libefc/efc_nport.h > new file mode 100644 > index 000000000000..b575ea205bbf > --- /dev/null > +++ b/drivers/scsi/elx/libefc/efc_nport.h > @@ -0,0 +1,50 @@ > +/* SPDX-License-Identifier: GPL-2.0 */ > +/* > + * Copyright (C) 2021 Broadcom. All Rights Reserved. The term > + * “Broadcom” refers to Broadcom Inc. and/or its subsidiaries. > + */ > + > +/** > + * EFC FC port (NPORT) exported declarations > + * > + */ > + > +#ifndef __EFC_NPORT_H__ > +#define __EFC_NPORT_H__ > + > +struct efc_nport * > +efc_nport_find(struct efc_domain *domain, u32 d_id); > +struct efc_nport * > +efc_nport_alloc(struct efc_domain *domain, uint64_t wwpn, uint64_t wwnn, > + u32 fc_id, bool enable_ini, bool enable_tgt); > +void > +efc_nport_free(struct efc_nport *nport); > +int > +efc_nport_attach(struct efc_nport *nport, u32 fc_id); > + > +void > +__efc_nport_allocated(struct efc_sm_ctx *ctx, > + enum efc_sm_event evt, void *arg); > +void > +__efc_nport_wait_shutdown(struct efc_sm_ctx *ctx, > + enum efc_sm_event evt, void *arg); > +void > +__efc_nport_wait_port_free(struct efc_sm_ctx *ctx, > + enum efc_sm_event evt, void *arg); > +void > +__efc_nport_vport_init(struct efc_sm_ctx *ctx, > + enum efc_sm_event evt, void *arg); > +void > +__efc_nport_vport_wait_alloc(struct efc_sm_ctx *ctx, > + enum efc_sm_event evt, void *arg); > +void > +__efc_nport_vport_allocated(struct efc_sm_ctx *ctx, > + enum efc_sm_event evt, void *arg); > +void > +__efc_nport_attached(struct efc_sm_ctx *ctx, > + enum efc_sm_event evt, void *arg); > + > +int > +efc_vport_start(struct efc_domain *domain); > + > +#endif /* __EFC_NPORT_H__ */ > Cheers, Hannes -- Dr. Hannes Reinecke Kernel Storage Architect hare@suse.de +49 911 74053 688 SUSE Software Solutions Germany GmbH, 90409 Nürnberg GF: F. Imendörffer, HRB 36809 (AG Nürnberg)
diff --git a/drivers/scsi/elx/libefc/efc_nport.c b/drivers/scsi/elx/libefc/efc_nport.c new file mode 100644 index 000000000000..bd50f0320631 --- /dev/null +++ b/drivers/scsi/elx/libefc/efc_nport.c @@ -0,0 +1,792 @@ +// SPDX-License-Identifier: GPL-2.0 +/* + * Copyright (C) 2021 Broadcom. All Rights Reserved. The term + * “Broadcom” refers to Broadcom Inc. and/or its subsidiaries. + */ + +/* + * NPORT + * + * Port object for physical port and NPIV ports. + */ + +/* + * NPORT REFERENCE COUNTING + * + * A nport reference should be taken when: + * - an nport is allocated + * - a vport populates associated nport + * - a remote node is allocated + * - a unsolicited frame is processed + * The reference should be dropped when: + * - the unsolicited frame processesing is done + * - the remote node is removed + * - the vport is removed + * - the nport is removed + */ + +#include "efc.h" + +void +efc_nport_cb(void *arg, int event, void *data) +{ + struct efc *efc = arg; + struct efc_nport *nport = data; + unsigned long flags = 0; + + efc_log_debug(efc, "nport event: %s\n", efc_sm_event_name(event)); + + spin_lock_irqsave(&efc->lock, flags); + efc_sm_post_event(&nport->sm, event, NULL); + spin_unlock_irqrestore(&efc->lock, flags); +} + +static struct efc_nport * +efc_nport_find_wwn(struct efc_domain *domain, uint64_t wwnn, uint64_t wwpn) +{ + struct efc_nport *nport = NULL; + + /* Find a nport, given the WWNN and WWPN */ + list_for_each_entry(nport, &domain->nport_list, list_entry) { + if (nport->wwnn == wwnn && nport->wwpn == wwpn) + return nport; + } + return NULL; +} + +static void +_efc_nport_free(struct kref *arg) +{ + struct efc_nport *nport = container_of(arg, struct efc_nport, ref); + + kfree(nport); +} + +struct efc_nport * +efc_nport_alloc(struct efc_domain *domain, uint64_t wwpn, uint64_t wwnn, + u32 fc_id, bool enable_ini, bool enable_tgt) +{ + struct efc_nport *nport; + + if (domain->efc->enable_ini) + enable_ini = 0; + + /* Return a failure if this nport has already been allocated */ + if ((wwpn != 0) || (wwnn != 0)) { + nport = efc_nport_find_wwn(domain, wwnn, wwpn); + if (nport) { + efc_log_err(domain->efc, + "Err: NPORT %016llX %016llX already allocated\n", + wwnn, wwpn); + return NULL; + } + } + + nport = kzalloc(sizeof(*nport), GFP_ATOMIC); + if (!nport) + return nport; + + /* initialize refcount */ + kref_init(&nport->ref); + nport->release = _efc_nport_free; + + nport->efc = domain->efc; + snprintf(nport->display_name, sizeof(nport->display_name), "------"); + nport->domain = domain; + xa_init(&nport->lookup); + nport->instance_index = domain->nport_count++; + nport->sm.app = nport; + nport->enable_ini = enable_ini; + nport->enable_tgt = enable_tgt; + nport->enable_rscn = (nport->enable_ini || + (nport->enable_tgt && enable_target_rscn(nport->efc))); + + /* Copy service parameters from domain */ + memcpy(nport->service_params, domain->service_params, + sizeof(struct fc_els_flogi)); + + /* Update requested fc_id */ + nport->fc_id = fc_id; + + /* Update the nport's service parameters for the new wwn's */ + nport->wwpn = wwpn; + nport->wwnn = wwnn; + snprintf(nport->wwnn_str, sizeof(nport->wwnn_str), "%016llX", + (unsigned long long)wwnn); + + /* + * if this is the "first" nport of the domain, + * then make it the "phys" nport + */ + if (list_empty(&domain->nport_list)) + domain->nport = nport; + + INIT_LIST_HEAD(&nport->list_entry); + list_add_tail(&nport->list_entry, &domain->nport_list); + + kref_get(&domain->ref); + + efc_log_debug(domain->efc, "New Nport [%s]\n", nport->display_name); + + return nport; +} + +void +efc_nport_free(struct efc_nport *nport) +{ + struct efc_domain *domain; + + if (!nport) + return; + + domain = nport->domain; + efc_log_debug(domain->efc, "[%s] free nport\n", nport->display_name); + list_del(&nport->list_entry); + /* + * if this is the physical nport, + * then clear it out of the domain + */ + if (nport == domain->nport) + domain->nport = NULL; + + xa_destroy(&nport->lookup); + xa_erase(&domain->lookup, nport->fc_id); + + if (list_empty(&domain->nport_list)) + efc_domain_post_event(domain, EFC_EVT_ALL_CHILD_NODES_FREE, + NULL); + + kref_put(&domain->ref, domain->release); + kref_put(&nport->ref, nport->release); + +} + +struct efc_nport * +efc_nport_find(struct efc_domain *domain, u32 d_id) +{ + struct efc_nport *nport; + + /* Find a nport object, given an FC_ID */ + nport = xa_load(&domain->lookup, d_id); + if (!nport || !kref_get_unless_zero(&nport->ref)) + return NULL; + + return nport; +} + +int +efc_nport_attach(struct efc_nport *nport, u32 fc_id) +{ + int rc; + struct efc_node *node; + struct efc *efc = nport->efc; + unsigned long index; + + /* Set our lookup */ + rc = xa_err(xa_store(&nport->domain->lookup, fc_id, nport, GFP_ATOMIC)); + if (rc) { + efc_log_err(efc, "Sport lookup store failed: %d\n", rc); + return rc; + } + + /* Update our display_name */ + efc_node_fcid_display(fc_id, nport->display_name, + sizeof(nport->display_name)); + + xa_for_each(&nport->lookup, index, node) { + efc_node_update_display_name(node); + } + + efc_log_debug(nport->efc, "[%s] attach nport: fc_id x%06x\n", + nport->display_name, fc_id); + + /* Register a nport, given an FC_ID */ + rc = efc_cmd_nport_attach(efc, nport, fc_id); + if (rc != EFC_HW_RTN_SUCCESS) { + efc_log_err(nport->efc, + "efc_hw_port_attach failed: %d\n", rc); + return EFC_FAIL; + } + return EFC_SUCCESS; +} + +static void +efc_nport_shutdown(struct efc_nport *nport) +{ + struct efc *efc = nport->efc; + struct efc_node *node; + unsigned long index; + + xa_for_each(&nport->lookup, index, node) { + if (!(node->rnode.fc_id == FC_FID_FLOGI && nport->is_vport)) { + efc_node_post_event(node, EFC_EVT_SHUTDOWN, NULL); + continue; + } + + /* + * If this is a vport, logout of the fabric + * controller so that it deletes the vport + * on the switch. + */ + /* if link is down, don't send logo */ + if (efc->link_status == EFC_LINK_STATUS_DOWN) { + efc_node_post_event(node, EFC_EVT_SHUTDOWN, NULL); + continue; + } + + efc_log_debug(efc, "[%s] nport shutdown vport, send logo\n", + node->display_name); + + if (!efc_send_logo(node)) { + /* sent LOGO, wait for response */ + efc_node_transition(node, __efc_d_wait_logo_rsp, NULL); + continue; + } + + /* + * failed to send LOGO, + * go ahead and cleanup node anyways + */ + node_printf(node, "Failed to send LOGO\n"); + efc_node_post_event(node, EFC_EVT_SHUTDOWN_EXPLICIT_LOGO, NULL); + } +} + +static void +efc_vport_link_down(struct efc_nport *nport) +{ + struct efc *efc = nport->efc; + struct efc_vport *vport; + + /* Clear the nport reference in the vport specification */ + list_for_each_entry(vport, &efc->vport_list, list_entry) { + if (vport->nport == nport) { + kref_put(&nport->ref, nport->release); + vport->nport = NULL; + break; + } + } +} + +static void +__efc_nport_common(const char *funcname, struct efc_sm_ctx *ctx, + enum efc_sm_event evt, void *arg) +{ + struct efc_nport *nport = ctx->app; + struct efc_domain *domain = nport->domain; + struct efc *efc = nport->efc; + + switch (evt) { + case EFC_EVT_ENTER: + case EFC_EVT_REENTER: + case EFC_EVT_EXIT: + case EFC_EVT_ALL_CHILD_NODES_FREE: + break; + case EFC_EVT_NPORT_ATTACH_OK: + efc_sm_transition(ctx, __efc_nport_attached, NULL); + break; + case EFC_EVT_SHUTDOWN: + /* Flag this nport as shutting down */ + nport->shutting_down = true; + + if (nport->is_vport) + efc_vport_link_down(nport); + + if (xa_empty(&nport->lookup)) { + /* Remove the nport from the domain's lookup table */ + xa_erase(&domain->lookup, nport->fc_id); + efc_sm_transition(ctx, __efc_nport_wait_port_free, + NULL); + if (efc_cmd_nport_free(efc, nport)) { + efc_log_debug(nport->efc, + "efc_hw_port_free failed\n"); + /* Not much we can do, free the nport anyways */ + efc_nport_free(nport); + } + } else { + /* sm: node list is not empty / shutdown nodes */ + efc_sm_transition(ctx, + __efc_nport_wait_shutdown, NULL); + efc_nport_shutdown(nport); + } + break; + default: + efc_log_debug(nport->efc, "[%s] %-20s %-20s not handled\n", + nport->display_name, funcname, + efc_sm_event_name(evt)); + } +} + +void +__efc_nport_allocated(struct efc_sm_ctx *ctx, + enum efc_sm_event evt, void *arg) +{ + struct efc_nport *nport = ctx->app; + struct efc_domain *domain = nport->domain; + + nport_sm_trace(nport); + + switch (evt) { + /* the physical nport is attached */ + case EFC_EVT_NPORT_ATTACH_OK: + WARN_ON(nport != domain->nport); + efc_sm_transition(ctx, __efc_nport_attached, NULL); + break; + + case EFC_EVT_NPORT_ALLOC_OK: + /* ignore */ + break; + default: + __efc_nport_common(__func__, ctx, evt, arg); + } +} + +void +__efc_nport_vport_init(struct efc_sm_ctx *ctx, + enum efc_sm_event evt, void *arg) +{ + struct efc_nport *nport = ctx->app; + struct efc *efc = nport->efc; + + nport_sm_trace(nport); + + switch (evt) { + case EFC_EVT_ENTER: { + __be64 be_wwpn = cpu_to_be64(nport->wwpn); + + if (nport->wwpn == 0) + efc_log_debug(efc, "vport: letting f/w select WWN\n"); + + if (nport->fc_id != U32_MAX) { + efc_log_debug(efc, "vport: hard coding port id: %x\n", + nport->fc_id); + } + + efc_sm_transition(ctx, __efc_nport_vport_wait_alloc, NULL); + /* If wwpn is zero, then we'll let the f/w */ + if (efc_cmd_nport_alloc(efc, nport, nport->domain, + nport->wwpn == 0 ? NULL : + (uint8_t *)&be_wwpn)) { + efc_log_err(efc, "Can't allocate port\n"); + break; + } + + break; + } + default: + __efc_nport_common(__func__, ctx, evt, arg); + } +} + +void +__efc_nport_vport_wait_alloc(struct efc_sm_ctx *ctx, + enum efc_sm_event evt, void *arg) +{ + struct efc_nport *nport = ctx->app; + struct efc *efc = nport->efc; + + nport_sm_trace(nport); + + switch (evt) { + case EFC_EVT_NPORT_ALLOC_OK: { + struct fc_els_flogi *sp; + + sp = (struct fc_els_flogi *)nport->service_params; + /* + * If we let f/w assign wwn's, + * then nport wwn's with those returned by hw + */ + if (nport->wwnn == 0) { + nport->wwnn = be64_to_cpu(nport->sli_wwnn); + nport->wwpn = be64_to_cpu(nport->sli_wwpn); + snprintf(nport->wwnn_str, sizeof(nport->wwnn_str), + "%016llX", nport->wwpn); + } + + /* Update the nport's service parameters */ + sp->fl_wwpn = cpu_to_be64(nport->wwpn); + sp->fl_wwnn = cpu_to_be64(nport->wwnn); + + /* + * if nport->fc_id is uninitialized, + * then request that the fabric node use FDISC + * to find an fc_id. + * Otherwise we're restoring vports, or we're in + * fabric emulation mode, so attach the fc_id + */ + if (nport->fc_id == U32_MAX) { + struct efc_node *fabric; + + fabric = efc_node_alloc(nport, FC_FID_FLOGI, false, + false); + if (!fabric) { + efc_log_err(efc, "efc_node_alloc() failed\n"); + return; + } + efc_node_transition(fabric, __efc_vport_fabric_init, + NULL); + } else { + snprintf(nport->wwnn_str, sizeof(nport->wwnn_str), + "%016llX", nport->wwpn); + efc_nport_attach(nport, nport->fc_id); + } + efc_sm_transition(ctx, __efc_nport_vport_allocated, NULL); + break; + } + default: + __efc_nport_common(__func__, ctx, evt, arg); + } +} + +void +__efc_nport_vport_allocated(struct efc_sm_ctx *ctx, + enum efc_sm_event evt, void *arg) +{ + struct efc_nport *nport = ctx->app; + struct efc *efc = nport->efc; + + nport_sm_trace(nport); + + /* + * This state is entered after the nport is allocated; + * it then waits for a fabric node + * FDISC to complete, which requests a nport attach. + * The nport attach complete is handled in this state. + */ + switch (evt) { + case EFC_EVT_NPORT_ATTACH_OK: { + struct efc_node *node; + + /* Find our fabric node, and forward this event */ + node = efc_node_find(nport, FC_FID_FLOGI); + if (!node) { + efc_log_debug(efc, "can't find node %06x\n", + FC_FID_FLOGI); + break; + } + /* sm: / forward nport attach to fabric node */ + efc_node_post_event(node, evt, NULL); + efc_sm_transition(ctx, __efc_nport_attached, NULL); + break; + } + default: + __efc_nport_common(__func__, ctx, evt, arg); + } +} + +static void +efc_vport_update_spec(struct efc_nport *nport) +{ + struct efc *efc = nport->efc; + struct efc_vport *vport; + unsigned long flags = 0; + + spin_lock_irqsave(&efc->vport_lock, flags); + list_for_each_entry(vport, &efc->vport_list, list_entry) { + if (vport->nport == nport) { + vport->wwnn = nport->wwnn; + vport->wwpn = nport->wwpn; + vport->tgt_data = nport->tgt_data; + vport->ini_data = nport->ini_data; + break; + } + } + spin_unlock_irqrestore(&efc->vport_lock, flags); +} + +void +__efc_nport_attached(struct efc_sm_ctx *ctx, + enum efc_sm_event evt, void *arg) +{ + struct efc_nport *nport = ctx->app; + struct efc *efc = nport->efc; + + nport_sm_trace(nport); + + switch (evt) { + case EFC_EVT_ENTER: { + struct efc_node *node; + unsigned long index; + + efc_log_debug(efc, + "[%s] NPORT attached WWPN %016llX WWNN %016llX\n", + nport->display_name, + nport->wwpn, nport->wwnn); + + xa_for_each(&nport->lookup, index, node) { + efc_node_update_display_name(node); + } + + nport->tgt_id = nport->fc_id; + + efc->tt.new_nport(efc, nport); + + /* + * Update the vport (if its not the physical nport) + * parameters + */ + if (nport->is_vport) + efc_vport_update_spec(nport); + break; + } + + case EFC_EVT_EXIT: + efc_log_debug(efc, + "[%s] NPORT deattached WWPN %016llX WWNN %016llX\n", + nport->display_name, + nport->wwpn, nport->wwnn); + + efc->tt.del_nport(efc, nport); + break; + default: + __efc_nport_common(__func__, ctx, evt, arg); + } +} + + +void +__efc_nport_wait_shutdown(struct efc_sm_ctx *ctx, + enum efc_sm_event evt, void *arg) +{ + struct efc_nport *nport = ctx->app; + struct efc_domain *domain = nport->domain; + struct efc *efc = nport->efc; + + nport_sm_trace(nport); + + switch (evt) { + case EFC_EVT_NPORT_ALLOC_OK: + case EFC_EVT_NPORT_ALLOC_FAIL: + case EFC_EVT_NPORT_ATTACH_OK: + case EFC_EVT_NPORT_ATTACH_FAIL: + /* ignore these events - just wait for the all free event */ + break; + + case EFC_EVT_ALL_CHILD_NODES_FREE: { + /* + * Remove the nport from the domain's + * sparse vector lookup table + */ + xa_erase(&domain->lookup, nport->fc_id); + efc_sm_transition(ctx, __efc_nport_wait_port_free, NULL); + if (efc_cmd_nport_free(efc, nport)) { + efc_log_err(nport->efc, "efc_hw_port_free failed\n"); + /* Not much we can do, free the nport anyways */ + efc_nport_free(nport); + } + break; + } + default: + __efc_nport_common(__func__, ctx, evt, arg); + } +} + +void +__efc_nport_wait_port_free(struct efc_sm_ctx *ctx, + enum efc_sm_event evt, void *arg) +{ + struct efc_nport *nport = ctx->app; + + nport_sm_trace(nport); + + switch (evt) { + case EFC_EVT_NPORT_ATTACH_OK: + /* Ignore as we are waiting for the free CB */ + break; + case EFC_EVT_NPORT_FREE_OK: { + /* All done, free myself */ + efc_nport_free(nport); + break; + } + default: + __efc_nport_common(__func__, ctx, evt, arg); + } +} + +static int +efc_vport_nport_alloc(struct efc_domain *domain, struct efc_vport *vport) +{ + struct efc_nport *nport; + + lockdep_assert_held(&domain->efc->lock); + + nport = efc_nport_alloc(domain, vport->wwpn, vport->wwnn, vport->fc_id, + vport->enable_ini, vport->enable_tgt); + vport->nport = nport; + if (!nport) + return EFC_FAIL; + + kref_get(&nport->ref); + nport->is_vport = true; + nport->tgt_data = vport->tgt_data; + nport->ini_data = vport->ini_data; + + efc_sm_transition(&nport->sm, __efc_nport_vport_init, NULL); + + return EFC_SUCCESS; +} + +int +efc_vport_start(struct efc_domain *domain) +{ + struct efc *efc = domain->efc; + struct efc_vport *vport; + struct efc_vport *next; + int rc = EFC_SUCCESS; + unsigned long flags = 0; + + /* Use the vport spec to find the associated vports and start them */ + spin_lock_irqsave(&efc->vport_lock, flags); + list_for_each_entry_safe(vport, next, &efc->vport_list, list_entry) { + if (!vport->nport) { + if (efc_vport_nport_alloc(domain, vport)) + rc = EFC_FAIL; + } + } + spin_unlock_irqrestore(&efc->vport_lock, flags); + + return rc; +} + +int +efc_nport_vport_new(struct efc_domain *domain, uint64_t wwpn, uint64_t wwnn, + u32 fc_id, bool ini, bool tgt, void *tgt_data, + void *ini_data) +{ + struct efc *efc = domain->efc; + struct efc_vport *vport; + int rc = EFC_SUCCESS; + unsigned long flags = 0; + + if (ini && domain->efc->enable_ini == 0) { + efc_log_debug(efc, + "driver initiator functionality not enabled\n"); + return EFC_FAIL; + } + + if (tgt && domain->efc->enable_tgt == 0) { + efc_log_debug(efc, + "driver target functionality not enabled\n"); + return EFC_FAIL; + } + + /* + * Create a vport spec if we need to recreate + * this vport after a link up event + */ + vport = efc_vport_create_spec(domain->efc, wwnn, wwpn, fc_id, ini, tgt, + tgt_data, ini_data); + if (!vport) { + efc_log_err(efc, "failed to create vport object entry\n"); + return EFC_FAIL; + } + + spin_lock_irqsave(&efc->lock, flags); + rc = efc_vport_nport_alloc(domain, vport); + spin_unlock_irqrestore(&efc->lock, flags); + + return rc; +} + +int +efc_nport_vport_del(struct efc *efc, struct efc_domain *domain, + u64 wwpn, uint64_t wwnn) +{ + struct efc_nport *nport; + int found = 0; + struct efc_vport *vport; + struct efc_vport *next; + unsigned long flags = 0; + + spin_lock_irqsave(&efc->vport_lock, flags); + /* walk the efc_vport_list and remove from there */ + list_for_each_entry_safe(vport, next, &efc->vport_list, list_entry) { + if (vport->wwpn == wwpn && vport->wwnn == wwnn) { + list_del(&vport->list_entry); + kfree(vport); + break; + } + } + spin_unlock_irqrestore(&efc->vport_lock, flags); + + if (!domain) { + /* No domain means no nport to look for */ + return EFC_SUCCESS; + } + + spin_lock_irqsave(&efc->lock, flags); + list_for_each_entry(nport, &domain->nport_list, list_entry) { + if (nport->wwpn == wwpn && nport->wwnn == wwnn) { + found = 1; + break; + } + } + + if (found) { + kref_put(&nport->ref, nport->release); + /* Shutdown this NPORT */ + efc_sm_post_event(&nport->sm, EFC_EVT_SHUTDOWN, NULL); + } + spin_unlock_irqrestore(&efc->lock, flags); + return EFC_SUCCESS; +} + +void +efc_vport_del_all(struct efc *efc) +{ + struct efc_vport *vport; + struct efc_vport *next; + unsigned long flags = 0; + + spin_lock_irqsave(&efc->vport_lock, flags); + list_for_each_entry_safe(vport, next, &efc->vport_list, list_entry) { + list_del(&vport->list_entry); + kfree(vport); + } + spin_unlock_irqrestore(&efc->vport_lock, flags); +} + +struct efc_vport * +efc_vport_create_spec(struct efc *efc, uint64_t wwnn, uint64_t wwpn, + u32 fc_id, bool enable_ini, + bool enable_tgt, void *tgt_data, void *ini_data) +{ + struct efc_vport *vport; + unsigned long flags = 0; + + /* + * walk the efc_vport_list and return failure + * if a valid(vport with non zero WWPN and WWNN) vport entry + * is already created + */ + spin_lock_irqsave(&efc->vport_lock, flags); + list_for_each_entry(vport, &efc->vport_list, list_entry) { + if ((wwpn && vport->wwpn == wwpn) && + (wwnn && vport->wwnn == wwnn)) { + efc_log_err(efc, + "Failed: VPORT %016llX %016llX already allocated\n", + wwnn, wwpn); + spin_unlock_irqrestore(&efc->vport_lock, flags); + return NULL; + } + } + + vport = kzalloc(sizeof(*vport), GFP_ATOMIC); + if (!vport) { + spin_unlock_irqrestore(&efc->vport_lock, flags); + return NULL; + } + + vport->wwnn = wwnn; + vport->wwpn = wwpn; + vport->fc_id = fc_id; + vport->enable_tgt = enable_tgt; + vport->enable_ini = enable_ini; + vport->tgt_data = tgt_data; + vport->ini_data = ini_data; + + INIT_LIST_HEAD(&vport->list_entry); + list_add_tail(&vport->list_entry, &efc->vport_list); + spin_unlock_irqrestore(&efc->vport_lock, flags); + return vport; +} diff --git a/drivers/scsi/elx/libefc/efc_nport.h b/drivers/scsi/elx/libefc/efc_nport.h new file mode 100644 index 000000000000..b575ea205bbf --- /dev/null +++ b/drivers/scsi/elx/libefc/efc_nport.h @@ -0,0 +1,50 @@ +/* SPDX-License-Identifier: GPL-2.0 */ +/* + * Copyright (C) 2021 Broadcom. All Rights Reserved. The term + * “Broadcom” refers to Broadcom Inc. and/or its subsidiaries. + */ + +/** + * EFC FC port (NPORT) exported declarations + * + */ + +#ifndef __EFC_NPORT_H__ +#define __EFC_NPORT_H__ + +struct efc_nport * +efc_nport_find(struct efc_domain *domain, u32 d_id); +struct efc_nport * +efc_nport_alloc(struct efc_domain *domain, uint64_t wwpn, uint64_t wwnn, + u32 fc_id, bool enable_ini, bool enable_tgt); +void +efc_nport_free(struct efc_nport *nport); +int +efc_nport_attach(struct efc_nport *nport, u32 fc_id); + +void +__efc_nport_allocated(struct efc_sm_ctx *ctx, + enum efc_sm_event evt, void *arg); +void +__efc_nport_wait_shutdown(struct efc_sm_ctx *ctx, + enum efc_sm_event evt, void *arg); +void +__efc_nport_wait_port_free(struct efc_sm_ctx *ctx, + enum efc_sm_event evt, void *arg); +void +__efc_nport_vport_init(struct efc_sm_ctx *ctx, + enum efc_sm_event evt, void *arg); +void +__efc_nport_vport_wait_alloc(struct efc_sm_ctx *ctx, + enum efc_sm_event evt, void *arg); +void +__efc_nport_vport_allocated(struct efc_sm_ctx *ctx, + enum efc_sm_event evt, void *arg); +void +__efc_nport_attached(struct efc_sm_ctx *ctx, + enum efc_sm_event evt, void *arg); + +int +efc_vport_start(struct efc_domain *domain); + +#endif /* __EFC_NPORT_H__ */