1089 lines
26 KiB
C
1089 lines
26 KiB
C
// SPDX-License-Identifier: GPL-2.0
|
|
/*
|
|
* Copyright (C) 2021 Broadcom. All Rights Reserved. The term
|
|
* “Broadcom” refers to Broadcom Inc. and/or its subsidiaries.
|
|
*/
|
|
|
|
/*
|
|
* domain_sm Domain State Machine: States
|
|
*/
|
|
|
|
#include "efc.h"
|
|
|
|
int
|
|
efc_domain_cb(void *arg, int event, void *data)
|
|
{
|
|
struct efc *efc = arg;
|
|
struct efc_domain *domain = NULL;
|
|
int rc = 0;
|
|
unsigned long flags = 0;
|
|
|
|
if (event != EFC_HW_DOMAIN_FOUND)
|
|
domain = data;
|
|
|
|
/* Accept domain callback events from the user driver */
|
|
spin_lock_irqsave(&efc->lock, flags);
|
|
switch (event) {
|
|
case EFC_HW_DOMAIN_FOUND: {
|
|
u64 fcf_wwn = 0;
|
|
struct efc_domain_record *drec = data;
|
|
|
|
/* extract the fcf_wwn */
|
|
fcf_wwn = be64_to_cpu(*((__be64 *)drec->wwn));
|
|
|
|
efc_log_debug(efc, "Domain found: wwn %016llX\n", fcf_wwn);
|
|
|
|
/* lookup domain, or allocate a new one */
|
|
domain = efc->domain;
|
|
if (!domain) {
|
|
domain = efc_domain_alloc(efc, fcf_wwn);
|
|
if (!domain) {
|
|
efc_log_err(efc, "efc_domain_alloc() failed\n");
|
|
rc = -1;
|
|
break;
|
|
}
|
|
efc_sm_transition(&domain->drvsm, __efc_domain_init,
|
|
NULL);
|
|
}
|
|
efc_domain_post_event(domain, EFC_EVT_DOMAIN_FOUND, drec);
|
|
break;
|
|
}
|
|
|
|
case EFC_HW_DOMAIN_LOST:
|
|
domain_trace(domain, "EFC_HW_DOMAIN_LOST:\n");
|
|
efc->hold_frames = true;
|
|
efc_domain_post_event(domain, EFC_EVT_DOMAIN_LOST, NULL);
|
|
break;
|
|
|
|
case EFC_HW_DOMAIN_ALLOC_OK:
|
|
domain_trace(domain, "EFC_HW_DOMAIN_ALLOC_OK:\n");
|
|
efc_domain_post_event(domain, EFC_EVT_DOMAIN_ALLOC_OK, NULL);
|
|
break;
|
|
|
|
case EFC_HW_DOMAIN_ALLOC_FAIL:
|
|
domain_trace(domain, "EFC_HW_DOMAIN_ALLOC_FAIL:\n");
|
|
efc_domain_post_event(domain, EFC_EVT_DOMAIN_ALLOC_FAIL,
|
|
NULL);
|
|
break;
|
|
|
|
case EFC_HW_DOMAIN_ATTACH_OK:
|
|
domain_trace(domain, "EFC_HW_DOMAIN_ATTACH_OK:\n");
|
|
efc_domain_post_event(domain, EFC_EVT_DOMAIN_ATTACH_OK, NULL);
|
|
break;
|
|
|
|
case EFC_HW_DOMAIN_ATTACH_FAIL:
|
|
domain_trace(domain, "EFC_HW_DOMAIN_ATTACH_FAIL:\n");
|
|
efc_domain_post_event(domain,
|
|
EFC_EVT_DOMAIN_ATTACH_FAIL, NULL);
|
|
break;
|
|
|
|
case EFC_HW_DOMAIN_FREE_OK:
|
|
domain_trace(domain, "EFC_HW_DOMAIN_FREE_OK:\n");
|
|
efc_domain_post_event(domain, EFC_EVT_DOMAIN_FREE_OK, NULL);
|
|
break;
|
|
|
|
case EFC_HW_DOMAIN_FREE_FAIL:
|
|
domain_trace(domain, "EFC_HW_DOMAIN_FREE_FAIL:\n");
|
|
efc_domain_post_event(domain, EFC_EVT_DOMAIN_FREE_FAIL, NULL);
|
|
break;
|
|
|
|
default:
|
|
efc_log_warn(efc, "unsupported event %#x\n", event);
|
|
}
|
|
spin_unlock_irqrestore(&efc->lock, flags);
|
|
|
|
if (efc->domain && domain->req_accept_frames) {
|
|
domain->req_accept_frames = false;
|
|
efc->hold_frames = false;
|
|
}
|
|
|
|
return rc;
|
|
}
|
|
|
|
static void
|
|
_efc_domain_free(struct kref *arg)
|
|
{
|
|
struct efc_domain *domain = container_of(arg, struct efc_domain, ref);
|
|
struct efc *efc = domain->efc;
|
|
|
|
if (efc->domain_free_cb)
|
|
(*efc->domain_free_cb)(efc, efc->domain_free_cb_arg);
|
|
|
|
kfree(domain);
|
|
}
|
|
|
|
void
|
|
efc_domain_free(struct efc_domain *domain)
|
|
{
|
|
struct efc *efc;
|
|
|
|
efc = domain->efc;
|
|
|
|
/* Hold frames to clear the domain pointer from the xport lookup */
|
|
efc->hold_frames = false;
|
|
|
|
efc_log_debug(efc, "Domain free: wwn %016llX\n", domain->fcf_wwn);
|
|
|
|
xa_destroy(&domain->lookup);
|
|
efc->domain = NULL;
|
|
kref_put(&domain->ref, domain->release);
|
|
}
|
|
|
|
struct efc_domain *
|
|
efc_domain_alloc(struct efc *efc, uint64_t fcf_wwn)
|
|
{
|
|
struct efc_domain *domain;
|
|
|
|
domain = kzalloc(sizeof(*domain), GFP_ATOMIC);
|
|
if (!domain)
|
|
return NULL;
|
|
|
|
domain->efc = efc;
|
|
domain->drvsm.app = domain;
|
|
|
|
/* initialize refcount */
|
|
kref_init(&domain->ref);
|
|
domain->release = _efc_domain_free;
|
|
|
|
xa_init(&domain->lookup);
|
|
|
|
INIT_LIST_HEAD(&domain->nport_list);
|
|
efc->domain = domain;
|
|
domain->fcf_wwn = fcf_wwn;
|
|
efc_log_debug(efc, "Domain allocated: wwn %016llX\n", domain->fcf_wwn);
|
|
|
|
return domain;
|
|
}
|
|
|
|
void
|
|
efc_register_domain_free_cb(struct efc *efc,
|
|
void (*callback)(struct efc *efc, void *arg),
|
|
void *arg)
|
|
{
|
|
/* Register a callback to be called when the domain is freed */
|
|
efc->domain_free_cb = callback;
|
|
efc->domain_free_cb_arg = arg;
|
|
if (!efc->domain && callback)
|
|
(*callback)(efc, arg);
|
|
}
|
|
|
|
static void
|
|
__efc_domain_common(const char *funcname, struct efc_sm_ctx *ctx,
|
|
enum efc_sm_event evt, void *arg)
|
|
{
|
|
struct efc_domain *domain = ctx->app;
|
|
|
|
switch (evt) {
|
|
case EFC_EVT_ENTER:
|
|
case EFC_EVT_REENTER:
|
|
case EFC_EVT_EXIT:
|
|
case EFC_EVT_ALL_CHILD_NODES_FREE:
|
|
/*
|
|
* this can arise if an FLOGI fails on the NPORT,
|
|
* and the NPORT is shutdown
|
|
*/
|
|
break;
|
|
default:
|
|
efc_log_warn(domain->efc, "%-20s %-20s not handled\n",
|
|
funcname, efc_sm_event_name(evt));
|
|
}
|
|
}
|
|
|
|
static void
|
|
__efc_domain_common_shutdown(const char *funcname, struct efc_sm_ctx *ctx,
|
|
enum efc_sm_event evt, void *arg)
|
|
{
|
|
struct efc_domain *domain = ctx->app;
|
|
|
|
switch (evt) {
|
|
case EFC_EVT_ENTER:
|
|
case EFC_EVT_REENTER:
|
|
case EFC_EVT_EXIT:
|
|
break;
|
|
case EFC_EVT_DOMAIN_FOUND:
|
|
/* save drec, mark domain_found_pending */
|
|
memcpy(&domain->pending_drec, arg,
|
|
sizeof(domain->pending_drec));
|
|
domain->domain_found_pending = true;
|
|
break;
|
|
case EFC_EVT_DOMAIN_LOST:
|
|
/* unmark domain_found_pending */
|
|
domain->domain_found_pending = false;
|
|
break;
|
|
|
|
default:
|
|
efc_log_warn(domain->efc, "%-20s %-20s not handled\n",
|
|
funcname, efc_sm_event_name(evt));
|
|
}
|
|
}
|
|
|
|
#define std_domain_state_decl(...)\
|
|
struct efc_domain *domain = NULL;\
|
|
struct efc *efc = NULL;\
|
|
\
|
|
WARN_ON(!ctx || !ctx->app);\
|
|
domain = ctx->app;\
|
|
WARN_ON(!domain->efc);\
|
|
efc = domain->efc
|
|
|
|
void
|
|
__efc_domain_init(struct efc_sm_ctx *ctx, enum efc_sm_event evt,
|
|
void *arg)
|
|
{
|
|
std_domain_state_decl();
|
|
|
|
domain_sm_trace(domain);
|
|
|
|
switch (evt) {
|
|
case EFC_EVT_ENTER:
|
|
domain->attached = false;
|
|
break;
|
|
|
|
case EFC_EVT_DOMAIN_FOUND: {
|
|
u32 i;
|
|
struct efc_domain_record *drec = arg;
|
|
struct efc_nport *nport;
|
|
|
|
u64 my_wwnn = efc->req_wwnn;
|
|
u64 my_wwpn = efc->req_wwpn;
|
|
__be64 bewwpn;
|
|
|
|
if (my_wwpn == 0 || my_wwnn == 0) {
|
|
efc_log_debug(efc, "using default hardware WWN config\n");
|
|
my_wwpn = efc->def_wwpn;
|
|
my_wwnn = efc->def_wwnn;
|
|
}
|
|
|
|
efc_log_debug(efc, "Create nport WWPN %016llX WWNN %016llX\n",
|
|
my_wwpn, my_wwnn);
|
|
|
|
/* Allocate a nport and transition to __efc_nport_allocated */
|
|
nport = efc_nport_alloc(domain, my_wwpn, my_wwnn, U32_MAX,
|
|
efc->enable_ini, efc->enable_tgt);
|
|
|
|
if (!nport) {
|
|
efc_log_err(efc, "efc_nport_alloc() failed\n");
|
|
break;
|
|
}
|
|
efc_sm_transition(&nport->sm, __efc_nport_allocated, NULL);
|
|
|
|
bewwpn = cpu_to_be64(nport->wwpn);
|
|
|
|
/* allocate struct efc_nport object for local port
|
|
* Note: drec->fc_id is ALPA from read_topology only if loop
|
|
*/
|
|
if (efc_cmd_nport_alloc(efc, nport, NULL, (uint8_t *)&bewwpn)) {
|
|
efc_log_err(efc, "Can't allocate port\n");
|
|
efc_nport_free(nport);
|
|
break;
|
|
}
|
|
|
|
domain->is_loop = drec->is_loop;
|
|
|
|
/*
|
|
* If the loop position map includes ALPA == 0,
|
|
* then we are in a public loop (NL_PORT)
|
|
* Note that the first element of the loopmap[]
|
|
* contains the count of elements, and if
|
|
* ALPA == 0 is present, it will occupy the first
|
|
* location after the count.
|
|
*/
|
|
domain->is_nlport = drec->map.loop[1] == 0x00;
|
|
|
|
if (!domain->is_loop) {
|
|
/* Initiate HW domain alloc */
|
|
if (efc_cmd_domain_alloc(efc, domain, drec->index)) {
|
|
efc_log_err(efc,
|
|
"Failed to initiate HW domain allocation\n");
|
|
break;
|
|
}
|
|
efc_sm_transition(ctx, __efc_domain_wait_alloc, arg);
|
|
break;
|
|
}
|
|
|
|
efc_log_debug(efc, "%s fc_id=%#x speed=%d\n",
|
|
drec->is_loop ?
|
|
(domain->is_nlport ?
|
|
"public-loop" : "loop") : "other",
|
|
drec->fc_id, drec->speed);
|
|
|
|
nport->fc_id = drec->fc_id;
|
|
nport->topology = EFC_NPORT_TOPO_FC_AL;
|
|
snprintf(nport->display_name, sizeof(nport->display_name),
|
|
"s%06x", drec->fc_id);
|
|
|
|
if (efc->enable_ini) {
|
|
u32 count = drec->map.loop[0];
|
|
|
|
efc_log_debug(efc, "%d position map entries\n",
|
|
count);
|
|
for (i = 1; i <= count; i++) {
|
|
if (drec->map.loop[i] != drec->fc_id) {
|
|
struct efc_node *node;
|
|
|
|
efc_log_debug(efc, "%#x -> %#x\n",
|
|
drec->fc_id,
|
|
drec->map.loop[i]);
|
|
node = efc_node_alloc(nport,
|
|
drec->map.loop[i],
|
|
false, true);
|
|
if (!node) {
|
|
efc_log_err(efc,
|
|
"efc_node_alloc() failed\n");
|
|
break;
|
|
}
|
|
efc_node_transition(node,
|
|
__efc_d_wait_loop,
|
|
NULL);
|
|
}
|
|
}
|
|
}
|
|
|
|
/* Initiate HW domain alloc */
|
|
if (efc_cmd_domain_alloc(efc, domain, drec->index)) {
|
|
efc_log_err(efc,
|
|
"Failed to initiate HW domain allocation\n");
|
|
break;
|
|
}
|
|
efc_sm_transition(ctx, __efc_domain_wait_alloc, arg);
|
|
break;
|
|
}
|
|
default:
|
|
__efc_domain_common(__func__, ctx, evt, arg);
|
|
}
|
|
}
|
|
|
|
void
|
|
__efc_domain_wait_alloc(struct efc_sm_ctx *ctx,
|
|
enum efc_sm_event evt, void *arg)
|
|
{
|
|
std_domain_state_decl();
|
|
|
|
domain_sm_trace(domain);
|
|
|
|
switch (evt) {
|
|
case EFC_EVT_DOMAIN_ALLOC_OK: {
|
|
struct fc_els_flogi *sp;
|
|
struct efc_nport *nport;
|
|
|
|
nport = domain->nport;
|
|
if (WARN_ON(!nport))
|
|
return;
|
|
|
|
sp = (struct fc_els_flogi *)nport->service_params;
|
|
|
|
/* Save the domain service parameters */
|
|
memcpy(domain->service_params + 4, domain->dma.virt,
|
|
sizeof(struct fc_els_flogi) - 4);
|
|
memcpy(nport->service_params + 4, domain->dma.virt,
|
|
sizeof(struct fc_els_flogi) - 4);
|
|
|
|
/*
|
|
* Update the nport's service parameters,
|
|
* user might have specified non-default names
|
|
*/
|
|
sp->fl_wwpn = cpu_to_be64(nport->wwpn);
|
|
sp->fl_wwnn = cpu_to_be64(nport->wwnn);
|
|
|
|
/*
|
|
* Take the loop topology path,
|
|
* unless we are an NL_PORT (public loop)
|
|
*/
|
|
if (domain->is_loop && !domain->is_nlport) {
|
|
/*
|
|
* For loop, we already have our FC ID
|
|
* and don't need fabric login.
|
|
* Transition to the allocated state and
|
|
* post an event to attach to
|
|
* the domain. Note that this breaks the
|
|
* normal action/transition
|
|
* pattern here to avoid a race with the
|
|
* domain attach callback.
|
|
*/
|
|
/* sm: is_loop / domain_attach */
|
|
efc_sm_transition(ctx, __efc_domain_allocated, NULL);
|
|
__efc_domain_attach_internal(domain, nport->fc_id);
|
|
break;
|
|
}
|
|
{
|
|
struct efc_node *node;
|
|
|
|
/* alloc fabric node, send FLOGI */
|
|
node = efc_node_find(nport, FC_FID_FLOGI);
|
|
if (node) {
|
|
efc_log_err(efc,
|
|
"Fabric Controller node already exists\n");
|
|
break;
|
|
}
|
|
node = efc_node_alloc(nport, FC_FID_FLOGI,
|
|
false, false);
|
|
if (!node) {
|
|
efc_log_err(efc,
|
|
"Error: efc_node_alloc() failed\n");
|
|
} else {
|
|
efc_node_transition(node,
|
|
__efc_fabric_init, NULL);
|
|
}
|
|
/* Accept frames */
|
|
domain->req_accept_frames = true;
|
|
}
|
|
/* sm: / start fabric logins */
|
|
efc_sm_transition(ctx, __efc_domain_allocated, NULL);
|
|
break;
|
|
}
|
|
|
|
case EFC_EVT_DOMAIN_ALLOC_FAIL:
|
|
efc_log_err(efc, "%s recv'd waiting for DOMAIN_ALLOC_OK;",
|
|
efc_sm_event_name(evt));
|
|
efc_log_err(efc, "shutting down domain\n");
|
|
domain->req_domain_free = true;
|
|
break;
|
|
|
|
case EFC_EVT_DOMAIN_FOUND:
|
|
/* Should not happen */
|
|
break;
|
|
|
|
case EFC_EVT_DOMAIN_LOST:
|
|
efc_log_debug(efc,
|
|
"%s received while waiting for hw_domain_alloc()\n",
|
|
efc_sm_event_name(evt));
|
|
efc_sm_transition(ctx, __efc_domain_wait_domain_lost, NULL);
|
|
break;
|
|
|
|
default:
|
|
__efc_domain_common(__func__, ctx, evt, arg);
|
|
}
|
|
}
|
|
|
|
void
|
|
__efc_domain_allocated(struct efc_sm_ctx *ctx,
|
|
enum efc_sm_event evt, void *arg)
|
|
{
|
|
std_domain_state_decl();
|
|
|
|
domain_sm_trace(domain);
|
|
|
|
switch (evt) {
|
|
case EFC_EVT_DOMAIN_REQ_ATTACH: {
|
|
int rc = 0;
|
|
u32 fc_id;
|
|
|
|
if (WARN_ON(!arg))
|
|
return;
|
|
|
|
fc_id = *((u32 *)arg);
|
|
efc_log_debug(efc, "Requesting hw domain attach fc_id x%x\n",
|
|
fc_id);
|
|
/* Update nport lookup */
|
|
rc = xa_err(xa_store(&domain->lookup, fc_id, domain->nport,
|
|
GFP_ATOMIC));
|
|
if (rc) {
|
|
efc_log_err(efc, "Sport lookup store failed: %d\n", rc);
|
|
return;
|
|
}
|
|
|
|
/* Update display name for the nport */
|
|
efc_node_fcid_display(fc_id, domain->nport->display_name,
|
|
sizeof(domain->nport->display_name));
|
|
|
|
/* Issue domain attach call */
|
|
rc = efc_cmd_domain_attach(efc, domain, fc_id);
|
|
if (rc) {
|
|
efc_log_err(efc, "efc_hw_domain_attach failed: %d\n",
|
|
rc);
|
|
return;
|
|
}
|
|
/* sm: / domain_attach */
|
|
efc_sm_transition(ctx, __efc_domain_wait_attach, NULL);
|
|
break;
|
|
}
|
|
|
|
case EFC_EVT_DOMAIN_FOUND:
|
|
/* Should not happen */
|
|
efc_log_err(efc, "%s: evt: %d should not happen\n",
|
|
__func__, evt);
|
|
break;
|
|
|
|
case EFC_EVT_DOMAIN_LOST: {
|
|
efc_log_debug(efc,
|
|
"%s received while in EFC_EVT_DOMAIN_REQ_ATTACH\n",
|
|
efc_sm_event_name(evt));
|
|
if (!list_empty(&domain->nport_list)) {
|
|
/*
|
|
* if there are nports, transition to
|
|
* wait state and send shutdown to each
|
|
* nport
|
|
*/
|
|
struct efc_nport *nport = NULL, *nport_next = NULL;
|
|
|
|
efc_sm_transition(ctx, __efc_domain_wait_nports_free,
|
|
NULL);
|
|
list_for_each_entry_safe(nport, nport_next,
|
|
&domain->nport_list,
|
|
list_entry) {
|
|
efc_sm_post_event(&nport->sm,
|
|
EFC_EVT_SHUTDOWN, NULL);
|
|
}
|
|
} else {
|
|
/* no nports exist, free domain */
|
|
efc_sm_transition(ctx, __efc_domain_wait_shutdown,
|
|
NULL);
|
|
if (efc_cmd_domain_free(efc, domain))
|
|
efc_log_err(efc, "hw_domain_free failed\n");
|
|
}
|
|
|
|
break;
|
|
}
|
|
|
|
default:
|
|
__efc_domain_common(__func__, ctx, evt, arg);
|
|
}
|
|
}
|
|
|
|
void
|
|
__efc_domain_wait_attach(struct efc_sm_ctx *ctx,
|
|
enum efc_sm_event evt, void *arg)
|
|
{
|
|
std_domain_state_decl();
|
|
|
|
domain_sm_trace(domain);
|
|
|
|
switch (evt) {
|
|
case EFC_EVT_DOMAIN_ATTACH_OK: {
|
|
struct efc_node *node = NULL;
|
|
struct efc_nport *nport, *next_nport;
|
|
unsigned long index;
|
|
|
|
/*
|
|
* Set domain notify pending state to avoid
|
|
* duplicate domain event post
|
|
*/
|
|
domain->domain_notify_pend = true;
|
|
|
|
/* Mark as attached */
|
|
domain->attached = true;
|
|
|
|
/* Transition to ready */
|
|
/* sm: / forward event to all nports and nodes */
|
|
efc_sm_transition(ctx, __efc_domain_ready, NULL);
|
|
|
|
/* We have an FCFI, so we can accept frames */
|
|
domain->req_accept_frames = true;
|
|
|
|
/*
|
|
* Notify all nodes that the domain attach request
|
|
* has completed
|
|
* Note: nport will have already received notification
|
|
* of nport attached as a result of the HW's port attach.
|
|
*/
|
|
list_for_each_entry_safe(nport, next_nport,
|
|
&domain->nport_list, list_entry) {
|
|
xa_for_each(&nport->lookup, index, node) {
|
|
efc_node_post_event(node,
|
|
EFC_EVT_DOMAIN_ATTACH_OK,
|
|
NULL);
|
|
}
|
|
}
|
|
domain->domain_notify_pend = false;
|
|
break;
|
|
}
|
|
|
|
case EFC_EVT_DOMAIN_ATTACH_FAIL:
|
|
efc_log_debug(efc,
|
|
"%s received while waiting for hw attach\n",
|
|
efc_sm_event_name(evt));
|
|
break;
|
|
|
|
case EFC_EVT_DOMAIN_FOUND:
|
|
/* Should not happen */
|
|
efc_log_err(efc, "%s: evt: %d should not happen\n",
|
|
__func__, evt);
|
|
break;
|
|
|
|
case EFC_EVT_DOMAIN_LOST:
|
|
/*
|
|
* Domain lost while waiting for an attach to complete,
|
|
* go to a state that waits for the domain attach to
|
|
* complete, then handle domain lost
|
|
*/
|
|
efc_sm_transition(ctx, __efc_domain_wait_domain_lost, NULL);
|
|
break;
|
|
|
|
case EFC_EVT_DOMAIN_REQ_ATTACH:
|
|
/*
|
|
* In P2P we can get an attach request from
|
|
* the other FLOGI path, so drop this one
|
|
*/
|
|
break;
|
|
|
|
default:
|
|
__efc_domain_common(__func__, ctx, evt, arg);
|
|
}
|
|
}
|
|
|
|
void
|
|
__efc_domain_ready(struct efc_sm_ctx *ctx, enum efc_sm_event evt, void *arg)
|
|
{
|
|
std_domain_state_decl();
|
|
|
|
domain_sm_trace(domain);
|
|
|
|
switch (evt) {
|
|
case EFC_EVT_ENTER: {
|
|
/* start any pending vports */
|
|
if (efc_vport_start(domain)) {
|
|
efc_log_debug(domain->efc,
|
|
"efc_vport_start didn't start vports\n");
|
|
}
|
|
break;
|
|
}
|
|
case EFC_EVT_DOMAIN_LOST: {
|
|
if (!list_empty(&domain->nport_list)) {
|
|
/*
|
|
* if there are nports, transition to wait state
|
|
* and send shutdown to each nport
|
|
*/
|
|
struct efc_nport *nport = NULL, *nport_next = NULL;
|
|
|
|
efc_sm_transition(ctx, __efc_domain_wait_nports_free,
|
|
NULL);
|
|
list_for_each_entry_safe(nport, nport_next,
|
|
&domain->nport_list,
|
|
list_entry) {
|
|
efc_sm_post_event(&nport->sm,
|
|
EFC_EVT_SHUTDOWN, NULL);
|
|
}
|
|
} else {
|
|
/* no nports exist, free domain */
|
|
efc_sm_transition(ctx, __efc_domain_wait_shutdown,
|
|
NULL);
|
|
if (efc_cmd_domain_free(efc, domain))
|
|
efc_log_err(efc, "hw_domain_free failed\n");
|
|
}
|
|
break;
|
|
}
|
|
|
|
case EFC_EVT_DOMAIN_FOUND:
|
|
/* Should not happen */
|
|
efc_log_err(efc, "%s: evt: %d should not happen\n",
|
|
__func__, evt);
|
|
break;
|
|
|
|
case EFC_EVT_DOMAIN_REQ_ATTACH: {
|
|
/* can happen during p2p */
|
|
u32 fc_id;
|
|
|
|
fc_id = *((u32 *)arg);
|
|
|
|
/* Assume that the domain is attached */
|
|
WARN_ON(!domain->attached);
|
|
|
|
/*
|
|
* Verify that the requested FC_ID
|
|
* is the same as the one we're working with
|
|
*/
|
|
WARN_ON(domain->nport->fc_id != fc_id);
|
|
break;
|
|
}
|
|
|
|
default:
|
|
__efc_domain_common(__func__, ctx, evt, arg);
|
|
}
|
|
}
|
|
|
|
void
|
|
__efc_domain_wait_nports_free(struct efc_sm_ctx *ctx, enum efc_sm_event evt,
|
|
void *arg)
|
|
{
|
|
std_domain_state_decl();
|
|
|
|
domain_sm_trace(domain);
|
|
|
|
/* Wait for nodes to free prior to the domain shutdown */
|
|
switch (evt) {
|
|
case EFC_EVT_ALL_CHILD_NODES_FREE: {
|
|
int rc;
|
|
|
|
/* sm: / efc_hw_domain_free */
|
|
efc_sm_transition(ctx, __efc_domain_wait_shutdown, NULL);
|
|
|
|
/* Request efc_hw_domain_free and wait for completion */
|
|
rc = efc_cmd_domain_free(efc, domain);
|
|
if (rc) {
|
|
efc_log_err(efc, "efc_hw_domain_free() failed: %d\n",
|
|
rc);
|
|
}
|
|
break;
|
|
}
|
|
default:
|
|
__efc_domain_common_shutdown(__func__, ctx, evt, arg);
|
|
}
|
|
}
|
|
|
|
void
|
|
__efc_domain_wait_shutdown(struct efc_sm_ctx *ctx,
|
|
enum efc_sm_event evt, void *arg)
|
|
{
|
|
std_domain_state_decl();
|
|
|
|
domain_sm_trace(domain);
|
|
|
|
switch (evt) {
|
|
case EFC_EVT_DOMAIN_FREE_OK:
|
|
/* sm: / domain_free */
|
|
if (domain->domain_found_pending) {
|
|
/*
|
|
* save fcf_wwn and drec from this domain,
|
|
* free current domain and allocate
|
|
* a new one with the same fcf_wwn
|
|
* could use a SLI-4 "re-register VPI"
|
|
* operation here?
|
|
*/
|
|
u64 fcf_wwn = domain->fcf_wwn;
|
|
struct efc_domain_record drec = domain->pending_drec;
|
|
|
|
efc_log_debug(efc, "Reallocating domain\n");
|
|
domain->req_domain_free = true;
|
|
domain = efc_domain_alloc(efc, fcf_wwn);
|
|
|
|
if (!domain) {
|
|
efc_log_err(efc,
|
|
"efc_domain_alloc() failed\n");
|
|
return;
|
|
}
|
|
/*
|
|
* got a new domain; at this point,
|
|
* there are at least two domains
|
|
* once the req_domain_free flag is processed,
|
|
* the associated domain will be removed.
|
|
*/
|
|
efc_sm_transition(&domain->drvsm, __efc_domain_init,
|
|
NULL);
|
|
efc_sm_post_event(&domain->drvsm,
|
|
EFC_EVT_DOMAIN_FOUND, &drec);
|
|
} else {
|
|
domain->req_domain_free = true;
|
|
}
|
|
break;
|
|
default:
|
|
__efc_domain_common_shutdown(__func__, ctx, evt, arg);
|
|
}
|
|
}
|
|
|
|
void
|
|
__efc_domain_wait_domain_lost(struct efc_sm_ctx *ctx,
|
|
enum efc_sm_event evt, void *arg)
|
|
{
|
|
std_domain_state_decl();
|
|
|
|
domain_sm_trace(domain);
|
|
|
|
/*
|
|
* Wait for the domain alloc/attach completion
|
|
* after receiving a domain lost.
|
|
*/
|
|
switch (evt) {
|
|
case EFC_EVT_DOMAIN_ALLOC_OK:
|
|
case EFC_EVT_DOMAIN_ATTACH_OK: {
|
|
if (!list_empty(&domain->nport_list)) {
|
|
/*
|
|
* if there are nports, transition to
|
|
* wait state and send shutdown to each nport
|
|
*/
|
|
struct efc_nport *nport = NULL, *nport_next = NULL;
|
|
|
|
efc_sm_transition(ctx, __efc_domain_wait_nports_free,
|
|
NULL);
|
|
list_for_each_entry_safe(nport, nport_next,
|
|
&domain->nport_list,
|
|
list_entry) {
|
|
efc_sm_post_event(&nport->sm,
|
|
EFC_EVT_SHUTDOWN, NULL);
|
|
}
|
|
} else {
|
|
/* no nports exist, free domain */
|
|
efc_sm_transition(ctx, __efc_domain_wait_shutdown,
|
|
NULL);
|
|
if (efc_cmd_domain_free(efc, domain))
|
|
efc_log_err(efc, "hw_domain_free() failed\n");
|
|
}
|
|
break;
|
|
}
|
|
case EFC_EVT_DOMAIN_ALLOC_FAIL:
|
|
case EFC_EVT_DOMAIN_ATTACH_FAIL:
|
|
efc_log_err(efc, "[domain] %-20s: failed\n",
|
|
efc_sm_event_name(evt));
|
|
break;
|
|
|
|
default:
|
|
__efc_domain_common_shutdown(__func__, ctx, evt, arg);
|
|
}
|
|
}
|
|
|
|
void
|
|
__efc_domain_attach_internal(struct efc_domain *domain, u32 s_id)
|
|
{
|
|
memcpy(domain->dma.virt,
|
|
((uint8_t *)domain->flogi_service_params) + 4,
|
|
sizeof(struct fc_els_flogi) - 4);
|
|
(void)efc_sm_post_event(&domain->drvsm, EFC_EVT_DOMAIN_REQ_ATTACH,
|
|
&s_id);
|
|
}
|
|
|
|
void
|
|
efc_domain_attach(struct efc_domain *domain, u32 s_id)
|
|
{
|
|
__efc_domain_attach_internal(domain, s_id);
|
|
}
|
|
|
|
int
|
|
efc_domain_post_event(struct efc_domain *domain,
|
|
enum efc_sm_event event, void *arg)
|
|
{
|
|
int rc;
|
|
bool req_domain_free;
|
|
|
|
rc = efc_sm_post_event(&domain->drvsm, event, arg);
|
|
|
|
req_domain_free = domain->req_domain_free;
|
|
domain->req_domain_free = false;
|
|
|
|
if (req_domain_free)
|
|
efc_domain_free(domain);
|
|
|
|
return rc;
|
|
}
|
|
|
|
static void
|
|
efct_domain_process_pending(struct efc_domain *domain)
|
|
{
|
|
struct efc *efc = domain->efc;
|
|
struct efc_hw_sequence *seq = NULL;
|
|
u32 processed = 0;
|
|
unsigned long flags = 0;
|
|
|
|
for (;;) {
|
|
/* need to check for hold frames condition after each frame
|
|
* processed because any given frame could cause a transition
|
|
* to a state that holds frames
|
|
*/
|
|
if (efc->hold_frames)
|
|
break;
|
|
|
|
/* Get next frame/sequence */
|
|
spin_lock_irqsave(&efc->pend_frames_lock, flags);
|
|
|
|
if (!list_empty(&efc->pend_frames)) {
|
|
seq = list_first_entry(&efc->pend_frames,
|
|
struct efc_hw_sequence, list_entry);
|
|
list_del(&seq->list_entry);
|
|
}
|
|
|
|
if (!seq) {
|
|
processed = efc->pend_frames_processed;
|
|
efc->pend_frames_processed = 0;
|
|
spin_unlock_irqrestore(&efc->pend_frames_lock, flags);
|
|
break;
|
|
}
|
|
efc->pend_frames_processed++;
|
|
|
|
spin_unlock_irqrestore(&efc->pend_frames_lock, flags);
|
|
|
|
/* now dispatch frame(s) to dispatch function */
|
|
if (efc_domain_dispatch_frame(domain, seq))
|
|
efc->tt.hw_seq_free(efc, seq);
|
|
|
|
seq = NULL;
|
|
}
|
|
|
|
if (processed != 0)
|
|
efc_log_debug(efc, "%u domain frames held and processed\n",
|
|
processed);
|
|
}
|
|
|
|
void
|
|
efc_dispatch_frame(struct efc *efc, struct efc_hw_sequence *seq)
|
|
{
|
|
struct efc_domain *domain = efc->domain;
|
|
|
|
/*
|
|
* If we are holding frames or the domain is not yet registered or
|
|
* there's already frames on the pending list,
|
|
* then add the new frame to pending list
|
|
*/
|
|
if (!domain || efc->hold_frames || !list_empty(&efc->pend_frames)) {
|
|
unsigned long flags = 0;
|
|
|
|
spin_lock_irqsave(&efc->pend_frames_lock, flags);
|
|
INIT_LIST_HEAD(&seq->list_entry);
|
|
list_add_tail(&seq->list_entry, &efc->pend_frames);
|
|
spin_unlock_irqrestore(&efc->pend_frames_lock, flags);
|
|
|
|
if (domain) {
|
|
/* immediately process pending frames */
|
|
efct_domain_process_pending(domain);
|
|
}
|
|
} else {
|
|
/*
|
|
* We are not holding frames and pending list is empty,
|
|
* just process frame. A non-zero return means the frame
|
|
* was not handled - so cleanup
|
|
*/
|
|
if (efc_domain_dispatch_frame(domain, seq))
|
|
efc->tt.hw_seq_free(efc, seq);
|
|
}
|
|
}
|
|
|
|
int
|
|
efc_domain_dispatch_frame(void *arg, struct efc_hw_sequence *seq)
|
|
{
|
|
struct efc_domain *domain = (struct efc_domain *)arg;
|
|
struct efc *efc = domain->efc;
|
|
struct fc_frame_header *hdr;
|
|
struct efc_node *node = NULL;
|
|
struct efc_nport *nport = NULL;
|
|
unsigned long flags = 0;
|
|
u32 s_id, d_id, rc = EFC_HW_SEQ_FREE;
|
|
|
|
if (!seq->header || !seq->header->dma.virt || !seq->payload->dma.virt) {
|
|
efc_log_err(efc, "Sequence header or payload is null\n");
|
|
return rc;
|
|
}
|
|
|
|
hdr = seq->header->dma.virt;
|
|
|
|
/* extract the s_id and d_id */
|
|
s_id = ntoh24(hdr->fh_s_id);
|
|
d_id = ntoh24(hdr->fh_d_id);
|
|
|
|
spin_lock_irqsave(&efc->lock, flags);
|
|
|
|
nport = efc_nport_find(domain, d_id);
|
|
if (!nport) {
|
|
if (hdr->fh_type == FC_TYPE_FCP) {
|
|
/* Drop frame */
|
|
efc_log_warn(efc, "FCP frame with invalid d_id x%x\n",
|
|
d_id);
|
|
goto out;
|
|
}
|
|
|
|
/* p2p will use this case */
|
|
nport = domain->nport;
|
|
if (!nport || !kref_get_unless_zero(&nport->ref)) {
|
|
efc_log_err(efc, "Physical nport is NULL\n");
|
|
goto out;
|
|
}
|
|
}
|
|
|
|
/* Lookup the node given the remote s_id */
|
|
node = efc_node_find(nport, s_id);
|
|
|
|
/* If not found, then create a new node */
|
|
if (!node) {
|
|
/*
|
|
* If this is solicited data or control based on R_CTL and
|
|
* there is no node context, then we can drop the frame
|
|
*/
|
|
if ((hdr->fh_r_ctl == FC_RCTL_DD_SOL_DATA) ||
|
|
(hdr->fh_r_ctl == FC_RCTL_DD_SOL_CTL)) {
|
|
efc_log_debug(efc, "sol data/ctrl frame without node\n");
|
|
goto out_release;
|
|
}
|
|
|
|
node = efc_node_alloc(nport, s_id, false, false);
|
|
if (!node) {
|
|
efc_log_err(efc, "efc_node_alloc() failed\n");
|
|
goto out_release;
|
|
}
|
|
/* don't send PLOGI on efc_d_init entry */
|
|
efc_node_init_device(node, false);
|
|
}
|
|
|
|
if (node->hold_frames || !list_empty(&node->pend_frames)) {
|
|
/* add frame to node's pending list */
|
|
spin_lock(&node->pend_frames_lock);
|
|
INIT_LIST_HEAD(&seq->list_entry);
|
|
list_add_tail(&seq->list_entry, &node->pend_frames);
|
|
spin_unlock(&node->pend_frames_lock);
|
|
rc = EFC_HW_SEQ_HOLD;
|
|
goto out_release;
|
|
}
|
|
|
|
/* now dispatch frame to the node frame handler */
|
|
efc_node_dispatch_frame(node, seq);
|
|
|
|
out_release:
|
|
kref_put(&nport->ref, nport->release);
|
|
out:
|
|
spin_unlock_irqrestore(&efc->lock, flags);
|
|
return rc;
|
|
}
|
|
|
|
void
|
|
efc_node_dispatch_frame(void *arg, struct efc_hw_sequence *seq)
|
|
{
|
|
struct fc_frame_header *hdr = seq->header->dma.virt;
|
|
u32 port_id;
|
|
struct efc_node *node = (struct efc_node *)arg;
|
|
struct efc *efc = node->efc;
|
|
|
|
port_id = ntoh24(hdr->fh_s_id);
|
|
|
|
if (WARN_ON(port_id != node->rnode.fc_id))
|
|
return;
|
|
|
|
if ((!(ntoh24(hdr->fh_f_ctl) & FC_FC_END_SEQ)) ||
|
|
!(ntoh24(hdr->fh_f_ctl) & FC_FC_SEQ_INIT)) {
|
|
node_printf(node,
|
|
"Drop frame hdr = %08x %08x %08x %08x %08x %08x\n",
|
|
cpu_to_be32(((u32 *)hdr)[0]),
|
|
cpu_to_be32(((u32 *)hdr)[1]),
|
|
cpu_to_be32(((u32 *)hdr)[2]),
|
|
cpu_to_be32(((u32 *)hdr)[3]),
|
|
cpu_to_be32(((u32 *)hdr)[4]),
|
|
cpu_to_be32(((u32 *)hdr)[5]));
|
|
return;
|
|
}
|
|
|
|
switch (hdr->fh_r_ctl) {
|
|
case FC_RCTL_ELS_REQ:
|
|
case FC_RCTL_ELS_REP:
|
|
efc_node_recv_els_frame(node, seq);
|
|
break;
|
|
|
|
case FC_RCTL_BA_ABTS:
|
|
case FC_RCTL_BA_ACC:
|
|
case FC_RCTL_BA_RJT:
|
|
case FC_RCTL_BA_NOP:
|
|
efc_log_err(efc, "Received ABTS:\n");
|
|
break;
|
|
|
|
case FC_RCTL_DD_UNSOL_CMD:
|
|
case FC_RCTL_DD_UNSOL_CTL:
|
|
switch (hdr->fh_type) {
|
|
case FC_TYPE_FCP:
|
|
if ((hdr->fh_r_ctl & 0xf) == FC_RCTL_DD_UNSOL_CMD) {
|
|
if (!node->fcp_enabled) {
|
|
efc_node_recv_fcp_cmd(node, seq);
|
|
break;
|
|
}
|
|
efc_log_err(efc, "Recvd FCP CMD. Drop IO\n");
|
|
} else if ((hdr->fh_r_ctl & 0xf) ==
|
|
FC_RCTL_DD_SOL_DATA) {
|
|
node_printf(node,
|
|
"solicited data recvd. Drop IO\n");
|
|
}
|
|
break;
|
|
|
|
case FC_TYPE_CT:
|
|
efc_node_recv_ct_frame(node, seq);
|
|
break;
|
|
default:
|
|
break;
|
|
}
|
|
break;
|
|
default:
|
|
efc_log_err(efc, "Unhandled frame rctl: %02x\n", hdr->fh_r_ctl);
|
|
}
|
|
}
|