blob: c6c6b659005fa3eadf35fa1c1c83917638fed4e6 [file] [log] [blame]
/*
* CDDL HEADER START
*
* The contents of this file are subject to the terms of the
* Common Development and Distribution License (the "License").
* You may not use this file except in compliance with the License.
*
* You can obtain a copy of the license at usr/src/OPENSOLARIS.LICENSE
* or http://www.opensolaris.org/os/licensing.
* See the License for the specific language governing permissions
* and limitations under the License.
*
* When distributing Covered Code, include this CDDL HEADER in each
* file and include the License file at usr/src/OPENSOLARIS.LICENSE.
* If applicable, add the following below this CDDL HEADER, with the
* fields enclosed by brackets "[]" replaced with your own identifying
* information: Portions Copyright [yyyy] [name of copyright owner]
*
* CDDL HEADER END
*/
/*
* Copyright 2009 Sun Microsystems, Inc. All rights reserved.
* Use is subject to license terms.
* Copyright (c) 2016 by Delphix. All rights reserved.
*/
/*
* gld - Generic LAN Driver Version 2, PSARC/1997/382
*
* This is a utility module that provides generic facilities for
* LAN drivers. The DLPI protocol and most STREAMS interfaces
* are handled here.
*
* It no longer provides compatibility with drivers
* implemented according to the GLD v0 documentation published
* in 1993. (See PSARC 2003/728)
*/
#include <sys/types.h>
#include <sys/errno.h>
#include <sys/stropts.h>
#include <sys/stream.h>
#include <sys/kmem.h>
#include <sys/stat.h>
#include <sys/modctl.h>
#include <sys/kstat.h>
#include <sys/debug.h>
#include <sys/note.h>
#include <sys/sysmacros.h>
#include <sys/byteorder.h>
#include <sys/strsun.h>
#include <sys/strsubr.h>
#include <sys/dlpi.h>
#include <sys/pattr.h>
#include <sys/ethernet.h>
#include <sys/ib/clients/ibd/ibd.h>
#include <sys/policy.h>
#include <sys/atomic.h>
#include <sys/multidata.h>
#include <sys/gld.h>
#include <sys/gldpriv.h>
#include <sys/ddi.h>
#include <sys/sunddi.h>
/*
* Macros to increment statistics.
*/
/*
* Increase kstats. Note this operation is not atomic. It can be used when
* GLDM_LOCK_HELD_WRITE(macinfo).
*/
#define BUMP(stats, vstats, stat, delta) do { \
((stats)->stat) += (delta); \
_NOTE(CONSTANTCONDITION) \
if ((vstats) != NULL) \
((struct gld_stats *)(vstats))->stat += (delta); \
_NOTE(CONSTANTCONDITION) \
} while (0)
#define ATOMIC_BUMP_STAT(stat, delta) do { \
_NOTE(CONSTANTCONDITION) \
if (sizeof ((stat)) == sizeof (uint32_t)) { \
atomic_add_32((uint32_t *)&(stat), (delta)); \
_NOTE(CONSTANTCONDITION) \
} else if (sizeof ((stat)) == sizeof (uint64_t)) { \
atomic_add_64((uint64_t *)&(stat), (delta)); \
} \
_NOTE(CONSTANTCONDITION) \
} while (0)
#define ATOMIC_BUMP(stats, vstats, stat, delta) do { \
ATOMIC_BUMP_STAT((stats)->stat, (delta)); \
_NOTE(CONSTANTCONDITION) \
if ((vstats) != NULL) { \
ATOMIC_BUMP_STAT(((struct gld_stats *)(vstats))->stat, \
(delta)); \
} \
_NOTE(CONSTANTCONDITION) \
} while (0)
#define UPDATE_STATS(stats, vstats, pktinfo, delta) { \
if ((pktinfo).isBroadcast) { \
ATOMIC_BUMP((stats), (vstats), \
glds_brdcstxmt, (delta)); \
} else if ((pktinfo).isMulticast) { \
ATOMIC_BUMP((stats), (vstats), glds_multixmt, (delta)); \
} \
ATOMIC_BUMP((stats), (vstats), glds_bytexmt64, \
((pktinfo).pktLen)); \
ATOMIC_BUMP((stats), (vstats), glds_pktxmt64, (delta)); \
}
#ifdef GLD_DEBUG
int gld_debug = GLDERRS;
#endif
/* called from gld_register */
static int gld_initstats(gld_mac_info_t *);
/* called from kstat mechanism, and from wsrv's get_statistics */
static int gld_update_kstat(kstat_t *, int);
/* statistics for additional vlans */
static int gld_init_vlan_stats(gld_vlan_t *);
static int gld_update_vlan_kstat(kstat_t *, int);
/* called from gld_getinfo */
static dev_info_t *gld_finddevinfo(dev_t);
/* called from wput, wsrv, unidata, and v0_sched to send a packet */
/* also from the source routing stuff for sending RDE protocol packets */
static int gld_start(queue_t *, mblk_t *, int, uint32_t);
static int gld_start_mdt(queue_t *, mblk_t *, int);
/* called from gld_start[_mdt] to loopback packet(s) in promiscuous mode */
static void gld_precv(gld_mac_info_t *, mblk_t *, uint32_t, struct gld_stats *);
static void gld_precv_mdt(gld_mac_info_t *, gld_vlan_t *, mblk_t *,
pdesc_t *, pktinfo_t *);
/* receive group: called from gld_recv and gld_precv* with maclock held */
static void gld_sendup(gld_mac_info_t *, pktinfo_t *, mblk_t *,
int (*)());
static int gld_accept(gld_t *, pktinfo_t *);
static int gld_mcmatch(gld_t *, pktinfo_t *);
static int gld_multicast(unsigned char *, gld_t *);
static int gld_paccept(gld_t *, pktinfo_t *);
static void gld_passon(gld_t *, mblk_t *, pktinfo_t *,
void (*)(queue_t *, mblk_t *));
static mblk_t *gld_addudind(gld_t *, mblk_t *, pktinfo_t *, boolean_t);
/* wsrv group: called from wsrv, single threaded per queue */
static int gld_ioctl(queue_t *, mblk_t *);
static void gld_fastpath(gld_t *, queue_t *, mblk_t *);
static int gld_cmds(queue_t *, mblk_t *);
static mblk_t *gld_bindack(queue_t *, mblk_t *);
static int gld_notify_req(queue_t *, mblk_t *);
static int gld_udqos(queue_t *, mblk_t *);
static int gld_bind(queue_t *, mblk_t *);
static int gld_unbind(queue_t *, mblk_t *);
static int gld_inforeq(queue_t *, mblk_t *);
static int gld_unitdata(queue_t *, mblk_t *);
static int gldattach(queue_t *, mblk_t *);
static int gldunattach(queue_t *, mblk_t *);
static int gld_enable_multi(queue_t *, mblk_t *);
static int gld_disable_multi(queue_t *, mblk_t *);
static void gld_send_disable_multi(gld_mac_info_t *, gld_mcast_t *);
static int gld_promisc(queue_t *, mblk_t *, t_uscalar_t, boolean_t);
static int gld_physaddr(queue_t *, mblk_t *);
static int gld_setaddr(queue_t *, mblk_t *);
static int gld_get_statistics(queue_t *, mblk_t *);
static int gld_cap(queue_t *, mblk_t *);
static int gld_cap_ack(queue_t *, mblk_t *);
static int gld_cap_enable(queue_t *, mblk_t *);
/* misc utilities, some requiring various mutexes held */
static int gld_start_mac(gld_mac_info_t *);
static void gld_stop_mac(gld_mac_info_t *);
static void gld_set_ipq(gld_t *);
static void gld_flushqueue(queue_t *);
static glddev_t *gld_devlookup(int);
static int gld_findminor(glddev_t *);
static void gldinsque(void *, void *);
static void gldremque(void *);
void gld_bitrevcopy(caddr_t, caddr_t, size_t);
void gld_bitreverse(uchar_t *, size_t);
char *gld_macaddr_sprintf(char *, unsigned char *, int);
static gld_vlan_t *gld_add_vlan(gld_mac_info_t *, uint32_t vid);
static void gld_rem_vlan(gld_vlan_t *);
gld_vlan_t *gld_find_vlan(gld_mac_info_t *, uint32_t);
gld_vlan_t *gld_get_vlan(gld_mac_info_t *, uint32_t);
#ifdef GLD_DEBUG
static void gld_check_assertions(void);
extern void gld_sr_dump(gld_mac_info_t *);
#endif
/*
* Allocate and zero-out "number" structures each of type "structure" in
* kernel memory.
*/
#define GLD_GETSTRUCT(structure, number) \
(kmem_zalloc((uint_t)(sizeof (structure) * (number)), KM_NOSLEEP))
#define abs(a) ((a) < 0 ? -(a) : a)
uint32_t gld_global_options = GLD_OPT_NO_ETHRXSNAP;
/*
* The device is of DL_ETHER type and is able to support VLAN by itself.
*/
#define VLAN_CAPABLE(macinfo) \
((macinfo)->gldm_type == DL_ETHER && \
(macinfo)->gldm_send_tagged != NULL)
/*
* The set of notifications generatable by GLD itself, the additional
* set that can be generated if the MAC driver provide the link-state
* tracking callback capability, and the set supported by the GLD
* notification code below.
*
* PLEASE keep these in sync with what the code actually does!
*/
static const uint32_t gld_internal_notes = DL_NOTE_PROMISC_ON_PHYS |
DL_NOTE_PROMISC_OFF_PHYS |
DL_NOTE_PHYS_ADDR;
static const uint32_t gld_linkstate_notes = DL_NOTE_LINK_DOWN |
DL_NOTE_LINK_UP |
DL_NOTE_SPEED;
static const uint32_t gld_supported_notes = DL_NOTE_PROMISC_ON_PHYS |
DL_NOTE_PROMISC_OFF_PHYS |
DL_NOTE_PHYS_ADDR |
DL_NOTE_LINK_DOWN |
DL_NOTE_LINK_UP |
DL_NOTE_SPEED;
/* Media must correspond to #defines in gld.h */
static char *gld_media[] = {
"unknown", /* GLDM_UNKNOWN - driver cannot determine media */
"aui", /* GLDM_AUI */
"bnc", /* GLDM_BNC */
"twpair", /* GLDM_TP */
"fiber", /* GLDM_FIBER */
"100baseT", /* GLDM_100BT */
"100vgAnyLan", /* GLDM_VGANYLAN */
"10baseT", /* GLDM_10BT */
"ring4", /* GLDM_RING4 */
"ring16", /* GLDM_RING16 */
"PHY/MII", /* GLDM_PHYMII */
"100baseTX", /* GLDM_100BTX */
"100baseT4", /* GLDM_100BT4 */
"unknown", /* skip */
"ipib", /* GLDM_IB */
};
/* Must correspond to #defines in gld.h */
static char *gld_duplex[] = {
"unknown", /* GLD_DUPLEX_UNKNOWN - not known or not applicable */
"half", /* GLD_DUPLEX_HALF */
"full" /* GLD_DUPLEX_FULL */
};
/*
* Interface types currently supported by GLD.
* If you add new types, you must check all "XXX" strings in the GLD source
* for implementation issues that may affect the support of your new type.
* In particular, any type with gldm_addrlen > 6, or gldm_saplen != -2, will
* require generalizing this GLD source to handle the new cases. In other
* words there are assumptions built into the code in a few places that must
* be fixed. Be sure to turn on DEBUG/ASSERT code when testing a new type.
*/
static gld_interface_t interfaces[] = {
/* Ethernet Bus */
{
DL_ETHER,
(uint_t)-1,
sizeof (struct ether_header),
gld_interpret_ether,
NULL,
gld_fastpath_ether,
gld_unitdata_ether,
gld_init_ether,
gld_uninit_ether,
"ether"
},
/* Fiber Distributed data interface */
{
DL_FDDI,
4352,
sizeof (struct fddi_mac_frm),
gld_interpret_fddi,
NULL,
gld_fastpath_fddi,
gld_unitdata_fddi,
gld_init_fddi,
gld_uninit_fddi,
"fddi"
},
/* Token Ring interface */
{
DL_TPR,
17914,
-1, /* variable header size */
gld_interpret_tr,
NULL,
gld_fastpath_tr,
gld_unitdata_tr,
gld_init_tr,
gld_uninit_tr,
"tpr"
},
/* Infiniband */
{
DL_IB,
4092,
sizeof (struct ipoib_header),
gld_interpret_ib,
gld_interpret_mdt_ib,
gld_fastpath_ib,
gld_unitdata_ib,
gld_init_ib,
gld_uninit_ib,
"ipib"
},
};
/*
* bit reversal lookup table.
*/
static uchar_t bit_rev[] = {
0x00, 0x80, 0x40, 0xc0, 0x20, 0xa0, 0x60, 0xe0, 0x10, 0x90, 0x50, 0xd0,
0x30, 0xb0, 0x70, 0xf0, 0x08, 0x88, 0x48, 0xc8, 0x28, 0xa8, 0x68, 0xe8,
0x18, 0x98, 0x58, 0xd8, 0x38, 0xb8, 0x78, 0xf8, 0x04, 0x84, 0x44, 0xc4,
0x24, 0xa4, 0x64, 0xe4, 0x14, 0x94, 0x54, 0xd4, 0x34, 0xb4, 0x74, 0xf4,
0x0c, 0x8c, 0x4c, 0xcc, 0x2c, 0xac, 0x6c, 0xec, 0x1c, 0x9c, 0x5c, 0xdc,
0x3c, 0xbc, 0x7c, 0xfc, 0x02, 0x82, 0x42, 0xc2, 0x22, 0xa2, 0x62, 0xe2,
0x12, 0x92, 0x52, 0xd2, 0x32, 0xb2, 0x72, 0xf2, 0x0a, 0x8a, 0x4a, 0xca,
0x2a, 0xaa, 0x6a, 0xea, 0x1a, 0x9a, 0x5a, 0xda, 0x3a, 0xba, 0x7a, 0xfa,
0x06, 0x86, 0x46, 0xc6, 0x26, 0xa6, 0x66, 0xe6, 0x16, 0x96, 0x56, 0xd6,
0x36, 0xb6, 0x76, 0xf6, 0x0e, 0x8e, 0x4e, 0xce, 0x2e, 0xae, 0x6e, 0xee,
0x1e, 0x9e, 0x5e, 0xde, 0x3e, 0xbe, 0x7e, 0xfe, 0x01, 0x81, 0x41, 0xc1,
0x21, 0xa1, 0x61, 0xe1, 0x11, 0x91, 0x51, 0xd1, 0x31, 0xb1, 0x71, 0xf1,
0x09, 0x89, 0x49, 0xc9, 0x29, 0xa9, 0x69, 0xe9, 0x19, 0x99, 0x59, 0xd9,
0x39, 0xb9, 0x79, 0xf9, 0x05, 0x85, 0x45, 0xc5, 0x25, 0xa5, 0x65, 0xe5,
0x15, 0x95, 0x55, 0xd5, 0x35, 0xb5, 0x75, 0xf5, 0x0d, 0x8d, 0x4d, 0xcd,
0x2d, 0xad, 0x6d, 0xed, 0x1d, 0x9d, 0x5d, 0xdd, 0x3d, 0xbd, 0x7d, 0xfd,
0x03, 0x83, 0x43, 0xc3, 0x23, 0xa3, 0x63, 0xe3, 0x13, 0x93, 0x53, 0xd3,
0x33, 0xb3, 0x73, 0xf3, 0x0b, 0x8b, 0x4b, 0xcb, 0x2b, 0xab, 0x6b, 0xeb,
0x1b, 0x9b, 0x5b, 0xdb, 0x3b, 0xbb, 0x7b, 0xfb, 0x07, 0x87, 0x47, 0xc7,
0x27, 0xa7, 0x67, 0xe7, 0x17, 0x97, 0x57, 0xd7, 0x37, 0xb7, 0x77, 0xf7,
0x0f, 0x8f, 0x4f, 0xcf, 0x2f, 0xaf, 0x6f, 0xef, 0x1f, 0x9f, 0x5f, 0xdf,
0x3f, 0xbf, 0x7f, 0xff,
};
/*
* User priorities, mapped from b_band.
*/
static uint32_t user_priority[] = {
0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0,
0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0,
1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1,
1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1,
2, 2, 2, 2, 2, 2, 2, 2, 2, 2, 2, 2, 2, 2, 2, 2,
2, 2, 2, 2, 2, 2, 2, 2, 2, 2, 2, 2, 2, 2, 2, 2,
3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3,
3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3, 3,
4, 4, 4, 4, 4, 4, 4, 4, 4, 4, 4, 4, 4, 4, 4, 4,
4, 4, 4, 4, 4, 4, 4, 4, 4, 4, 4, 4, 4, 4, 4, 4,
5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5,
5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5, 5,
6, 6, 6, 6, 6, 6, 6, 6, 6, 6, 6, 6, 6, 6, 6, 6,
6, 6, 6, 6, 6, 6, 6, 6, 6, 6, 6, 6, 6, 6, 6, 6,
7, 7, 7, 7, 7, 7, 7, 7, 7, 7, 7, 7, 7, 7, 7, 7,
7, 7, 7, 7, 7, 7, 7, 7, 7, 7, 7, 7, 7, 7, 7, 7
};
#define UPRI(gld, band) ((band != 0) ? user_priority[(band)] : (gld)->gld_upri)
static struct glddevice gld_device_list; /* Per-system root of GLD tables */
/*
* Module linkage information for the kernel.
*/
static struct modldrv modlmisc = {
&mod_miscops, /* Type of module - a utility provider */
"Generic LAN Driver (" GLD_VERSION_STRING ")"
#ifdef GLD_DEBUG
" DEBUG"
#endif
};
static struct modlinkage modlinkage = {
MODREV_1, &modlmisc, NULL
};
int
_init(void)
{
int e;
/* initialize gld_device_list mutex */
mutex_init(&gld_device_list.gld_devlock, NULL, MUTEX_DRIVER, NULL);
/* initialize device driver (per-major) list */
gld_device_list.gld_next =
gld_device_list.gld_prev = &gld_device_list;
if ((e = mod_install(&modlinkage)) != 0)
mutex_destroy(&gld_device_list.gld_devlock);
return (e);
}
int
_fini(void)
{
int e;
if ((e = mod_remove(&modlinkage)) != 0)
return (e);
ASSERT(gld_device_list.gld_next ==
(glddev_t *)&gld_device_list.gld_next);
ASSERT(gld_device_list.gld_prev ==
(glddev_t *)&gld_device_list.gld_next);
mutex_destroy(&gld_device_list.gld_devlock);
return (e);
}
int
_info(struct modinfo *modinfop)
{
return (mod_info(&modlinkage, modinfop));
}
/*
* GLD service routines
*/
/* So this gld binary maybe can be forward compatible with future v2 drivers */
#define GLD_MAC_RESERVED (16 * sizeof (caddr_t))
/*ARGSUSED*/
gld_mac_info_t *
gld_mac_alloc(dev_info_t *devinfo)
{
gld_mac_info_t *macinfo;
macinfo = kmem_zalloc(sizeof (gld_mac_info_t) + GLD_MAC_RESERVED,
KM_SLEEP);
/*
* The setting of gldm_driver_version will not be documented or allowed
* until a future release.
*/
macinfo->gldm_driver_version = GLD_VERSION_200;
/*
* GLD's version. This also is undocumented for now, but will be
* available if needed in the future.
*/
macinfo->gldm_GLD_version = GLD_VERSION;
return (macinfo);
}
/*
* gld_mac_free must be called after the driver has removed interrupts
* and completely stopped calling gld_recv() and gld_sched(). At that
* point the interrupt routine is guaranteed by the system to have been
* exited and the maclock is no longer needed. Of course, it is
* expected (required) that (assuming gld_register() succeeded),
* gld_unregister() was called before gld_mac_free().
*/
void
gld_mac_free(gld_mac_info_t *macinfo)
{
ASSERT(macinfo);
ASSERT(macinfo->gldm_GLD_version == GLD_VERSION);
/*
* Assert that if we made it through gld_register, then we must
* have unregistered.
*/
ASSERT(!GLDM_LOCK_INITED(macinfo) ||
(macinfo->gldm_GLD_flags & GLD_UNREGISTERED));
GLDM_LOCK_DESTROY(macinfo);
kmem_free(macinfo, sizeof (gld_mac_info_t) + GLD_MAC_RESERVED);
}
/*
* gld_register -- called once per device instance (PPA)
*
* During its attach routine, a real device driver will register with GLD
* so that later opens and dl_attach_reqs will work. The arguments are the
* devinfo pointer, the device name, and a macinfo structure describing the
* physical device instance.
*/
int
gld_register(dev_info_t *devinfo, char *devname, gld_mac_info_t *macinfo)
{
int mediatype;
int major = ddi_name_to_major(devname), i;
glddev_t *glddev;
gld_mac_pvt_t *mac_pvt;
char minordev[32];
char pbuf[3*GLD_MAX_ADDRLEN];
gld_interface_t *ifp;
ASSERT(devinfo != NULL);
ASSERT(macinfo != NULL);
if (macinfo->gldm_driver_version != GLD_VERSION)
return (DDI_FAILURE);
mediatype = macinfo->gldm_type;
/*
* Entry points should be ready for us.
* ioctl is optional.
* set_multicast and get_stats are optional in v0.
* intr is only required if you add an interrupt.
*/
ASSERT(macinfo->gldm_reset != NULL);
ASSERT(macinfo->gldm_start != NULL);
ASSERT(macinfo->gldm_stop != NULL);
ASSERT(macinfo->gldm_set_mac_addr != NULL);
ASSERT(macinfo->gldm_set_promiscuous != NULL);
ASSERT(macinfo->gldm_send != NULL);
ASSERT(macinfo->gldm_maxpkt >= macinfo->gldm_minpkt);
ASSERT(macinfo->gldm_GLD_version == GLD_VERSION);
ASSERT(macinfo->gldm_broadcast_addr != NULL);
ASSERT(macinfo->gldm_vendor_addr != NULL);
ASSERT(macinfo->gldm_ident != NULL);
if (macinfo->gldm_addrlen > GLD_MAX_ADDRLEN) {
cmn_err(CE_WARN, "GLD: %s driver gldm_addrlen %d > %d not sup"
"ported", devname, macinfo->gldm_addrlen, GLD_MAX_ADDRLEN);
return (DDI_FAILURE);
}
/*
* GLD only functions properly with saplen == -2
*/
if (macinfo->gldm_saplen != -2) {
cmn_err(CE_WARN, "GLD: %s driver gldm_saplen %d != -2 "
"not supported", devname, macinfo->gldm_saplen);
return (DDI_FAILURE);
}
/* see gld_rsrv() */
if (ddi_getprop(DDI_DEV_T_NONE, devinfo, 0, "fast_recv", 0))
macinfo->gldm_options |= GLDOPT_FAST_RECV;
mutex_enter(&gld_device_list.gld_devlock);
glddev = gld_devlookup(major);
/*
* Allocate per-driver (major) data structure if necessary
*/
if (glddev == NULL) {
/* first occurrence of this device name (major number) */
glddev = GLD_GETSTRUCT(glddev_t, 1);
if (glddev == NULL) {
mutex_exit(&gld_device_list.gld_devlock);
return (DDI_FAILURE);
}
(void) strncpy(glddev->gld_name, devname,
sizeof (glddev->gld_name) - 1);
glddev->gld_major = major;
glddev->gld_nextminor = GLD_MIN_CLONE_MINOR;
glddev->gld_mac_next = glddev->gld_mac_prev =
(gld_mac_info_t *)&glddev->gld_mac_next;
glddev->gld_str_next = glddev->gld_str_prev =
(gld_t *)&glddev->gld_str_next;
mutex_init(&glddev->gld_devlock, NULL, MUTEX_DRIVER, NULL);
/* allow increase of number of supported multicast addrs */
glddev->gld_multisize = ddi_getprop(DDI_DEV_T_NONE,
devinfo, 0, "multisize", GLD_MAX_MULTICAST);
/*
* Optionally restrict DLPI provider style
*
* -1 - don't create style 1 nodes
* -2 - don't create style 2 nodes
*/
glddev->gld_styles = ddi_getprop(DDI_DEV_T_NONE, devinfo, 0,
"gld-provider-styles", 0);
/* Stuff that's needed before any PPA gets attached */
glddev->gld_type = macinfo->gldm_type;
glddev->gld_minsdu = macinfo->gldm_minpkt;
glddev->gld_saplen = macinfo->gldm_saplen;
glddev->gld_addrlen = macinfo->gldm_addrlen;
glddev->gld_broadcast = kmem_zalloc(macinfo->gldm_addrlen,
KM_SLEEP);
bcopy(macinfo->gldm_broadcast_addr,
glddev->gld_broadcast, macinfo->gldm_addrlen);
glddev->gld_maxsdu = macinfo->gldm_maxpkt;
gldinsque(glddev, gld_device_list.gld_prev);
}
glddev->gld_ndevice++;
/* Now glddev can't go away until we unregister this mac (or fail) */
mutex_exit(&gld_device_list.gld_devlock);
/*
* Per-instance initialization
*/
/*
* Initialize per-mac structure that is private to GLD.
* Set up interface pointer. These are device class specific pointers
* used to handle FDDI/TR/ETHER/IPoIB specific packets.
*/
for (i = 0; i < sizeof (interfaces)/sizeof (*interfaces); i++) {
if (mediatype != interfaces[i].mac_type)
continue;
macinfo->gldm_mac_pvt = kmem_zalloc(sizeof (gld_mac_pvt_t),
KM_SLEEP);
((gld_mac_pvt_t *)macinfo->gldm_mac_pvt)->interfacep = ifp =
&interfaces[i];
break;
}
if (ifp == NULL) {
cmn_err(CE_WARN, "GLD: this version does not support %s driver "
"of type %d", devname, mediatype);
goto failure;
}
/*
* Driver can only register MTU within legal media range.
*/
if (macinfo->gldm_maxpkt > ifp->mtu_size) {
cmn_err(CE_WARN, "GLD: oversize MTU is specified by driver %s",
devname);
goto failure;
}
/*
* Correct margin size if it is not set.
*/
if (VLAN_CAPABLE(macinfo) && (macinfo->gldm_margin == 0))
macinfo->gldm_margin = VTAG_SIZE;
/*
* For now, only Infiniband drivers can use MDT. Do not add
* support for Ethernet, FDDI or TR.
*/
if (macinfo->gldm_mdt_pre != NULL) {
if (mediatype != DL_IB) {
cmn_err(CE_WARN, "GLD: MDT not supported for %s "
"driver of type %d", devname, mediatype);
goto failure;
}
/*
* Validate entry points.
*/
if ((macinfo->gldm_mdt_send == NULL) ||
(macinfo->gldm_mdt_post == NULL)) {
cmn_err(CE_WARN, "GLD: invalid MDT entry points for "
"%s driver of type %d", devname, mediatype);
goto failure;
}
macinfo->gldm_options |= GLDOPT_MDT;
}
mac_pvt = (gld_mac_pvt_t *)macinfo->gldm_mac_pvt;
mac_pvt->major_dev = glddev;
mac_pvt->curr_macaddr = kmem_zalloc(macinfo->gldm_addrlen, KM_SLEEP);
/*
* XXX Do bit-reversed devices store gldm_vendor in canonical
* format or in wire format? Also gldm_broadcast. For now
* we are assuming canonical, but I'm not sure that makes the
* most sense for ease of driver implementation.
*/
bcopy(macinfo->gldm_vendor_addr, mac_pvt->curr_macaddr,
macinfo->gldm_addrlen);
mac_pvt->statistics = kmem_zalloc(sizeof (struct gld_stats), KM_SLEEP);
/*
* The available set of notifications is those generatable by GLD
* itself, plus those corresponding to the capabilities of the MAC
* driver, intersected with those supported by gld_notify_ind() above.
*/
mac_pvt->notifications = gld_internal_notes;
if (macinfo->gldm_capabilities & GLD_CAP_LINKSTATE)
mac_pvt->notifications |= gld_linkstate_notes;
mac_pvt->notifications &= gld_supported_notes;
GLDM_LOCK_INIT(macinfo);
ddi_set_driver_private(devinfo, macinfo);
/*
* Now atomically get a PPA and put ourselves on the mac list.
*/
mutex_enter(&glddev->gld_devlock);
#ifdef DEBUG
if (macinfo->gldm_ppa != ddi_get_instance(devinfo))
cmn_err(CE_WARN, "%s%d instance != ppa %d",
ddi_driver_name(devinfo), ddi_get_instance(devinfo),
macinfo->gldm_ppa);
#endif
/*
* Create style 2 node (gated by gld-provider-styles property).
*
* NOTE: When the CLONE_DEV flag is specified to
* ddi_create_minor_node() the minor number argument is
* immaterial. Opens of that node will go via the clone
* driver and gld_open() will always be passed a dev_t with
* minor of zero.
*/
if (glddev->gld_styles != -2) {
if (ddi_create_minor_node(devinfo, glddev->gld_name, S_IFCHR,
0, DDI_NT_NET, CLONE_DEV) == DDI_FAILURE) {
mutex_exit(&glddev->gld_devlock);
goto late_failure;
}
}
/*
* Create style 1 node (gated by gld-provider-styles property)
*/
if (glddev->gld_styles != -1) {
(void) sprintf(minordev, "%s%d", glddev->gld_name,
macinfo->gldm_ppa);
if (ddi_create_minor_node(devinfo, minordev, S_IFCHR,
GLD_STYLE1_PPA_TO_MINOR(macinfo->gldm_ppa), DDI_NT_NET,
0) != DDI_SUCCESS) {
mutex_exit(&glddev->gld_devlock);
goto late_failure;
}
}
/* add ourselves to this major device's linked list of instances */
gldinsque(macinfo, glddev->gld_mac_prev);
mutex_exit(&glddev->gld_devlock);
/*
* Unfortunately we need the ppa before we call gld_initstats();
* otherwise we would like to do this just above the mutex_enter
* above. In which case we could have set MAC_READY inside the
* mutex and we wouldn't have needed to check it in open and
* DL_ATTACH. We wouldn't like to do the initstats/kstat_create
* inside the mutex because it might get taken in our kstat_update
* routine and cause a deadlock with kstat_chain_lock.
*/
/* gld_initstats() calls (*ifp->init)() */
if (gld_initstats(macinfo) != GLD_SUCCESS) {
mutex_enter(&glddev->gld_devlock);
gldremque(macinfo);
mutex_exit(&glddev->gld_devlock);
goto late_failure;
}
/*
* Need to indicate we are NOW ready to process interrupts;
* any interrupt before this is set is for someone else.
* This flag is also now used to tell open, et. al. that this
* mac is now fully ready and available for use.
*/
GLDM_LOCK(macinfo, RW_WRITER);
macinfo->gldm_GLD_flags |= GLD_MAC_READY;
GLDM_UNLOCK(macinfo);
/* log local ethernet address -- XXX not DDI compliant */
if (macinfo->gldm_addrlen == sizeof (struct ether_addr))
(void) localetheraddr(
(struct ether_addr *)macinfo->gldm_vendor_addr, NULL);
/* now put announcement into the message buffer */
cmn_err(CE_CONT, "!%s%d: %s: type \"%s\" mac address %s\n",
glddev->gld_name,
macinfo->gldm_ppa, macinfo->gldm_ident,
mac_pvt->interfacep->mac_string,
gld_macaddr_sprintf(pbuf, macinfo->gldm_vendor_addr,
macinfo->gldm_addrlen));
ddi_report_dev(devinfo);
return (DDI_SUCCESS);
late_failure:
ddi_remove_minor_node(devinfo, NULL);
GLDM_LOCK_DESTROY(macinfo);
if (mac_pvt->curr_macaddr != NULL)
kmem_free(mac_pvt->curr_macaddr, macinfo->gldm_addrlen);
if (mac_pvt->statistics != NULL)
kmem_free(mac_pvt->statistics, sizeof (struct gld_stats));
kmem_free(macinfo->gldm_mac_pvt, sizeof (gld_mac_pvt_t));
macinfo->gldm_mac_pvt = NULL;
failure:
mutex_enter(&gld_device_list.gld_devlock);
glddev->gld_ndevice--;
/*
* Note that just because this goes to zero here does not necessarily
* mean that we were the one who added the glddev above. It's
* possible that the first mac unattached while were were in here
* failing to attach the second mac. But we're now the last.
*/
if (glddev->gld_ndevice == 0) {
/* There should be no macinfos left */
ASSERT(glddev->gld_mac_next ==
(gld_mac_info_t *)&glddev->gld_mac_next);
ASSERT(glddev->gld_mac_prev ==
(gld_mac_info_t *)&glddev->gld_mac_next);
/*
* There should be no DL_UNATTACHED streams: the system
* should not have detached the "first" devinfo which has
* all the open style 2 streams.
*
* XXX This is not clear. See gld_getinfo and Bug 1165519
*/
ASSERT(glddev->gld_str_next == (gld_t *)&glddev->gld_str_next);
ASSERT(glddev->gld_str_prev == (gld_t *)&glddev->gld_str_next);
gldremque(glddev);
mutex_destroy(&glddev->gld_devlock);
if (glddev->gld_broadcast != NULL)
kmem_free(glddev->gld_broadcast, glddev->gld_addrlen);
kmem_free(glddev, sizeof (glddev_t));
}
mutex_exit(&gld_device_list.gld_devlock);
return (DDI_FAILURE);
}
/*
* gld_unregister (macinfo)
* remove the macinfo structure from local structures
* this is cleanup for a driver to be unloaded
*/
int
gld_unregister(gld_mac_info_t *macinfo)
{
gld_mac_pvt_t *mac_pvt = (gld_mac_pvt_t *)macinfo->gldm_mac_pvt;
glddev_t *glddev = mac_pvt->major_dev;
gld_interface_t *ifp;
int multisize = sizeof (gld_mcast_t) * glddev->gld_multisize;
mutex_enter(&glddev->gld_devlock);
GLDM_LOCK(macinfo, RW_WRITER);
if (mac_pvt->nvlan > 0) {
GLDM_UNLOCK(macinfo);
mutex_exit(&glddev->gld_devlock);
return (DDI_FAILURE);
}
#ifdef GLD_DEBUG
{
int i;
for (i = 0; i < VLAN_HASHSZ; i++) {
if ((mac_pvt->vlan_hash[i] != NULL))
cmn_err(CE_PANIC,
"%s, line %d: "
"mac_pvt->vlan_hash[%d] != NULL",
__FILE__, __LINE__, i);
}
}
#endif
/* Delete this mac */
gldremque(macinfo);
/* Disallow further entries to gld_recv() and gld_sched() */
macinfo->gldm_GLD_flags |= GLD_UNREGISTERED;
GLDM_UNLOCK(macinfo);
mutex_exit(&glddev->gld_devlock);
ifp = ((gld_mac_pvt_t *)macinfo->gldm_mac_pvt)->interfacep;
(*ifp->uninit)(macinfo);
ASSERT(mac_pvt->kstatp);
kstat_delete(mac_pvt->kstatp);
ASSERT(GLDM_LOCK_INITED(macinfo));
kmem_free(mac_pvt->curr_macaddr, macinfo->gldm_addrlen);
kmem_free(mac_pvt->statistics, sizeof (struct gld_stats));
if (mac_pvt->mcast_table != NULL)
kmem_free(mac_pvt->mcast_table, multisize);
kmem_free(macinfo->gldm_mac_pvt, sizeof (gld_mac_pvt_t));
macinfo->gldm_mac_pvt = (caddr_t)NULL;
/* We now have one fewer instance for this major device */
mutex_enter(&gld_device_list.gld_devlock);
glddev->gld_ndevice--;
if (glddev->gld_ndevice == 0) {
/* There should be no macinfos left */
ASSERT(glddev->gld_mac_next ==
(gld_mac_info_t *)&glddev->gld_mac_next);
ASSERT(glddev->gld_mac_prev ==
(gld_mac_info_t *)&glddev->gld_mac_next);
/*
* There should be no DL_UNATTACHED streams: the system
* should not have detached the "first" devinfo which has
* all the open style 2 streams.
*
* XXX This is not clear. See gld_getinfo and Bug 1165519
*/
ASSERT(glddev->gld_str_next == (gld_t *)&glddev->gld_str_next);
ASSERT(glddev->gld_str_prev == (gld_t *)&glddev->gld_str_next);
ddi_remove_minor_node(macinfo->gldm_devinfo, NULL);
gldremque(glddev);
mutex_destroy(&glddev->gld_devlock);
if (glddev->gld_broadcast != NULL)
kmem_free(glddev->gld_broadcast, glddev->gld_addrlen);
kmem_free(glddev, sizeof (glddev_t));
}
mutex_exit(&gld_device_list.gld_devlock);
return (DDI_SUCCESS);
}
/*
* gld_initstats
* called from gld_register
*/
static int
gld_initstats(gld_mac_info_t *macinfo)
{
gld_mac_pvt_t *mac_pvt = (gld_mac_pvt_t *)macinfo->gldm_mac_pvt;
struct gldkstats *sp;
glddev_t *glddev;
kstat_t *ksp;
gld_interface_t *ifp;
glddev = mac_pvt->major_dev;
if ((ksp = kstat_create(glddev->gld_name, macinfo->gldm_ppa,
NULL, "net", KSTAT_TYPE_NAMED,
sizeof (struct gldkstats) / sizeof (kstat_named_t), 0)) == NULL) {
cmn_err(CE_WARN,
"GLD: failed to create kstat structure for %s%d",
glddev->gld_name, macinfo->gldm_ppa);
return (GLD_FAILURE);
}
mac_pvt->kstatp = ksp;
ksp->ks_update = gld_update_kstat;
ksp->ks_private = (void *)macinfo;
sp = ksp->ks_data;
kstat_named_init(&sp->glds_pktrcv, "ipackets", KSTAT_DATA_UINT32);
kstat_named_init(&sp->glds_pktxmt, "opackets", KSTAT_DATA_UINT32);
kstat_named_init(&sp->glds_errrcv, "ierrors", KSTAT_DATA_ULONG);
kstat_named_init(&sp->glds_errxmt, "oerrors", KSTAT_DATA_ULONG);
kstat_named_init(&sp->glds_bytexmt, "obytes", KSTAT_DATA_UINT32);
kstat_named_init(&sp->glds_bytercv, "rbytes", KSTAT_DATA_UINT32);
kstat_named_init(&sp->glds_multixmt, "multixmt", KSTAT_DATA_ULONG);
kstat_named_init(&sp->glds_multircv, "multircv", KSTAT_DATA_ULONG);
kstat_named_init(&sp->glds_brdcstxmt, "brdcstxmt", KSTAT_DATA_ULONG);
kstat_named_init(&sp->glds_brdcstrcv, "brdcstrcv", KSTAT_DATA_ULONG);
kstat_named_init(&sp->glds_blocked, "blocked", KSTAT_DATA_ULONG);
kstat_named_init(&sp->glds_noxmtbuf, "noxmtbuf", KSTAT_DATA_ULONG);
kstat_named_init(&sp->glds_norcvbuf, "norcvbuf", KSTAT_DATA_ULONG);
kstat_named_init(&sp->glds_xmtretry, "xmtretry", KSTAT_DATA_ULONG);
kstat_named_init(&sp->glds_intr, "intr", KSTAT_DATA_ULONG);
kstat_named_init(&sp->glds_pktrcv64, "ipackets64", KSTAT_DATA_UINT64);
kstat_named_init(&sp->glds_pktxmt64, "opackets64", KSTAT_DATA_UINT64);
kstat_named_init(&sp->glds_bytexmt64, "obytes64", KSTAT_DATA_UINT64);
kstat_named_init(&sp->glds_bytercv64, "rbytes64", KSTAT_DATA_UINT64);
kstat_named_init(&sp->glds_unknowns, "unknowns", KSTAT_DATA_ULONG);
kstat_named_init(&sp->glds_speed, "ifspeed", KSTAT_DATA_UINT64);
kstat_named_init(&sp->glds_media, "media", KSTAT_DATA_CHAR);
kstat_named_init(&sp->glds_prom, "promisc", KSTAT_DATA_CHAR);
kstat_named_init(&sp->glds_overflow, "oflo", KSTAT_DATA_ULONG);
kstat_named_init(&sp->glds_underflow, "uflo", KSTAT_DATA_ULONG);
kstat_named_init(&sp->glds_missed, "missed", KSTAT_DATA_ULONG);
kstat_named_init(&sp->glds_xmtbadinterp, "xmt_badinterp",
KSTAT_DATA_UINT32);
kstat_named_init(&sp->glds_rcvbadinterp, "rcv_badinterp",
KSTAT_DATA_UINT32);
ifp = ((gld_mac_pvt_t *)macinfo->gldm_mac_pvt)->interfacep;
(*ifp->init)(macinfo);
kstat_install(ksp);
return (GLD_SUCCESS);
}
/* called from kstat mechanism, and from wsrv's get_statistics_req */
static int
gld_update_kstat(kstat_t *ksp, int rw)
{
gld_mac_info_t *macinfo;
gld_mac_pvt_t *mac_pvt;
struct gldkstats *gsp;
struct gld_stats *stats;
if (rw == KSTAT_WRITE)
return (EACCES);
macinfo = (gld_mac_info_t *)ksp->ks_private;
ASSERT(macinfo != NULL);
GLDM_LOCK(macinfo, RW_WRITER);
if (!(macinfo->gldm_GLD_flags & GLD_MAC_READY)) {
GLDM_UNLOCK(macinfo);
return (EIO); /* this one's not ready yet */
}
if (macinfo->gldm_GLD_flags & GLD_UNREGISTERED) {
GLDM_UNLOCK(macinfo);
return (EIO); /* this one's not ready any more */
}
mac_pvt = (gld_mac_pvt_t *)macinfo->gldm_mac_pvt;
gsp = mac_pvt->kstatp->ks_data;
ASSERT(gsp);
stats = mac_pvt->statistics;
if (macinfo->gldm_get_stats)
(void) (*macinfo->gldm_get_stats)(macinfo, stats);
gsp->glds_pktxmt.value.ui32 = stats->glds_pktxmt64 & 0xffffffff;
gsp->glds_bytexmt.value.ui32 = stats->glds_bytexmt64 & 0xffffffff;
gsp->glds_multixmt.value.ul = stats->glds_multixmt;
gsp->glds_brdcstxmt.value.ul = stats->glds_brdcstxmt;
gsp->glds_noxmtbuf.value.ul = stats->glds_noxmtbuf; /* 0 for now */
gsp->glds_xmtretry.value.ul = stats->glds_xmtretry;
gsp->glds_pktxmt64.value.ui64 = stats->glds_pktxmt64;
gsp->glds_bytexmt64.value.ui64 = stats->glds_bytexmt64;
gsp->glds_xmtbadinterp.value.ui32 = stats->glds_xmtbadinterp;
gsp->glds_pktrcv.value.ui32 = stats->glds_pktrcv64 & 0xffffffff;
gsp->glds_errxmt.value.ul = stats->glds_errxmt;
gsp->glds_errrcv.value.ul = stats->glds_errrcv;
gsp->glds_bytercv.value.ui32 = stats->glds_bytercv64 & 0xffffffff;
gsp->glds_multircv.value.ul = stats->glds_multircv;
gsp->glds_brdcstrcv.value.ul = stats->glds_brdcstrcv;
gsp->glds_blocked.value.ul = stats->glds_blocked;
gsp->glds_overflow.value.ul = stats->glds_overflow;
gsp->glds_underflow.value.ul = stats->glds_underflow;
gsp->glds_missed.value.ul = stats->glds_missed;
gsp->glds_norcvbuf.value.ul = stats->glds_norcvbuf +
stats->glds_gldnorcvbuf;
gsp->glds_intr.value.ul = stats->glds_intr;
gsp->glds_speed.value.ui64 = stats->glds_speed;
gsp->glds_unknowns.value.ul = stats->glds_unknowns;
gsp->glds_pktrcv64.value.ui64 = stats->glds_pktrcv64;
gsp->glds_bytercv64.value.ui64 = stats->glds_bytercv64;
gsp->glds_rcvbadinterp.value.ui32 = stats->glds_rcvbadinterp;
if (mac_pvt->nprom)
(void) strcpy(gsp->glds_prom.value.c, "phys");
else if (mac_pvt->nprom_multi)
(void) strcpy(gsp->glds_prom.value.c, "multi");
else
(void) strcpy(gsp->glds_prom.value.c, "off");
(void) strcpy(gsp->glds_media.value.c, gld_media[
stats->glds_media < sizeof (gld_media) / sizeof (gld_media[0])
? stats->glds_media : 0]);
switch (macinfo->gldm_type) {
case DL_ETHER:
gsp->glds_frame.value.ul = stats->glds_frame;
gsp->glds_crc.value.ul = stats->glds_crc;
gsp->glds_collisions.value.ul = stats->glds_collisions;
gsp->glds_excoll.value.ul = stats->glds_excoll;
gsp->glds_defer.value.ul = stats->glds_defer;
gsp->glds_short.value.ul = stats->glds_short;
gsp->glds_xmtlatecoll.value.ul = stats->glds_xmtlatecoll;
gsp->glds_nocarrier.value.ul = stats->glds_nocarrier;
gsp->glds_dot3_first_coll.value.ui32 =
stats->glds_dot3_first_coll;
gsp->glds_dot3_multi_coll.value.ui32 =
stats->glds_dot3_multi_coll;
gsp->glds_dot3_sqe_error.value.ui32 =
stats->glds_dot3_sqe_error;
gsp->glds_dot3_mac_xmt_error.value.ui32 =
stats->glds_dot3_mac_xmt_error;
gsp->glds_dot3_mac_rcv_error.value.ui32 =
stats->glds_dot3_mac_rcv_error;
gsp->glds_dot3_frame_too_long.value.ui32 =
stats->glds_dot3_frame_too_long;
(void) strcpy(gsp->glds_duplex.value.c, gld_duplex[
stats->glds_duplex <
sizeof (gld_duplex) / sizeof (gld_duplex[0]) ?
stats->glds_duplex : 0]);
break;
case DL_TPR:
gsp->glds_dot5_line_error.value.ui32 =
stats->glds_dot5_line_error;
gsp->glds_dot5_burst_error.value.ui32 =
stats->glds_dot5_burst_error;
gsp->glds_dot5_signal_loss.value.ui32 =
stats->glds_dot5_signal_loss;
gsp->glds_dot5_ace_error.value.ui32 =
stats->glds_dot5_ace_error;
gsp->glds_dot5_internal_error.value.ui32 =
stats->glds_dot5_internal_error;
gsp->glds_dot5_lost_frame_error.value.ui32 =
stats->glds_dot5_lost_frame_error;
gsp->glds_dot5_frame_copied_error.value.ui32 =
stats->glds_dot5_frame_copied_error;
gsp->glds_dot5_token_error.value.ui32 =
stats->glds_dot5_token_error;
gsp->glds_dot5_freq_error.value.ui32 =
stats->glds_dot5_freq_error;
break;
case DL_FDDI:
gsp->glds_fddi_mac_error.value.ui32 =
stats->glds_fddi_mac_error;
gsp->glds_fddi_mac_lost.value.ui32 =
stats->glds_fddi_mac_lost;
gsp->glds_fddi_mac_token.value.ui32 =
stats->glds_fddi_mac_token;
gsp->glds_fddi_mac_tvx_expired.value.ui32 =
stats->glds_fddi_mac_tvx_expired;
gsp->glds_fddi_mac_late.value.ui32 =
stats->glds_fddi_mac_late;
gsp->glds_fddi_mac_ring_op.value.ui32 =
stats->glds_fddi_mac_ring_op;
break;
case DL_IB:
break;
default:
break;
}
GLDM_UNLOCK(macinfo);
#ifdef GLD_DEBUG
gld_check_assertions();
if (gld_debug & GLDRDE)
gld_sr_dump(macinfo);
#endif
return (0);
}
static int
gld_init_vlan_stats(gld_vlan_t *vlan)
{
gld_mac_info_t *mac = vlan->gldv_mac;
gld_mac_pvt_t *mac_pvt = (gld_mac_pvt_t *)mac->gldm_mac_pvt;
struct gldkstats *sp;
glddev_t *glddev;
kstat_t *ksp;
char *name;
int instance;
glddev = mac_pvt->major_dev;
name = glddev->gld_name;
instance = (vlan->gldv_id * GLD_VLAN_SCALE) + mac->gldm_ppa;
if ((ksp = kstat_create(name, instance,
NULL, "net", KSTAT_TYPE_NAMED,
sizeof (struct gldkstats) / sizeof (kstat_named_t), 0)) == NULL) {
cmn_err(CE_WARN,
"GLD: failed to create kstat structure for %s%d",
name, instance);
return (GLD_FAILURE);
}
vlan->gldv_kstatp = ksp;
ksp->ks_update = gld_update_vlan_kstat;
ksp->ks_private = (void *)vlan;
sp = ksp->ks_data;
kstat_named_init(&sp->glds_pktrcv, "ipackets", KSTAT_DATA_UINT32);
kstat_named_init(&sp->glds_pktxmt, "opackets", KSTAT_DATA_UINT32);
kstat_named_init(&sp->glds_errrcv, "ierrors", KSTAT_DATA_ULONG);
kstat_named_init(&sp->glds_errxmt, "oerrors", KSTAT_DATA_ULONG);
kstat_named_init(&sp->glds_bytexmt, "obytes", KSTAT_DATA_UINT32);
kstat_named_init(&sp->glds_bytercv, "rbytes", KSTAT_DATA_UINT32);
kstat_named_init(&sp->glds_multixmt, "multixmt", KSTAT_DATA_ULONG);
kstat_named_init(&sp->glds_multircv, "multircv", KSTAT_DATA_ULONG);
kstat_named_init(&sp->glds_brdcstxmt, "brdcstxmt", KSTAT_DATA_ULONG);
kstat_named_init(&sp->glds_brdcstrcv, "brdcstrcv", KSTAT_DATA_ULONG);
kstat_named_init(&sp->glds_blocked, "blocked", KSTAT_DATA_ULONG);
kstat_named_init(&sp->glds_noxmtbuf, "noxmtbuf", KSTAT_DATA_ULONG);
kstat_named_init(&sp->glds_norcvbuf, "norcvbuf", KSTAT_DATA_ULONG);
kstat_named_init(&sp->glds_xmtretry, "xmtretry", KSTAT_DATA_ULONG);
kstat_named_init(&sp->glds_intr, "intr", KSTAT_DATA_ULONG);
kstat_named_init(&sp->glds_pktrcv64, "ipackets64", KSTAT_DATA_UINT64);
kstat_named_init(&sp->glds_pktxmt64, "opackets64", KSTAT_DATA_UINT64);
kstat_named_init(&sp->glds_bytexmt64, "obytes64", KSTAT_DATA_UINT64);
kstat_named_init(&sp->glds_bytercv64, "rbytes64", KSTAT_DATA_UINT64);
kstat_named_init(&sp->glds_unknowns, "unknowns", KSTAT_DATA_ULONG);
kstat_named_init(&sp->glds_speed, "ifspeed", KSTAT_DATA_UINT64);
kstat_named_init(&sp->glds_media, "media", KSTAT_DATA_CHAR);
kstat_named_init(&sp->glds_prom, "promisc", KSTAT_DATA_CHAR);
kstat_named_init(&sp->glds_overflow, "oflo", KSTAT_DATA_ULONG);
kstat_named_init(&sp->glds_underflow, "uflo", KSTAT_DATA_ULONG);
kstat_named_init(&sp->glds_missed, "missed", KSTAT_DATA_ULONG);
kstat_named_init(&sp->glds_xmtbadinterp, "xmt_badinterp",
KSTAT_DATA_UINT32);
kstat_named_init(&sp->glds_rcvbadinterp, "rcv_badinterp",
KSTAT_DATA_UINT32);
kstat_install(ksp);
return (GLD_SUCCESS);
}
static int
gld_update_vlan_kstat(kstat_t *ksp, int rw)
{
gld_vlan_t *vlan;
gld_mac_info_t *macinfo;
struct gldkstats *gsp;
struct gld_stats *stats;
gld_mac_pvt_t *mac_pvt;
uint32_t media;
if (rw == KSTAT_WRITE)
return (EACCES);
vlan = (gld_vlan_t *)ksp->ks_private;
ASSERT(vlan != NULL);
macinfo = vlan->gldv_mac;
GLDM_LOCK(macinfo, RW_WRITER);
mac_pvt = (gld_mac_pvt_t *)macinfo->gldm_mac_pvt;
gsp = vlan->gldv_kstatp->ks_data;
ASSERT(gsp);
stats = vlan->gldv_stats;
gsp->glds_pktxmt.value.ui32 = stats->glds_pktxmt64 & 0xffffffff;
gsp->glds_bytexmt.value.ui32 = stats->glds_bytexmt64 & 0xffffffff;
gsp->glds_errxmt.value.ul = stats->glds_errxmt;
gsp->glds_multixmt.value.ul = stats->glds_multixmt;
gsp->glds_brdcstxmt.value.ul = stats->glds_brdcstxmt;
gsp->glds_noxmtbuf.value.ul = stats->glds_noxmtbuf;
gsp->glds_xmtretry.value.ul = stats->glds_xmtretry;
gsp->glds_pktxmt64.value.ui64 = stats->glds_pktxmt64;
gsp->glds_bytexmt64.value.ui64 = stats->glds_bytexmt64;
gsp->glds_pktrcv.value.ui32 = stats->glds_pktrcv64 & 0xffffffff;
gsp->glds_bytercv.value.ui32 = stats->glds_bytercv64 & 0xffffffff;
gsp->glds_errrcv.value.ul = stats->glds_errrcv;
gsp->glds_multircv.value.ul = stats->glds_multircv;
gsp->glds_brdcstrcv.value.ul = stats->glds_brdcstrcv;
gsp->glds_blocked.value.ul = stats->glds_blocked;
gsp->glds_pktrcv64.value.ui64 = stats->glds_pktrcv64;
gsp->glds_bytercv64.value.ui64 = stats->glds_bytercv64;
gsp->glds_unknowns.value.ul = stats->glds_unknowns;
gsp->glds_xmtbadinterp.value.ui32 = stats->glds_xmtbadinterp;
gsp->glds_rcvbadinterp.value.ui32 = stats->glds_rcvbadinterp;
gsp->glds_speed.value.ui64 = mac_pvt->statistics->glds_speed;
media = mac_pvt->statistics->glds_media;
(void) strcpy(gsp->glds_media.value.c,
gld_media[media < sizeof (gld_media) / sizeof (gld_media[0]) ?
media : 0]);
GLDM_UNLOCK(macinfo);
return (0);
}
/*
* The device dependent driver specifies gld_getinfo as its getinfo routine.
*/
/*ARGSUSED*/
int
gld_getinfo(dev_info_t *dip, ddi_info_cmd_t cmd, void *arg, void **resultp)
{
dev_info_t *devinfo;
minor_t minor = getminor((dev_t)arg);
int rc = DDI_FAILURE;
switch (cmd) {
case DDI_INFO_DEVT2DEVINFO:
if ((devinfo = gld_finddevinfo((dev_t)arg)) != NULL) {
*(dev_info_t **)resultp = devinfo;
rc = DDI_SUCCESS;
}
break;
case DDI_INFO_DEVT2INSTANCE:
/* Need static mapping for deferred attach */
if (minor == GLD_USE_STYLE2) {
/*
* Style 2: this minor number does not correspond to
* any particular instance number.
*/
rc = DDI_FAILURE;
} else if (minor <= GLD_MAX_STYLE1_MINOR) {
/* Style 1: calculate the PPA from the minor */
*resultp = (void *)(uintptr_t)
GLD_STYLE1_MINOR_TO_PPA(minor);
rc = DDI_SUCCESS;
} else {
/* Clone: look for it. Not a static mapping */
if ((devinfo = gld_finddevinfo((dev_t)arg)) != NULL) {
*resultp = (void *)(uintptr_t)
ddi_get_instance(devinfo);
rc = DDI_SUCCESS;
}
}
break;
}
return (rc);
}
/* called from gld_getinfo */
dev_info_t *
gld_finddevinfo(dev_t dev)
{
minor_t minor = getminor(dev);
glddev_t *device;
gld_mac_info_t *mac;
gld_vlan_t *vlan;
gld_t *str;
dev_info_t *devinfo = NULL;
int i;
if (minor == GLD_USE_STYLE2) {
/*
* Style 2: this minor number does not correspond to
* any particular instance number.
*
* XXX We don't know what to say. See Bug 1165519.
*/
return (NULL);
}
mutex_enter(&gld_device_list.gld_devlock); /* hold the device */
device = gld_devlookup(getmajor(dev));
if (device == NULL) {
/* There are no attached instances of this device */
mutex_exit(&gld_device_list.gld_devlock);
return (NULL);
}
/*
* Search all attached macs and streams.
*
* XXX We don't bother checking the DL_UNATTACHED streams since
* we don't know what devinfo we should report back even if we
* found the minor. Maybe we should associate streams that are
* not currently attached to a PPA with the "first" devinfo node
* of the major device to attach -- the one that created the
* minor node for the generic device.
*/
mutex_enter(&device->gld_devlock);
for (mac = device->gld_mac_next;
mac != (gld_mac_info_t *)&device->gld_mac_next;
mac = mac->gldm_next) {
gld_mac_pvt_t *pvt = (gld_mac_pvt_t *)mac->gldm_mac_pvt;
if (!(mac->gldm_GLD_flags & GLD_MAC_READY))
continue; /* this one's not ready yet */
if (minor <= GLD_MAX_STYLE1_MINOR) {
/* Style 1 -- look for the corresponding PPA */
if (minor == GLD_STYLE1_PPA_TO_MINOR(mac->gldm_ppa)) {
devinfo = mac->gldm_devinfo;
goto out; /* found it! */
} else
continue; /* not this PPA */
}
/* We are looking for a clone */
for (i = 0; i < VLAN_HASHSZ; i++) {
for (vlan = pvt->vlan_hash[i];
vlan != NULL; vlan = vlan->gldv_next) {
for (str = vlan->gldv_str_next;
str != (gld_t *)&vlan->gldv_str_next;
str = str->gld_next) {
ASSERT(str->gld_mac_info == mac);
if (minor == str->gld_minor) {
devinfo = mac->gldm_devinfo;
goto out;
}
}
}
}
}
out:
mutex_exit(&device->gld_devlock);
mutex_exit(&gld_device_list.gld_devlock);
return (devinfo);
}
/*
* STREAMS open routine. The device dependent driver specifies this as its
* open entry point.
*/
/*ARGSUSED2*/
int
gld_open(queue_t *q, dev_t *dev, int flag, int sflag, cred_t *cred)
{
gld_mac_pvt_t *mac_pvt;
gld_t *gld;
glddev_t *glddev;
gld_mac_info_t *macinfo;
minor_t minor = getminor(*dev);
gld_vlan_t *vlan;
t_uscalar_t ppa;
ASSERT(q != NULL);
if (minor > GLD_MAX_STYLE1_MINOR)
return (ENXIO);
ASSERT(q->q_ptr == NULL); /* Clone device gives us a fresh Q */
/* Find our per-major glddev_t structure */
mutex_enter(&gld_device_list.gld_devlock);
glddev = gld_devlookup(getmajor(*dev));
/*
* This glddev will hang around since detach (and therefore
* gld_unregister) can't run while we're here in the open routine.
*/
mutex_exit(&gld_device_list.gld_devlock);
if (glddev == NULL)
return (ENXIO);
#ifdef GLD_DEBUG
if (gld_debug & GLDPROT) {
if (minor == GLD_USE_STYLE2)
cmn_err(CE_NOTE, "gld_open(%p, Style 2)", (void *)q);
else
cmn_err(CE_NOTE, "gld_open(%p, Style 1, minor = %d)",
(void *)q, minor);
}
#endif
/*
* get a per-stream structure and link things together so we
* can easily find them later.
*/
gld = kmem_zalloc(sizeof (gld_t), KM_SLEEP);
/*
* fill in the structure and state info
*/
gld->gld_qptr = q;
gld->gld_device = glddev;
gld->gld_state = DL_UNATTACHED;
/*
* we must atomically find a free minor number and add the stream
* to a list, because gld_findminor has to traverse the lists to
* determine which minor numbers are free.
*/
mutex_enter(&glddev->gld_devlock);
/* find a free minor device number for the clone */
gld->gld_minor = gld_findminor(glddev);
if (gld->gld_minor == 0) {
mutex_exit(&glddev->gld_devlock);
kmem_free(gld, sizeof (gld_t));
return (ENOSR);
}
#ifdef GLD_VERBOSE_DEBUG
if (gld_debug & GLDPROT)
cmn_err(CE_NOTE, "gld_open() gld ptr: %p minor: %d",
(void *)gld, gld->gld_minor);
#endif
if (minor == GLD_USE_STYLE2) {
gld->gld_style = DL_STYLE2;
*dev = makedevice(getmajor(*dev), gld->gld_minor);
WR(q)->q_ptr = q->q_ptr = (caddr_t)gld;
gldinsque(gld, glddev->gld_str_prev);
#ifdef GLD_VERBOSE_DEBUG
if (gld_debug & GLDPROT)
cmn_err(CE_NOTE, "GLDstruct added to device list");
#endif
(void) qassociate(q, -1);
goto done;
}
gld->gld_style = DL_STYLE1;
/* the PPA is actually 1 less than the minordev */
ppa = GLD_STYLE1_MINOR_TO_PPA(minor);
for (macinfo = glddev->gld_mac_next;
macinfo != (gld_mac_info_t *)(&glddev->gld_mac_next);
macinfo = macinfo->gldm_next) {
ASSERT(macinfo != NULL);
if (macinfo->gldm_ppa != ppa)
continue;
if (!(macinfo->gldm_GLD_flags & GLD_MAC_READY))
continue; /* this one's not ready yet */
/*
* we found the correct PPA
*/
GLDM_LOCK(macinfo, RW_WRITER);
gld->gld_mac_info = macinfo;
if (macinfo->gldm_send_tagged != NULL)
gld->gld_send = macinfo->gldm_send_tagged;
else
gld->gld_send = macinfo->gldm_send;
/* now ready for action */
gld->gld_state = DL_UNBOUND;
if ((vlan = gld_get_vlan(macinfo, VLAN_VID_NONE)) == NULL) {
GLDM_UNLOCK(macinfo);
mutex_exit(&glddev->gld_devlock);
kmem_free(gld, sizeof (gld_t));
return (EIO);
}
mac_pvt = (gld_mac_pvt_t *)macinfo->gldm_mac_pvt;
if (!mac_pvt->started) {
if (gld_start_mac(macinfo) != GLD_SUCCESS) {
gld_rem_vlan(vlan);
GLDM_UNLOCK(macinfo);
mutex_exit(&glddev->gld_devlock);
kmem_free(gld, sizeof (gld_t));
return (EIO);
}
}
gld->gld_vlan = vlan;
vlan->gldv_nstreams++;
gldinsque(gld, vlan->gldv_str_prev);
*dev = makedevice(getmajor(*dev), gld->gld_minor);
WR(q)->q_ptr = q->q_ptr = (caddr_t)gld;
GLDM_UNLOCK(macinfo);
#ifdef GLD_VERBOSE_DEBUG
if (gld_debug & GLDPROT)
cmn_err(CE_NOTE,
"GLDstruct added to instance list");
#endif
break;
}
if (gld->gld_state == DL_UNATTACHED) {
mutex_exit(&glddev->gld_devlock);
kmem_free(gld, sizeof (gld_t));
return (ENXIO);
}
done:
mutex_exit(&glddev->gld_devlock);
noenable(WR(q)); /* We'll do the qenables manually */
qprocson(q); /* start the queues running */
qenable(WR(q));
return (0);
}
/*
* normal stream close call checks current status and cleans up
* data structures that were dynamically allocated
*/
/*ARGSUSED1*/
int
gld_close(queue_t *q, int flag, cred_t *cred)
{
gld_t *gld = (gld_t *)q->q_ptr;
glddev_t *glddev = gld->gld_device;
ASSERT(q);
ASSERT(gld);
#ifdef GLD_DEBUG
if (gld_debug & GLDPROT) {
cmn_err(CE_NOTE, "gld_close(%p, Style %d)",
(void *)q, (gld->gld_style & 0x1) + 1);
}
#endif
/* Hold all device streams lists still while we check for a macinfo */
mutex_enter(&glddev->gld_devlock);
if (gld->gld_mac_info != NULL) {
/* If there's a macinfo, block recv while we change state */
GLDM_LOCK(gld->gld_mac_info, RW_WRITER);
gld->gld_flags |= GLD_STR_CLOSING; /* no more rcv putnexts */
GLDM_UNLOCK(gld->gld_mac_info);
} else {
/* no mac DL_ATTACHED right now */
gld->gld_flags |= GLD_STR_CLOSING;
}
mutex_exit(&glddev->gld_devlock);
/*
* qprocsoff before we call gld_unbind/gldunattach, so that
* we know wsrv isn't in there trying to undo what we're doing.
*/
qprocsoff(q);
ASSERT(gld->gld_wput_count == 0);
gld->gld_wput_count = 0; /* just in case */
if (gld->gld_state == DL_IDLE) {
/* Need to unbind */
ASSERT(gld->gld_mac_info != NULL);
(void) gld_unbind(WR(q), NULL);
}
if (gld->gld_state == DL_UNBOUND) {
/*
* Need to unattach
* For style 2 stream, gldunattach also
* associate queue with NULL dip
*/
ASSERT(gld->gld_mac_info != NULL);
(void) gldunattach(WR(q), NULL);
}
/* disassociate the stream from the device */
q->q_ptr = WR(q)->q_ptr = NULL;
/*
* Since we unattached above (if necessary), we know that we're
* on the per-major list of unattached streams, rather than a
* per-PPA list. So we know we should hold the devlock.
*/
mutex_enter(&glddev->gld_devlock);
gldremque(gld); /* remove from Style 2 list */
mutex_exit(&glddev->gld_devlock);
kmem_free(gld, sizeof (gld_t));
return (0);
}
/*
* gld_rsrv (q)
* simple read service procedure
* purpose is to avoid the time it takes for packets
* to move through IP so we can get them off the board
* as fast as possible due to limited PC resources.
*
* This is not normally used in the current implementation. It
* can be selected with the undocumented property "fast_recv".
* If that property is set, gld_recv will send the packet
* upstream with a putq() rather than a putnext(), thus causing
* this routine to be scheduled.
*/
int
gld_rsrv(queue_t *q)
{
mblk_t *mp;
while ((mp = getq(q)) != NULL) {
if (canputnext(q)) {
putnext(q, mp);
} else {
freemsg(mp);
}
}
return (0);
}
/*
* gld_wput (q, mp)
* general gld stream write put routine. Receives fastpath data from upper
* modules and processes it immediately. ioctl and M_PROTO/M_PCPROTO are
* queued for later processing by the service procedure.
*/
int
gld_wput(queue_t *q, mblk_t *mp)
{
gld_t *gld = (gld_t *)(q->q_ptr);
int rc;
boolean_t multidata = B_TRUE;
uint32_t upri;
#ifdef GLD_DEBUG
if (gld_debug & GLDTRACE)
cmn_err(CE_NOTE, "gld_wput(%p %p): type %x",
(void *)q, (void *)mp, DB_TYPE(mp));
#endif
switch (DB_TYPE(mp)) {
case M_DATA:
/* fast data / raw support */
/* we must be DL_ATTACHED and DL_BOUND to do this */
/* Tricky to access memory without taking the mutex */
if ((gld->gld_flags & (GLD_RAW | GLD_FAST)) == 0 ||
gld->gld_state != DL_IDLE) {
merror(q, mp, EPROTO);
break;
}
/*
* Cleanup MBLK_VTAG in case it is set by other
* modules. MBLK_VTAG is used to save the vtag information.
*/
GLD_CLEAR_MBLK_VTAG(mp);
multidata = B_FALSE;
/* FALLTHROUGH */
case M_MULTIDATA:
/* Only call gld_start() directly if nothing queued ahead */
/* No guarantees about ordering with different threads */
if (q->q_first)
goto use_wsrv;
/*
* This can happen if wsrv has taken off the last mblk but
* is still processing it.
*/
membar_consumer();
if (gld->gld_in_wsrv)
goto use_wsrv;
/*
* Keep a count of current wput calls to start.
* Nonzero count delays any attempted DL_UNBIND.
* See comments above gld_start().
*/
atomic_inc_32((uint32_t *)&gld->gld_wput_count);
membar_enter();
/* Recheck state now wput_count is set to prevent DL_UNBIND */
/* If this Q is in process of DL_UNBIND, don't call start */
if (gld->gld_state != DL_IDLE || gld->gld_in_unbind) {
/* Extremely unlikely */
atomic_dec_32((uint32_t *)&gld->gld_wput_count);
goto use_wsrv;
}
/*
* Get the priority value. Note that in raw mode, the
* per-packet priority value kept in b_band is ignored.
*/
upri = (gld->gld_flags & GLD_RAW) ? gld->gld_upri :
UPRI(gld, mp->b_band);
rc = (multidata) ? gld_start_mdt(q, mp, GLD_WPUT) :
gld_start(q, mp, GLD_WPUT, upri);
/* Allow DL_UNBIND again */
membar_exit();
atomic_dec_32((uint32_t *)&gld->gld_wput_count);
if (rc == GLD_NORESOURCES)
qenable(q);
break; /* Done with this packet */
use_wsrv:
/* Q not empty, in DL_DETACH, or start gave NORESOURCES */
(void) putq(q, mp);
qenable(q);
break;
case M_IOCTL:
/* ioctl relies on wsrv single threading per queue */
(void) putq(q, mp);
qenable(q);
break;
case M_CTL:
(void) putq(q, mp);
qenable(q);
break;
case M_FLUSH: /* canonical flush handling */
/* XXX Should these be FLUSHALL? */
if (*mp->b_rptr & FLUSHW)
flushq(q, 0);
if (*mp->b_rptr & FLUSHR) {
flushq(RD(q), 0);
*mp->b_rptr &= ~FLUSHW;
qreply(q, mp);
} else
freemsg(mp);
break;
case M_PROTO:
case M_PCPROTO:
/* these rely on wsrv single threading per queue */
(void) putq(q, mp);
qenable(q);
break;
default:
#ifdef GLD_DEBUG
if (gld_debug & GLDETRACE)
cmn_err(CE_WARN,
"gld: Unexpected packet type from queue: 0x%x",
DB_TYPE(mp));
#endif
freemsg(mp);
}
return (0);
}
/*
* gld_wsrv - Incoming messages are processed according to the DLPI protocol
* specification.
*
* wsrv is single-threaded per Q. We make use of this to avoid taking the
* lock for reading data items that are only ever written by us.
*/
int
gld_wsrv(queue_t *q)
{
mblk_t *mp;
gld_t *gld = (gld_t *)q->q_ptr;
gld_mac_info_t *macinfo;
union DL_primitives *prim;
int err;
boolean_t multidata;
uint32_t upri;
#ifdef GLD_DEBUG
if (gld_debug & GLDTRACE)
cmn_err(CE_NOTE, "gld_wsrv(%p)", (void *)q);
#endif
ASSERT(!gld->gld_in_wsrv);
gld->gld_xwait = B_FALSE; /* We are now going to process this Q */
if (q->q_first == NULL)
return (0);
macinfo = gld->gld_mac_info;
/*
* Help wput avoid a call to gld_start if there might be a message
* previously queued by that thread being processed here.
*/
gld->gld_in_wsrv = B_TRUE;
membar_enter();
while ((mp = getq(q)) != NULL) {
switch (DB_TYPE(mp)) {
case M_DATA:
case M_MULTIDATA:
multidata = (DB_TYPE(mp) == M_MULTIDATA);
/*
* retry of a previously processed UNITDATA_REQ
* or is a RAW or FAST message from above.
*/
if (macinfo == NULL) {
/* No longer attached to a PPA, drop packet */
freemsg(mp);
break;
}
gld->gld_sched_ran = B_FALSE;
membar_enter();
/*
* Get the priority value. Note that in raw mode, the
* per-packet priority value kept in b_band is ignored.
*/
upri = (gld->gld_flags & GLD_RAW) ? gld->gld_upri :
UPRI(gld, mp->b_band);
err = (multidata) ? gld_start_mdt(q, mp, GLD_WSRV) :
gld_start(q, mp, GLD_WSRV, upri);
if (err == GLD_NORESOURCES) {
/* gld_sched will qenable us later */
gld->gld_xwait = B_TRUE; /* want qenable */
membar_enter();
/*
* v2: we're not holding the lock; it's
* possible that the driver could have already
* called gld_sched (following up on its
* return of GLD_NORESOURCES), before we got a
* chance to do the putbq() and set gld_xwait.
* So if we saw a call to gld_sched that
* examined this queue, since our call to
* gld_start() above, then it's possible we've
* already seen the only call to gld_sched()
* we're ever going to see. So we better retry
* transmitting this packet right now.
*/
if (gld->gld_sched_ran) {
#ifdef GLD_DEBUG
if (gld_debug & GLDTRACE)
cmn_err(CE_NOTE, "gld_wsrv: "
"sched was called");
#endif
break; /* try again right now */
}
gld->gld_in_wsrv = B_FALSE;
return (0);
}
break;
case M_IOCTL:
(void) gld_ioctl(q, mp);
break;
case M_CTL:
if (macinfo == NULL) {
freemsg(mp);
break;
}
if (macinfo->gldm_mctl != NULL) {
GLDM_LOCK(macinfo, RW_WRITER);
(void) (*macinfo->gldm_mctl) (macinfo, q, mp);
GLDM_UNLOCK(macinfo);
} else {
/* This driver doesn't recognize, just drop */
freemsg(mp);
}
break;
case M_PROTO: /* Will be an DLPI message of some type */
case M_PCPROTO:
if ((err = gld_cmds(q, mp)) != GLDE_OK) {
if (err == GLDE_RETRY) {
gld->gld_in_wsrv = B_FALSE;
return (0); /* quit while we're ahead */
}
prim = (union DL_primitives *)mp->b_rptr;
dlerrorack(q, mp, prim->dl_primitive, err, 0);
}
break;
default:
/* This should never happen */
#ifdef GLD_DEBUG
if (gld_debug & GLDERRS)
cmn_err(CE_WARN,
"gld_wsrv: db_type(%x) not supported",
mp->b_datap->db_type);
#endif
freemsg(mp); /* unknown types are discarded */
break;
}
}
membar_exit();
gld->gld_in_wsrv = B_FALSE;
return (0);
}
/*
* gld_start() can get called from gld_wput(), gld_wsrv(), or gld_unitdata().
*
* We only come directly from wput() in the GLD_FAST (fastpath) or RAW case.
*
* In particular, we must avoid calling gld_precv*() if we came from wput().
* gld_precv*() is where we, on the transmit side, loop back our outgoing
* packets to the receive side if we are in physical promiscuous mode.
* Since the receive side holds a lock across its call to the upstream
* putnext, and that upstream module could well have looped back to our
* wput() routine on the same thread, we cannot call gld_precv* from here
* for fear of causing a recursive lock entry in our receive code.
*
* There is a problem here when coming from gld_wput(). While wput
* only comes here if the queue is attached to a PPA and bound to a SAP
* and there are no messages on the queue ahead of the M_DATA that could
* change that, it is theoretically possible that another thread could
* now wput a DL_UNBIND and a DL_DETACH message, and the wsrv() routine
* could wake up and process them, before we finish processing this
* send of the M_DATA. This can only possibly happen on a Style 2 RAW or
* FAST (fastpath) stream: non RAW/FAST streams always go through wsrv(),
* and Style 1 streams only DL_DETACH in the close routine, where
* qprocsoff() protects us. If this happens we could end up calling
* gldm_send() after we have detached the stream and possibly called
* gldm_stop(). Worse, once the number of attached streams goes to zero,
* detach/unregister could be called, and the macinfo could go away entirely.
*
* No one has ever seen this happen.
*
* It is some trouble to fix this, and we would rather not add any mutex
* logic into the wput() routine, which is supposed to be a "fast"
* path.
*
* What I've done is use an atomic counter to keep a count of the number
* of threads currently calling gld_start() from wput() on this stream.
* If DL_DETACH sees this as nonzero, it putbqs the request back onto
* the queue and qenables, hoping to have better luck next time. Since
* people shouldn't be trying to send after they've asked to DL_DETACH,
* hopefully very soon all the wput=>start threads should have returned
* and the DL_DETACH will succeed. It's hard to test this since the odds
* of the failure even trying to happen are so small. I probably could
* have ignored the whole issue and never been the worse for it.
*
* Because some GLDv2 Ethernet drivers do not allow the size of transmitted
* packet to be greater than ETHERMAX, we must first strip the VLAN tag
* from a tagged packet before passing it to the driver's gld_send() entry
* point function, and pass the VLAN tag as a separate argument. The
* gld_send() function may fail. In that case, the packet will need to be
* queued in order to be processed again in GLD's service routine. As the
* VTAG has already been stripped at that time, we save the VTAG information
* in (the unused fields of) dblk using GLD_SAVE_MBLK_VTAG(), so that the
* VTAG can also be queued and be able to be got when gld_start() is called
* next time from gld_wsrv().
*
* Some rules to use GLD_{CLEAR|SAVE}_MBLK_VTAG macros:
*
* - GLD_SAVE_MBLK_VTAG() must be called to save the VTAG information each time
* the message is queued by putbq().
*
* - GLD_CLEAR_MBLK_VTAG() must be called to clear the bogus VTAG information
* (if any) in dblk before the message is passed to the gld_start() function.
*/
static int
gld_start(queue_t *q, mblk_t *mp, int caller, uint32_t upri)
{
mblk_t *nmp;
gld_t *gld = (gld_t *)q->q_ptr;
gld_mac_info_t *macinfo;
gld_mac_pvt_t *mac_pvt;
int rc;
gld_interface_t *ifp;
pktinfo_t pktinfo;
uint32_t vtag, vid;
uint32_t raw_vtag = 0;
gld_vlan_t *vlan;
struct gld_stats *stats0, *stats = NULL;
ASSERT(DB_TYPE(mp) == M_DATA);
macinfo = gld->gld_mac_info;
mac_pvt = (gld_mac_pvt_t *)macinfo->gldm_mac_pvt;
ifp = mac_pvt->interfacep;
vlan = (gld_vlan_t *)gld->gld_vlan;
vid = vlan->gldv_id;
/*
* If this interface is a VLAN, the kstats of corresponding
* "VLAN 0" should also be updated. Note that the gld_vlan_t
* structure for VLAN 0 might not exist if there are no DLPI
* consumers attaching on VLAN 0. Fortunately we can directly
* access VLAN 0's kstats from macinfo.
*
* Therefore, stats0 (VLAN 0's kstats) must always be
* updated, and stats must to be updated if it is not NULL.
*/
stats0 = mac_pvt->statistics;
if (vid != VLAN_VID_NONE)
stats = vlan->gldv_stats;
if ((*ifp->interpreter)(macinfo, mp, &pktinfo, GLD_TX) != 0) {
#ifdef GLD_DEBUG
if (gld_debug & GLDERRS)
cmn_err(CE_WARN,
"gld_start: failed to interpret outbound packet");
#endif
goto badarg;
}
vtag = VLAN_VID_NONE;
raw_vtag = GLD_GET_MBLK_VTAG(mp);
if (GLD_VTAG_TCI(raw_vtag) != 0) {
uint16_t raw_pri, raw_vid, evid;
/*
* Tagged packet.
*/
raw_pri = GLD_VTAG_PRI(raw_vtag);
raw_vid = GLD_VTAG_VID(raw_vtag);
GLD_CLEAR_MBLK_VTAG(mp);
if (gld->gld_flags & GLD_RAW) {
/*
* In raw mode, we only expect untagged packets or
* special priority-tagged packets on a VLAN stream.
* Drop the packet if its VID is not zero.
*/
if (vid != VLAN_VID_NONE && raw_vid != VLAN_VID_NONE)
goto badarg;
/*
* If it is raw mode, use the per-stream priority if
* the priority is not specified in the packet.
* Otherwise, ignore the priority bits in the packet.
*/
upri = (raw_pri != 0) ? raw_pri : upri;
}
if (vid == VLAN_VID_NONE && vid != raw_vid) {
gld_vlan_t *tmp_vlan;
/*
* This link is a physical link but the packet is
* a VLAN tagged packet, the kstats of corresponding
* VLAN (if any) should also be updated.
*/
tmp_vlan = gld_find_vlan(macinfo, raw_vid);
if (tmp_vlan != NULL)
stats = tmp_vlan->gldv_stats;
}
evid = (vid == VLAN_VID_NONE) ? raw_vid : vid;
if (evid != VLAN_VID_NONE || upri != 0)
vtag = GLD_MAKE_VTAG(upri, VLAN_CFI_ETHER, evid);
} else {
/*
* Untagged packet:
* Get vtag from the attached PPA of this stream.
*/
if ((vid != VLAN_VID_NONE) ||
((macinfo->gldm_type == DL_ETHER) && (upri != 0))) {
vtag = GLD_MAKE_VTAG(upri, VLAN_CFI_ETHER, vid);
}
}
/*
* We're not holding the lock for this check. If the promiscuous
* state is in flux it doesn't matter much if we get this wrong.
*/
if (mac_pvt->nprom > 0) {
/*
* We want to loopback to the receive side, but to avoid
* recursive lock entry: if we came from wput(), which
* could have looped back via IP from our own receive
* interrupt thread, we decline this request. wput()
* will then queue the packet for wsrv(). This means
* that when snoop is running we don't get the advantage
* of the wput() multithreaded direct entry to the
* driver's send routine.
*/
if (caller == GLD_WPUT) {
GLD_SAVE_MBLK_VTAG(mp, raw_vtag);
(void) putbq(q, mp);
return (GLD_NORESOURCES);
}
if (macinfo->gldm_capabilities & GLD_CAP_ZEROCOPY)
nmp = dupmsg_noloan(mp);
else
nmp = dupmsg(mp);
} else
nmp = NULL; /* we need no loopback */
if (ifp->hdr_size > 0 &&
pktinfo.pktLen > ifp->hdr_size + (vtag == 0 ? 0 : VTAG_SIZE) +
macinfo->gldm_maxpkt) {
if (nmp)
freemsg(nmp); /* free the duped message */
#ifdef GLD_DEBUG
if (gld_debug & GLDERRS)
cmn_err(CE_WARN,
"gld_start: oversize outbound packet, size %d,"
"max %d", pktinfo.pktLen,
ifp->hdr_size + (vtag == 0 ? 0 : VTAG_SIZE) +
macinfo->gldm_maxpkt);
#endif
goto badarg;
}
rc = (*gld->gld_send)(macinfo, mp, vtag);
if (rc != GLD_SUCCESS) {
if (rc == GLD_NORESOURCES) {
ATOMIC_BUMP(stats0, stats, glds_xmtretry, 1);
GLD_SAVE_MBLK_VTAG(mp, raw_vtag);
(void) putbq(q, mp);
} else {
/* transmit error; drop the packet */
freemsg(mp);
/* We're supposed to count failed attempts as well */
UPDATE_STATS(stats0, stats, pktinfo, 1);
#ifdef GLD_DEBUG
if (gld_debug & GLDERRS)
cmn_err(CE_WARN,
"gld_start: gldm_send failed %d", rc);
#endif
}
if (nmp)
freemsg(nmp); /* free the dupped message */
return (rc);
}
UPDATE_STATS(stats0, stats, pktinfo, 1);
/*
* Loopback case. The message needs to be returned back on
* the read side. This would silently fail if the dupmsg fails
* above. This is probably OK, if there is no memory to dup the
* block, then there isn't much we could do anyway.
*/
if (nmp) {
GLDM_LOCK(macinfo, RW_WRITER);
gld_precv(macinfo, nmp, vtag, stats);
GLDM_UNLOCK(macinfo);
}
return (GLD_SUCCESS);
badarg:
freemsg(mp);
ATOMIC_BUMP(stats0, stats, glds_xmtbadinterp, 1);
return (GLD_BADARG);
}
/*
* With MDT V.2 a single message mp can have one header area and multiple
* payload areas. A packet is described by dl_pkt_info, and each packet can
* span multiple payload areas (currently with TCP, each packet will have one
* header and at the most two payload areas). MACs might have a limit on the
* number of payload segments (i.e. per packet scatter-gather limit), and
* MDT V.2 has a way of specifying that with mdt_span_limit; the MAC driver
* might also have a limit on the total number of payloads in a message, and
* that is specified by mdt_max_pld.
*/
static int
gld_start_mdt(queue_t *q, mblk_t *mp, int caller)
{
mblk_t *nextmp;
gld_t *gld = (gld_t *)q->q_ptr;
gld_mac_info_t *macinfo = gld->gld_mac_info;
gld_mac_pvt_t *mac_pvt = (gld_mac_pvt_t *)macinfo->gldm_mac_pvt;
int numpacks, mdtpacks;
gld_interface_t *ifp = mac_pvt->interfacep;
pktinfo_t pktinfo;
gld_vlan_t *vlan = (gld_vlan_t *)gld->gld_vlan;
boolean_t doloop = B_FALSE;
multidata_t *dlmdp;
pdescinfo_t pinfo;
pdesc_t *dl_pkt;
void *cookie;
uint_t totLen = 0;
ASSERT(DB_TYPE(mp) == M_MULTIDATA);
/*
* We're not holding the lock for this check. If the promiscuous
* state is in flux it doesn't matter much if we get this wrong.
*/
if (mac_pvt->nprom > 0) {
/*
* We want to loopback to the receive side, but to avoid
* recursive lock entry: if we came from wput(), which
* could have looped back via IP from our own receive
* interrupt thread, we decline this request. wput()
* will then queue the packet for wsrv(). This means
* that when snoop is running we don't get the advantage
* of the wput() multithreaded direct entry to the
* driver's send routine.
*/
if (caller == GLD_WPUT) {
(void) putbq(q, mp);
return (GLD_NORESOURCES);
}
doloop = B_TRUE;
/*
* unlike the M_DATA case, we don't have to call
* dupmsg_noloan here because mmd_transform
* (called by gld_precv_mdt) will make a copy of
* each dblk.
*/
}
while (mp != NULL) {
/*
* The lower layer driver only gets a single multidata
* message; this also makes it easier to handle noresources.
*/
nextmp = mp->b_cont;
mp->b_cont = NULL;
/*
* Get number of packets in this message; if nothing
* to transmit, go to next message.
*/
dlmdp = mmd_getmultidata(mp);
if ((mdtpacks = (int)mmd_getcnt(dlmdp, NULL, NULL)) == 0) {
freemsg(mp);
mp = nextmp;
continue;
}
/*
* Run interpreter to populate media specific pktinfo fields.
* This collects per MDT message information like sap,
* broad/multicast etc.
*/
(void) (*ifp->interpreter_mdt)(macinfo, mp, NULL, &pktinfo,
GLD_MDT_TX);
numpacks = (*macinfo->gldm_mdt_pre)(macinfo, mp, &cookie);
if (numpacks > 0) {
/*
* Driver indicates it can transmit at least 1, and
* possibly all, packets in MDT message.
*/
int count = numpacks;
for (dl_pkt = mmd_getfirstpdesc(dlmdp, &pinfo);
(dl_pkt != NULL);
dl_pkt = mmd_getnextpdesc(dl_pkt, &pinfo)) {
/*
* Format this packet by adding link header and
* adjusting pdescinfo to include it; get
* packet length.
*/
(void) (*ifp->interpreter_mdt)(macinfo, NULL,
&pinfo, &pktinfo, GLD_MDT_TXPKT);
totLen += pktinfo.pktLen;
/*
* Loop back packet before handing to the
* driver.
*/
if (doloop &&
mmd_adjpdesc(dl_pkt, &pinfo) != NULL) {
GLDM_LOCK(macinfo, RW_WRITER);
gld_precv_mdt(macinfo, vlan, mp,
dl_pkt, &pktinfo);
GLDM_UNLOCK(macinfo);
}
/*
* And send off to driver.
*/
(*macinfo->gldm_mdt_send)(macinfo, cookie,
&pinfo);
/*
* Be careful not to invoke getnextpdesc if we
* already sent the last packet, since driver
* might have posted it to hardware causing a
* completion and freemsg() so the MDT data
* structures might not be valid anymore.
*/
if (--count == 0)
break;
}
(*macinfo->gldm_mdt_post)(macinfo, mp, cookie);
pktinfo.pktLen = totLen;
UPDATE_STATS(vlan->gldv_stats, NULL, pktinfo, numpacks);
/*
* In the noresources case (when driver indicates it
* can not transmit all packets in the MDT message),
* adjust to skip the first few packets on retrial.
*/
if (numpacks != mdtpacks) {
/*
* Release already processed packet descriptors.
*/
for (count = 0; count < numpacks; count++) {
dl_pkt = mmd_getfirstpdesc(dlmdp,
&pinfo);
mmd_rempdesc(dl_pkt);
}
ATOMIC_BUMP(vlan->gldv_stats, NULL,
glds_xmtretry, 1);
mp->b_cont = nextmp;
(void) putbq(q, mp);
return (GLD_NORESOURCES);
}
} else if (numpacks == 0) {
/*
* Driver indicates it can not transmit any packets
* currently and will request retrial later.
*/
ATOMIC_BUMP(vlan->gldv_stats, NULL, glds_xmtretry, 1);
mp->b_cont = nextmp;
(void) putbq(q, mp);
return (GLD_NORESOURCES);
} else {
ASSERT(numpacks == -1);
/*
* We're supposed to count failed attempts as well.
*/
dl_pkt = mmd_getfirstpdesc(dlmdp, &pinfo);
while (dl_pkt != NULL) {
/*
* Call interpreter to determine total packet
* bytes that are being dropped.
*/
(void) (*ifp->interpreter_mdt)(macinfo, NULL,
&pinfo, &pktinfo, GLD_MDT_TXPKT);
totLen += pktinfo.pktLen;
dl_pkt = mmd_getnextpdesc(dl_pkt, &pinfo);
}
pktinfo.pktLen = totLen;
UPDATE_STATS(vlan->gldv_stats, NULL, pktinfo, mdtpacks);
/*
* Transmit error; drop the message, move on
* to the next one.
*/
freemsg(mp);
}
/*
* Process the next multidata block, if there is one.
*/
mp = nextmp;
}
return (GLD_SUCCESS);
}
/*
* gld_intr (macinfo)
*/
uint_t
gld_intr(gld_mac_info_t *macinfo)
{
ASSERT(macinfo != NULL);
if (!(macinfo->gldm_GLD_flags & GLD_MAC_READY))
return (DDI_INTR_UNCLAIMED);
return ((*macinfo->gldm_intr)(macinfo));
}
/*
* gld_sched (macinfo)
*
* This routine scans the streams that refer to a specific macinfo
* structure and causes the STREAMS scheduler to try to run them if
* they are marked as waiting for the transmit buffer.
*/
void
gld_sched(gld_mac_info_t *macinfo)
{
gld_mac_pvt_t *mac_pvt;
gld_t *gld;
gld_vlan_t *vlan;
int i;
ASSERT(macinfo != NULL);
GLDM_LOCK(macinfo, RW_WRITER);
if (macinfo->gldm_GLD_flags & GLD_UNREGISTERED) {
/* We're probably being called from a leftover interrupt */
GLDM_UNLOCK(macinfo);
return;
}
mac_pvt = (gld_mac_pvt_t *)macinfo->gldm_mac_pvt;
for (i = 0; i < VLAN_HASHSZ; i++) {
for (vlan = mac_pvt->vlan_hash[i];
vlan != NULL; vlan = vlan->gldv_next) {
for (gld = vlan->gldv_str_next;
gld != (gld_t *)&vlan->gldv_str_next;
gld = gld->gld_next) {
ASSERT(gld->gld_mac_info == macinfo);
gld->gld_sched_ran = B_TRUE;
membar_enter();
if (gld->gld_xwait) {
gld->gld_xwait = B_FALSE;
qenable(WR(gld->gld_qptr));
}
}
}
}
GLDM_UNLOCK(macinfo);
}
/*
* gld_precv (macinfo, mp, vtag, stats)
* called from gld_start to loopback a packet when in promiscuous mode
*
* VLAN 0's statistics need to be updated. If stats is not NULL,
* it needs to be updated as well.
*/
static void
gld_precv(gld_mac_info_t *macinfo, mblk_t *mp, uint32_t vtag,
struct gld_stats *stats)
{
gld_mac_pvt_t *mac_pvt;
gld_interface_t *ifp;
pktinfo_t pktinfo;
ASSERT(GLDM_LOCK_HELD_WRITE(macinfo));
mac_pvt = (gld_mac_pvt_t *)macinfo->gldm_mac_pvt;
ifp = mac_pvt->interfacep;
/*
* call the media specific packet interpreter routine
*/
if ((*ifp->interpreter)(macinfo, mp, &pktinfo, GLD_RXLOOP) != 0) {
freemsg(mp);
BUMP(mac_pvt->statistics, stats, glds_rcvbadinterp, 1);
#ifdef GLD_DEBUG
if (gld_debug & GLDERRS)
cmn_err(CE_WARN,
"gld_precv: interpreter failed");
#endif
return;
}
/*
* Update the vtag information.
*/
pktinfo.isTagged = (vtag != VLAN_VID_NONE);
pktinfo.vid = GLD_VTAG_VID(vtag);
pktinfo.cfi = GLD_VTAG_CFI(vtag);
pktinfo.user_pri = GLD_VTAG_PRI(vtag);
gld_sendup(macinfo, &pktinfo, mp, gld_paccept);
}
/*
* Called from gld_start_mdt to loopback packet(s) when in promiscuous mode.
* Note that 'vlan' is always a physical link, because MDT can only be
* enabled on non-VLAN streams.
*/
/*ARGSUSED*/
static void
gld_precv_mdt(gld_mac_info_t *macinfo, gld_vlan_t *vlan, mblk_t *mp,
pdesc_t *dl_pkt, pktinfo_t *pktinfo)
{
mblk_t *adjmp;
gld_mac_pvt_t *mac_pvt = (gld_mac_pvt_t *)macinfo->gldm_mac_pvt;
gld_interface_t *ifp = mac_pvt->interfacep;
ASSERT(GLDM_LOCK_HELD_WRITE(macinfo));
/*
* Get source/destination.
*/
(void) (*ifp->interpreter_mdt)(macinfo, mp, NULL, pktinfo,
GLD_MDT_RXLOOP);
if ((adjmp = mmd_transform(dl_pkt)) != NULL)
gld_sendup(macinfo, pktinfo, adjmp, gld_paccept);
}
/*
* gld_recv (macinfo, mp)
* called with an mac-level packet in a mblock; take the maclock,
* try the ip4q and ip6q hack, and otherwise call gld_sendup.
*
* V0 drivers already are holding the mutex when they call us.
*/
void
gld_recv(gld_mac_info_t *macinfo, mblk_t *mp)
{
gld_recv_tagged(macinfo, mp, VLAN_VTAG_NONE);
}
void
gld_recv_tagged(gld_mac_info_t *macinfo, mblk_t *mp, uint32_t vtag)
{
gld_mac_pvt_t *mac_pvt;
char pbuf[3*GLD_MAX_ADDRLEN];
pktinfo_t pktinfo;
gld_interface_t *ifp;
queue_t *ipq = NULL;
gld_vlan_t *vlan = NULL, *vlan0 = NULL, *vlann = NULL;
struct gld_stats *stats0, *stats = NULL;
uint32_t vid;
int err;
ASSERT(macinfo != NULL);
ASSERT(mp->b_datap->db_ref);
GLDM_LOCK(macinfo, RW_READER);
if (macinfo->gldm_GLD_flags & GLD_UNREGISTERED) {
/* We're probably being called from a leftover interrupt */
freemsg(mp);
goto done;
}
/*
* If this packet is a VLAN tagged packet, the kstats of corresponding
* "VLAN 0" should also be updated. We can directly access VLAN 0's
* kstats from macinfo.
*
* Further, the packets needs to be passed to VLAN 0 if there is
* any DLPI consumer on VLAN 0 who is interested in tagged packets
* (DL_PROMISC_SAP is on or is bounded to ETHERTYPE_VLAN SAP).
*/
mac_pvt = (gld_mac_pvt_t *)macinfo->gldm_mac_pvt;
stats0 = mac_pvt->statistics;
vid = GLD_VTAG_VID(vtag);
vlan0 = gld_find_vlan(macinfo, VLAN_VID_NONE);
if (vid != VLAN_VID_NONE) {
/*
* If there are no physical DLPI consumers interested in the
* VLAN packet, clear vlan0.
*/
if ((vlan0 != NULL) && (vlan0->gldv_nvlan_sap == 0))
vlan0 = NULL;
/*
* vlann is the VLAN with the same VID as the VLAN packet.
*/
vlann = gld_find_vlan(macinfo, vid);
if (vlann != NULL)
stats = vlann->gldv_stats;
}
vlan = (vid == VLAN_VID_NONE) ? vlan0 : vlann;
ifp = mac_pvt->interfacep;
err = (*ifp->interpreter)(macinfo, mp, &pktinfo, GLD_RXQUICK);
BUMP(stats0, stats, glds_bytercv64, pktinfo.pktLen);
BUMP(stats0, stats, glds_pktrcv64, 1);
if ((vlann == NULL) && (vlan0 == NULL)) {
freemsg(mp);
goto done;
}
/*
* Check whether underlying media code supports the IPQ hack:
*
* - the interpreter could quickly parse the packet
* - the device type supports IPQ (ethernet and IPoIB)
* - there is one, and only one, IP stream bound (to this VLAN)
* - that stream is a "fastpath" stream
* - the packet is of type ETHERTYPE_IP or ETHERTYPE_IPV6
* - there are no streams in promiscuous mode (on this VLAN)
* - if this packet is tagged, there is no need to send this
* packet to physical streams
*/
if ((err != 0) && ((vlan != NULL) && (vlan->gldv_nprom == 0)) &&
(vlan == vlan0 || vlan0 == NULL)) {
switch (pktinfo.ethertype) {
case ETHERTYPE_IP:
ipq = vlan->gldv_ipq;
break;
case ETHERTYPE_IPV6:
ipq = vlan->gldv_ipv6q;
break;
}
}
/*
* Special case for IP; we can simply do the putnext here, if:
* o The IPQ hack is possible (ipq != NULL).
* o the packet is specifically for me, and therefore:
* - the packet is not multicast or broadcast (fastpath only
* wants unicast packets).
*
* o the stream is not asserting flow control.
*/
if (ipq != NULL &&
pktinfo.isForMe &&
canputnext(ipq)) {
/*
* Skip the mac header. We know there is no LLC1/SNAP header
* in this packet
*/
mp->b_rptr += pktinfo.macLen;
putnext(ipq, mp);
goto done;
}
/*
* call the media specific packet interpreter routine
*/
if ((*ifp->interpreter)(macinfo, mp, &pktinfo, GLD_RX) != 0) {
BUMP(stats0, stats, glds_rcvbadinterp, 1);
#ifdef GLD_DEBUG
if (gld_debug & GLDERRS)
cmn_err(CE_WARN,
"gld_recv_tagged: interpreter failed");
#endif
freemsg(mp);
goto done;
}
/*
* This is safe even if vtag is VLAN_VTAG_NONE
*/
pktinfo.vid = vid;
pktinfo.cfi = GLD_VTAG_CFI(vtag);
#ifdef GLD_DEBUG