xref: /illumos-gate/usr/src/uts/common/io/hxge/hxge_send.c (revision 86ef0a63)
13dec9fcdSqs /*
23dec9fcdSqs  * CDDL HEADER START
33dec9fcdSqs  *
43dec9fcdSqs  * The contents of this file are subject to the terms of the
53dec9fcdSqs  * Common Development and Distribution License (the "License").
63dec9fcdSqs  * You may not use this file except in compliance with the License.
73dec9fcdSqs  *
83dec9fcdSqs  * You can obtain a copy of the license at usr/src/OPENSOLARIS.LICENSE
93dec9fcdSqs  * or http://www.opensolaris.org/os/licensing.
103dec9fcdSqs  * See the License for the specific language governing permissions
113dec9fcdSqs  * and limitations under the License.
123dec9fcdSqs  *
133dec9fcdSqs  * When distributing Covered Code, include this CDDL HEADER in each
143dec9fcdSqs  * file and include the License file at usr/src/OPENSOLARIS.LICENSE.
153dec9fcdSqs  * If applicable, add the following below this CDDL HEADER, with the
163dec9fcdSqs  * fields enclosed by brackets "[]" replaced with your own identifying
173dec9fcdSqs  * information: Portions Copyright [yyyy] [name of copyright owner]
183dec9fcdSqs  *
193dec9fcdSqs  * CDDL HEADER END
203dec9fcdSqs  */
21c1374a13SSurya Prakki 
223dec9fcdSqs /*
230dc2366fSVenugopal Iyer  * Copyright 2010 Sun Microsystems, Inc.  All rights reserved.
243dec9fcdSqs  * Use is subject to license terms.
253dec9fcdSqs  */
263dec9fcdSqs 
273dec9fcdSqs #include <hxge_impl.h>
283dec9fcdSqs 
293dec9fcdSqs extern uint32_t hxge_reclaim_pending;
303dec9fcdSqs extern uint32_t hxge_bcopy_thresh;
313dec9fcdSqs extern uint32_t hxge_dvma_thresh;
323dec9fcdSqs extern uint32_t hxge_dma_stream_thresh;
333dec9fcdSqs extern uint32_t	hxge_tx_minfree;
343dec9fcdSqs extern uint32_t	hxge_tx_intr_thres;
353dec9fcdSqs extern uint32_t	hxge_tx_max_gathers;
363dec9fcdSqs extern uint32_t	hxge_tx_tiny_pack;
373dec9fcdSqs extern uint32_t	hxge_tx_use_bcopy;
383dec9fcdSqs 
391ed83081SMichael Speer static int hxge_start(p_hxge_t hxgep, p_tx_ring_t tx_ring_p, p_mblk_t mp);
403dec9fcdSqs 
411ed83081SMichael Speer void
hxge_tx_ring_task(void * arg)421ed83081SMichael Speer hxge_tx_ring_task(void *arg)
431ed83081SMichael Speer {
441ed83081SMichael Speer 	p_tx_ring_t	ring = (p_tx_ring_t)arg;
451ed83081SMichael Speer 
461ed83081SMichael Speer 	MUTEX_ENTER(&ring->lock);
471ed83081SMichael Speer 	(void) hxge_txdma_reclaim(ring->hxgep, ring, 0);
481ed83081SMichael Speer 	MUTEX_EXIT(&ring->lock);
491ed83081SMichael Speer 
501ed83081SMichael Speer 	mac_tx_ring_update(ring->hxgep->mach, ring->ring_handle);
511ed83081SMichael Speer }
523dec9fcdSqs 
531ed83081SMichael Speer static void
hxge_tx_ring_dispatch(p_tx_ring_t ring)541ed83081SMichael Speer hxge_tx_ring_dispatch(p_tx_ring_t ring)
551ed83081SMichael Speer {
561ed83081SMichael Speer 	/*
571ed83081SMichael Speer 	 * Kick the ring task to reclaim some buffers.
581ed83081SMichael Speer 	 */
591ed83081SMichael Speer 	(void) ddi_taskq_dispatch(ring->taskq,
601ed83081SMichael Speer 	    hxge_tx_ring_task, (void *)ring, DDI_SLEEP);
611ed83081SMichael Speer }
621ed83081SMichael Speer 
631ed83081SMichael Speer mblk_t *
hxge_tx_ring_send(void * arg,mblk_t * mp)641ed83081SMichael Speer hxge_tx_ring_send(void *arg, mblk_t *mp)
651ed83081SMichael Speer {
66*86ef0a63SRichard Lowe 	p_hxge_ring_handle_t	rhp = (p_hxge_ring_handle_t)arg;
671ed83081SMichael Speer 	p_hxge_t		hxgep;
681ed83081SMichael Speer 	p_tx_ring_t		tx_ring_p;
691ed83081SMichael Speer 	int			status;
701ed83081SMichael Speer 
711ed83081SMichael Speer 	ASSERT(rhp != NULL);
721ed83081SMichael Speer 	ASSERT((rhp->index >= 0) && (rhp->index < HXGE_MAX_TDCS));
731ed83081SMichael Speer 
741ed83081SMichael Speer 	hxgep = rhp->hxgep;
751ed83081SMichael Speer 	tx_ring_p = hxgep->tx_rings->rings[rhp->index];
761ed83081SMichael Speer 	ASSERT(hxgep == tx_ring_p->hxgep);
771ed83081SMichael Speer 
781ed83081SMichael Speer 	status = hxge_start(hxgep, tx_ring_p, mp);
791ed83081SMichael Speer 	if (status != 0) {
801ed83081SMichael Speer 		hxge_tx_ring_dispatch(tx_ring_p);
811ed83081SMichael Speer 		return (mp);
821ed83081SMichael Speer 	}
831ed83081SMichael Speer 
841ed83081SMichael Speer 	return ((mblk_t *)NULL);
851ed83081SMichael Speer }
861ed83081SMichael Speer 
871ed83081SMichael Speer static int
hxge_start(p_hxge_t hxgep,p_tx_ring_t tx_ring_p,p_mblk_t mp)883dec9fcdSqs hxge_start(p_hxge_t hxgep, p_tx_ring_t tx_ring_p, p_mblk_t mp)
893dec9fcdSqs {
90*86ef0a63SRichard Lowe 	int			dma_status, status = 0;
91*86ef0a63SRichard Lowe 	p_tx_desc_t		tx_desc_ring_vp;
923dec9fcdSqs 	hpi_handle_t		hpi_desc_handle;
93*86ef0a63SRichard Lowe 	hxge_os_dma_handle_t	tx_desc_dma_handle;
94*86ef0a63SRichard Lowe 	p_tx_desc_t		tx_desc_p;
95*86ef0a63SRichard Lowe 	p_tx_msg_t		tx_msg_ring;
96*86ef0a63SRichard Lowe 	p_tx_msg_t		tx_msg_p;
973dec9fcdSqs 	tx_desc_t		tx_desc, *tmp_desc_p;
983dec9fcdSqs 	tx_desc_t		sop_tx_desc, *sop_tx_desc_p;
993dec9fcdSqs 	p_tx_pkt_header_t	hdrp;
1003dec9fcdSqs 	p_tx_pkt_hdr_all_t	pkthdrp;
1013dec9fcdSqs 	uint8_t			npads = 0;
102*86ef0a63SRichard Lowe 	uint64_t		dma_ioaddr;
1033dec9fcdSqs 	uint32_t		dma_flags;
1043dec9fcdSqs 	int			last_bidx;
105*86ef0a63SRichard Lowe 	uint8_t			*b_rptr;
106*86ef0a63SRichard Lowe 	caddr_t			kaddr;
1073dec9fcdSqs 	uint32_t		nmblks;
1083dec9fcdSqs 	uint32_t		ngathers;
1093dec9fcdSqs 	uint32_t		clen;
110*86ef0a63SRichard Lowe 	int			len;
1113dec9fcdSqs 	uint32_t		pkt_len, pack_len, min_len;
1123dec9fcdSqs 	uint32_t		bcopy_thresh;
113*86ef0a63SRichard Lowe 	int			i, cur_index, sop_index;
1143dec9fcdSqs 	uint16_t		tail_index;
1153dec9fcdSqs 	boolean_t		tail_wrap = B_FALSE;
1163dec9fcdSqs 	hxge_dma_common_t	desc_area;
117*86ef0a63SRichard Lowe 	hxge_os_dma_handle_t	dma_handle;
118*86ef0a63SRichard Lowe 	ddi_dma_cookie_t	dma_cookie;
1193dec9fcdSqs 	hpi_handle_t		hpi_handle;
120*86ef0a63SRichard Lowe 	p_mblk_t		nmp;
1213dec9fcdSqs 	p_mblk_t		t_mp;
122*86ef0a63SRichard Lowe 	uint32_t		ncookies;
123*86ef0a63SRichard Lowe 	boolean_t		good_packet;
124*86ef0a63SRichard Lowe 	boolean_t		mark_mode = B_FALSE;
125*86ef0a63SRichard Lowe 	p_hxge_stats_t		statsp;
1263dec9fcdSqs 	p_hxge_tx_ring_stats_t	tdc_stats;
127*86ef0a63SRichard Lowe 	t_uscalar_t		start_offset = 0;
128*86ef0a63SRichard Lowe 	t_uscalar_t		stuff_offset = 0;
129*86ef0a63SRichard Lowe 	t_uscalar_t		end_offset = 0;
130*86ef0a63SRichard Lowe 	t_uscalar_t		value = 0;
131*86ef0a63SRichard Lowe 	t_uscalar_t		cksum_flags = 0;
1323dec9fcdSqs 	boolean_t		cksum_on = B_FALSE;
1333dec9fcdSqs 	uint32_t		boff = 0;
1343dec9fcdSqs 	uint64_t		tot_xfer_len = 0, tmp_len = 0;
1353dec9fcdSqs 	boolean_t		header_set = B_FALSE;
1363dec9fcdSqs 	tdc_tdr_kick_t		kick;
137cf6ef894SMichael Speer 	uint32_t		offset;
1383dec9fcdSqs #ifdef HXGE_DEBUG
139*86ef0a63SRichard Lowe 	p_tx_desc_t		tx_desc_ring_pp;
140*86ef0a63SRichard Lowe 	p_tx_desc_t		tx_desc_pp;
1413dec9fcdSqs 	tx_desc_t		*save_desc_p;
1423dec9fcdSqs 	int			dump_len;
1433dec9fcdSqs 	int			sad_len;
1443dec9fcdSqs 	uint64_t		sad;
1453dec9fcdSqs 	int			xfer_len;
1463dec9fcdSqs 	uint32_t		msgsize;
1473dec9fcdSqs #endif
1483dec9fcdSqs 
1493dec9fcdSqs 	HXGE_DEBUG_MSG((hxgep, TX_CTL,
1503dec9fcdSqs 	    "==> hxge_start: tx dma channel %d", tx_ring_p->tdc));
1513dec9fcdSqs 	HXGE_DEBUG_MSG((hxgep, TX_CTL,
1523dec9fcdSqs 	    "==> hxge_start: Starting tdc %d desc pending %d",
1533dec9fcdSqs 	    tx_ring_p->tdc, tx_ring_p->descs_pending));
1543dec9fcdSqs 
1553dec9fcdSqs 	statsp = hxgep->statsp;
1563dec9fcdSqs 
1573dec9fcdSqs 	if (hxgep->statsp->port_stats.lb_mode == hxge_lb_normal) {
1583dec9fcdSqs 		if (!statsp->mac_stats.link_up) {
1593dec9fcdSqs 			freemsg(mp);
1603dec9fcdSqs 			HXGE_DEBUG_MSG((hxgep, TX_CTL, "==> hxge_start: "
1613dec9fcdSqs 			    "link not up or LB mode"));
1623dec9fcdSqs 			goto hxge_start_fail1;
1633dec9fcdSqs 		}
1643dec9fcdSqs 	}
1653dec9fcdSqs 
1660dc2366fSVenugopal Iyer 	mac_hcksum_get(mp, &start_offset, &stuff_offset, &end_offset, &value,
1670dc2366fSVenugopal Iyer 	    &cksum_flags);
1683dec9fcdSqs 	if (!HXGE_IS_VLAN_PACKET(mp->b_rptr)) {
1693dec9fcdSqs 		start_offset += sizeof (ether_header_t);
1703dec9fcdSqs 		stuff_offset += sizeof (ether_header_t);
1713dec9fcdSqs 	} else {
1723dec9fcdSqs 		start_offset += sizeof (struct ether_vlan_header);
1733dec9fcdSqs 		stuff_offset += sizeof (struct ether_vlan_header);
1743dec9fcdSqs 	}
1753dec9fcdSqs 
1763dec9fcdSqs 	if (cksum_flags & HCK_PARTIALCKSUM) {
1773dec9fcdSqs 		HXGE_DEBUG_MSG((hxgep, TX_CTL,
1783dec9fcdSqs 		    "==> hxge_start: mp $%p len %d "
1793dec9fcdSqs 		    "cksum_flags 0x%x (partial checksum) ",
1803dec9fcdSqs 		    mp, MBLKL(mp), cksum_flags));
1813dec9fcdSqs 		cksum_on = B_TRUE;
1823dec9fcdSqs 	}
1833dec9fcdSqs 
1843dec9fcdSqs 	MUTEX_ENTER(&tx_ring_p->lock);
1853dec9fcdSqs start_again:
1863dec9fcdSqs 	ngathers = 0;
1873dec9fcdSqs 	sop_index = tx_ring_p->wr_index;
1883dec9fcdSqs #ifdef	HXGE_DEBUG
1893dec9fcdSqs 	if (tx_ring_p->descs_pending) {
1903dec9fcdSqs 		HXGE_DEBUG_MSG((hxgep, TX_CTL,
1913dec9fcdSqs 		    "==> hxge_start: desc pending %d ",
1923dec9fcdSqs 		    tx_ring_p->descs_pending));
1933dec9fcdSqs 	}
1943dec9fcdSqs 
1953dec9fcdSqs 	dump_len = (int)(MBLKL(mp));
1963dec9fcdSqs 	dump_len = (dump_len > 128) ? 128: dump_len;
1973dec9fcdSqs 
1983dec9fcdSqs 	HXGE_DEBUG_MSG((hxgep, TX_CTL,
1993dec9fcdSqs 	    "==> hxge_start: tdc %d: dumping ...: b_rptr $%p "
2003dec9fcdSqs 	    "(Before header reserve: ORIGINAL LEN %d)",
2013dec9fcdSqs 	    tx_ring_p->tdc, mp->b_rptr, dump_len));
2023dec9fcdSqs 
2033dec9fcdSqs 	HXGE_DEBUG_MSG((hxgep, TX_CTL,
2043dec9fcdSqs 	    "==> hxge_start: dump packets (IP ORIGINAL b_rptr $%p): %s",
2053dec9fcdSqs 	    mp->b_rptr, hxge_dump_packet((char *)mp->b_rptr, dump_len)));
2063dec9fcdSqs #endif
2073dec9fcdSqs 
2083dec9fcdSqs 	tdc_stats = tx_ring_p->tdc_stats;
2093dec9fcdSqs 	mark_mode = (tx_ring_p->descs_pending &&
2103dec9fcdSqs 	    ((tx_ring_p->tx_ring_size - tx_ring_p->descs_pending) <
2113dec9fcdSqs 	    hxge_tx_minfree));
2123dec9fcdSqs 
2133dec9fcdSqs 	HXGE_DEBUG_MSG((hxgep, TX_CTL,
2143dec9fcdSqs 	    "TX Descriptor ring is channel %d mark mode %d",
2153dec9fcdSqs 	    tx_ring_p->tdc, mark_mode));
2163dec9fcdSqs 
2173dec9fcdSqs 	if (!hxge_txdma_reclaim(hxgep, tx_ring_p, hxge_tx_minfree)) {
2183dec9fcdSqs 		HXGE_DEBUG_MSG((hxgep, TX_CTL,
2193dec9fcdSqs 		    "TX Descriptor ring is full: channel %d", tx_ring_p->tdc));
2203dec9fcdSqs 		HXGE_DEBUG_MSG((hxgep, TX_CTL,
2213dec9fcdSqs 		    "TX Descriptor ring is full: channel %d", tx_ring_p->tdc));
22275d94465SJosef 'Jeff' Sipek 		(void) atomic_cas_32((uint32_t *)&tx_ring_p->queueing, 0, 1);
2233dec9fcdSqs 		tdc_stats->tx_no_desc++;
2243dec9fcdSqs 		MUTEX_EXIT(&tx_ring_p->lock);
2253dec9fcdSqs 		status = 1;
2263dec9fcdSqs 		goto hxge_start_fail1;
2273dec9fcdSqs 	}
2283dec9fcdSqs 
2293dec9fcdSqs 	nmp = mp;
2303dec9fcdSqs 	i = sop_index = tx_ring_p->wr_index;
2313dec9fcdSqs 	nmblks = 0;
2323dec9fcdSqs 	ngathers = 0;
2333dec9fcdSqs 	pkt_len = 0;
2343dec9fcdSqs 	pack_len = 0;
2353dec9fcdSqs 	clen = 0;
2363dec9fcdSqs 	last_bidx = -1;
2373dec9fcdSqs 	good_packet = B_TRUE;
2383dec9fcdSqs 
2393dec9fcdSqs 	desc_area = tx_ring_p->tdc_desc;
2403dec9fcdSqs 	hpi_handle = desc_area.hpi_handle;
2413dec9fcdSqs 	hpi_desc_handle.regh = (hxge_os_acc_handle_t)
2423dec9fcdSqs 	    DMA_COMMON_ACC_HANDLE(desc_area);
243fe930412Sqs 	hpi_desc_handle.hxgep = hxgep;
2443dec9fcdSqs 	tx_desc_ring_vp = (p_tx_desc_t)DMA_COMMON_VPTR(desc_area);
2453dec9fcdSqs #ifdef	HXGE_DEBUG
2463dec9fcdSqs 	tx_desc_ring_pp = (p_tx_desc_t)DMA_COMMON_IOADDR(desc_area);
2473dec9fcdSqs #endif
2483dec9fcdSqs 	tx_desc_dma_handle = (hxge_os_dma_handle_t)DMA_COMMON_HANDLE(desc_area);
2493dec9fcdSqs 	tx_msg_ring = tx_ring_p->tx_msg_ring;
2503dec9fcdSqs 
2513dec9fcdSqs 	HXGE_DEBUG_MSG((hxgep, TX_CTL, "==> hxge_start: wr_index %d i %d",
2523dec9fcdSqs 	    sop_index, i));
2533dec9fcdSqs 
2543dec9fcdSqs #ifdef	HXGE_DEBUG
2553dec9fcdSqs 	msgsize = msgdsize(nmp);
2563dec9fcdSqs 	HXGE_DEBUG_MSG((hxgep, TX_CTL,
2573dec9fcdSqs 	    "==> hxge_start(1): wr_index %d i %d msgdsize %d",
2583dec9fcdSqs 	    sop_index, i, msgsize));
2593dec9fcdSqs #endif
2603dec9fcdSqs 	/*
2613dec9fcdSqs 	 * The first 16 bytes of the premapped buffer are reserved
2623dec9fcdSqs 	 * for header. No padding will be used.
2633dec9fcdSqs 	 */
2643dec9fcdSqs 	pkt_len = pack_len = boff = TX_PKT_HEADER_SIZE;
2653dec9fcdSqs 	if (hxge_tx_use_bcopy) {
2663dec9fcdSqs 		bcopy_thresh = (hxge_bcopy_thresh - TX_PKT_HEADER_SIZE);
2673dec9fcdSqs 	} else {
2683dec9fcdSqs 		bcopy_thresh = (TX_BCOPY_SIZE - TX_PKT_HEADER_SIZE);
2693dec9fcdSqs 	}
2703dec9fcdSqs 	while (nmp) {
2713dec9fcdSqs 		good_packet = B_TRUE;
2723dec9fcdSqs 		b_rptr = nmp->b_rptr;
2733dec9fcdSqs 		len = MBLKL(nmp);
2743dec9fcdSqs 		if (len <= 0) {
2753dec9fcdSqs 			nmp = nmp->b_cont;
2763dec9fcdSqs 			continue;
2773dec9fcdSqs 		}
2783dec9fcdSqs 		nmblks++;
2793dec9fcdSqs 
2803dec9fcdSqs 		HXGE_DEBUG_MSG((hxgep, TX_CTL, "==> hxge_start(1): nmblks %d "
2813dec9fcdSqs 		    "len %d pkt_len %d pack_len %d",
2823dec9fcdSqs 		    nmblks, len, pkt_len, pack_len));
2833dec9fcdSqs 		/*
2843dec9fcdSqs 		 * Hardware limits the transfer length to 4K.
2853dec9fcdSqs 		 * If len is more than 4K, we need to break
2863dec9fcdSqs 		 * nmp into two chunks: Make first chunk smaller
2873dec9fcdSqs 		 * than 4K. The second chunk will be broken into
2883dec9fcdSqs 		 * less than 4K (if needed) during the next pass.
2893dec9fcdSqs 		 */
2903dec9fcdSqs 		if (len > (TX_MAX_TRANSFER_LENGTH - TX_PKT_HEADER_SIZE)) {
2913dec9fcdSqs 			if ((t_mp = dupb(nmp)) != NULL) {
2923dec9fcdSqs 				nmp->b_wptr = nmp->b_rptr +
2933dec9fcdSqs 				    (TX_MAX_TRANSFER_LENGTH -
2943dec9fcdSqs 				    TX_PKT_HEADER_SIZE);
2953dec9fcdSqs 				t_mp->b_rptr = nmp->b_wptr;
2963dec9fcdSqs 				t_mp->b_cont = nmp->b_cont;
2973dec9fcdSqs 				nmp->b_cont = t_mp;
2983dec9fcdSqs 				len = MBLKL(nmp);
2993dec9fcdSqs 			} else {
3003dec9fcdSqs 				good_packet = B_FALSE;
3013dec9fcdSqs 				goto hxge_start_fail2;
3023dec9fcdSqs 			}
3033dec9fcdSqs 		}
3043dec9fcdSqs 		tx_desc.value = 0;
3053dec9fcdSqs 		tx_desc_p = &tx_desc_ring_vp[i];
3063dec9fcdSqs #ifdef	HXGE_DEBUG
3073dec9fcdSqs 		tx_desc_pp = &tx_desc_ring_pp[i];
3083dec9fcdSqs #endif
3093dec9fcdSqs 		tx_msg_p = &tx_msg_ring[i];
3103dec9fcdSqs 		hpi_desc_handle.regp = (uint64_t)tx_desc_p;
3113dec9fcdSqs 		if (!header_set &&
3123dec9fcdSqs 		    ((!hxge_tx_use_bcopy && (len > TX_BCOPY_SIZE)) ||
3133dec9fcdSqs 		    (len >= bcopy_thresh))) {
3143dec9fcdSqs 			header_set = B_TRUE;
3153dec9fcdSqs 			bcopy_thresh += TX_PKT_HEADER_SIZE;
3163dec9fcdSqs 			boff = 0;
3173dec9fcdSqs 			pack_len = 0;
3183dec9fcdSqs 			kaddr = (caddr_t)DMA_COMMON_VPTR(tx_msg_p->buf_dma);
3193dec9fcdSqs 			hdrp = (p_tx_pkt_header_t)kaddr;
3203dec9fcdSqs 			clen = pkt_len;
3213dec9fcdSqs 			dma_handle = tx_msg_p->buf_dma_handle;
3223dec9fcdSqs 			dma_ioaddr = DMA_COMMON_IOADDR(tx_msg_p->buf_dma);
323cf6ef894SMichael Speer 			offset = tx_msg_p->offset_index * hxge_bcopy_thresh;
3243dec9fcdSqs 			(void) ddi_dma_sync(dma_handle,
325cf6ef894SMichael Speer 			    offset, hxge_bcopy_thresh, DDI_DMA_SYNC_FORDEV);
3263dec9fcdSqs 
3273dec9fcdSqs 			tx_msg_p->flags.dma_type = USE_BCOPY;
3283dec9fcdSqs 			goto hxge_start_control_header_only;
3293dec9fcdSqs 		}
3303dec9fcdSqs 
3313dec9fcdSqs 		pkt_len += len;
3323dec9fcdSqs 		pack_len += len;
3333dec9fcdSqs 
3343dec9fcdSqs 		HXGE_DEBUG_MSG((hxgep, TX_CTL,
3353dec9fcdSqs 		    "==> hxge_start(3): desc entry %d DESC IOADDR $%p "
3363dec9fcdSqs 		    "desc_vp $%p tx_desc_p $%p desc_pp $%p tx_desc_pp $%p "
3373dec9fcdSqs 		    "len %d pkt_len %d pack_len %d",
3383dec9fcdSqs 		    i,
3393dec9fcdSqs 		    DMA_COMMON_IOADDR(desc_area),
3403dec9fcdSqs 		    tx_desc_ring_vp, tx_desc_p,
3413dec9fcdSqs 		    tx_desc_ring_pp, tx_desc_pp,
3423dec9fcdSqs 		    len, pkt_len, pack_len));
3433dec9fcdSqs 
3443dec9fcdSqs 		if (len < bcopy_thresh) {
3453dec9fcdSqs 			HXGE_DEBUG_MSG((hxgep, TX_CTL,
3463dec9fcdSqs 			    "==> hxge_start(4): USE BCOPY: "));
3473dec9fcdSqs 			if (hxge_tx_tiny_pack) {
3483dec9fcdSqs 				uint32_t blst = TXDMA_DESC_NEXT_INDEX(i, -1,
3493dec9fcdSqs 				    tx_ring_p->tx_wrap_mask);
3503dec9fcdSqs 				HXGE_DEBUG_MSG((hxgep, TX_CTL,
3513dec9fcdSqs 				    "==> hxge_start(5): pack"));
3523dec9fcdSqs 				if ((pack_len <= bcopy_thresh) &&
3533dec9fcdSqs 				    (last_bidx == blst)) {
3543dec9fcdSqs 					HXGE_DEBUG_MSG((hxgep, TX_CTL,
3553dec9fcdSqs 					    "==> hxge_start: pack(6) "
3563dec9fcdSqs 					    "(pkt_len %d pack_len %d)",
3573dec9fcdSqs 					    pkt_len, pack_len));
3583dec9fcdSqs 					i = blst;
3593dec9fcdSqs 					tx_desc_p = &tx_desc_ring_vp[i];
3603dec9fcdSqs #ifdef	HXGE_DEBUG
3613dec9fcdSqs 					tx_desc_pp = &tx_desc_ring_pp[i];
3623dec9fcdSqs #endif
3633dec9fcdSqs 					tx_msg_p = &tx_msg_ring[i];
3643dec9fcdSqs 					boff = pack_len - len;
3653dec9fcdSqs 					ngathers--;
3663dec9fcdSqs 				} else if (pack_len > bcopy_thresh &&
3673dec9fcdSqs 				    header_set) {
3683dec9fcdSqs 					pack_len = len;
3693dec9fcdSqs 					boff = 0;
3703dec9fcdSqs 					bcopy_thresh = hxge_bcopy_thresh;
3713dec9fcdSqs 					HXGE_DEBUG_MSG((hxgep, TX_CTL,
3723dec9fcdSqs 					    "==> hxge_start(7): > max NEW "
3733dec9fcdSqs 					    "bcopy thresh %d "
3743dec9fcdSqs 					    "pkt_len %d pack_len %d(next)",
3753dec9fcdSqs 					    bcopy_thresh, pkt_len, pack_len));
3763dec9fcdSqs 				}
3773dec9fcdSqs 				last_bidx = i;
3783dec9fcdSqs 			}
3793dec9fcdSqs 			kaddr = (caddr_t)DMA_COMMON_VPTR(tx_msg_p->buf_dma);
3803dec9fcdSqs 			if ((boff == TX_PKT_HEADER_SIZE) && (nmblks == 1)) {
3813dec9fcdSqs 				hdrp = (p_tx_pkt_header_t)kaddr;
3823dec9fcdSqs 				header_set = B_TRUE;
3833dec9fcdSqs 				HXGE_DEBUG_MSG((hxgep, TX_CTL,
3843dec9fcdSqs 				    "==> hxge_start(7_x2): "
3853dec9fcdSqs 				    "pkt_len %d pack_len %d (new hdrp $%p)",
3863dec9fcdSqs 				    pkt_len, pack_len, hdrp));
3873dec9fcdSqs 			}
3883dec9fcdSqs 			tx_msg_p->flags.dma_type = USE_BCOPY;
3893dec9fcdSqs 			kaddr += boff;
3903dec9fcdSqs 			HXGE_DEBUG_MSG((hxgep, TX_CTL,
3913dec9fcdSqs 			    "==> hxge_start(8): USE BCOPY: before bcopy "
3923dec9fcdSqs 			    "DESC IOADDR $%p entry %d bcopy packets %d "
3933dec9fcdSqs 			    "bcopy kaddr $%p bcopy ioaddr (SAD) $%p "
3943dec9fcdSqs 			    "bcopy clen %d bcopy boff %d",
3953dec9fcdSqs 			    DMA_COMMON_IOADDR(desc_area), i,
3963dec9fcdSqs 			    tdc_stats->tx_hdr_pkts, kaddr, dma_ioaddr,
3973dec9fcdSqs 			    clen, boff));
3983dec9fcdSqs 			HXGE_DEBUG_MSG((hxgep, TX_CTL,
3993dec9fcdSqs 			    "==> hxge_start: 1USE BCOPY: "));
4003dec9fcdSqs 			HXGE_DEBUG_MSG((hxgep, TX_CTL,
4013dec9fcdSqs 			    "==> hxge_start: 2USE BCOPY: "));
4023dec9fcdSqs 			HXGE_DEBUG_MSG((hxgep, TX_CTL, "==> hxge_start: "
4033dec9fcdSqs 			    "last USE BCOPY: copy from b_rptr $%p "
4043dec9fcdSqs 			    "to KADDR $%p (len %d offset %d",
4053dec9fcdSqs 			    b_rptr, kaddr, len, boff));
4063dec9fcdSqs 			bcopy(b_rptr, kaddr, len);
4073dec9fcdSqs #ifdef	HXGE_DEBUG
4083dec9fcdSqs 			dump_len = (len > 128) ? 128: len;
4093dec9fcdSqs 			HXGE_DEBUG_MSG((hxgep, TX_CTL,
4103dec9fcdSqs 			    "==> hxge_start: dump packets "
4113dec9fcdSqs 			    "(After BCOPY len %d)"
4123dec9fcdSqs 			    "(b_rptr $%p): %s", len, nmp->b_rptr,
4133dec9fcdSqs 			    hxge_dump_packet((char *)nmp->b_rptr,
4143dec9fcdSqs 			    dump_len)));
4153dec9fcdSqs #endif
4163dec9fcdSqs 			dma_handle = tx_msg_p->buf_dma_handle;
4173dec9fcdSqs 			dma_ioaddr = DMA_COMMON_IOADDR(tx_msg_p->buf_dma);
418cf6ef894SMichael Speer 			offset = tx_msg_p->offset_index * hxge_bcopy_thresh;
4193dec9fcdSqs 			(void) ddi_dma_sync(dma_handle,
420cf6ef894SMichael Speer 			    offset, hxge_bcopy_thresh, DDI_DMA_SYNC_FORDEV);
4213dec9fcdSqs 			clen = len + boff;
4223dec9fcdSqs 			tdc_stats->tx_hdr_pkts++;
4233dec9fcdSqs 			HXGE_DEBUG_MSG((hxgep, TX_CTL, "==> hxge_start(9): "
4243dec9fcdSqs 			    "USE BCOPY: DESC IOADDR $%p entry %d "
4253dec9fcdSqs 			    "bcopy packets %d bcopy kaddr $%p "
4263dec9fcdSqs 			    "bcopy ioaddr (SAD) $%p bcopy clen %d "
4273dec9fcdSqs 			    "bcopy boff %d",
4283dec9fcdSqs 			    DMA_COMMON_IOADDR(desc_area), i,
4293dec9fcdSqs 			    tdc_stats->tx_hdr_pkts, kaddr, dma_ioaddr,
4303dec9fcdSqs 			    clen, boff));
4313dec9fcdSqs 		} else {
4323dec9fcdSqs 			HXGE_DEBUG_MSG((hxgep, TX_CTL,
4333dec9fcdSqs 			    "==> hxge_start(12): USE DVMA: len %d", len));
4343dec9fcdSqs 			tx_msg_p->flags.dma_type = USE_DMA;
4353dec9fcdSqs 			dma_flags = DDI_DMA_WRITE;
4363dec9fcdSqs 			if (len < hxge_dma_stream_thresh) {
4373dec9fcdSqs 				dma_flags |= DDI_DMA_CONSISTENT;
4383dec9fcdSqs 			} else {
4393dec9fcdSqs 				dma_flags |= DDI_DMA_STREAMING;
4403dec9fcdSqs 			}
4413dec9fcdSqs 
4423dec9fcdSqs 			dma_handle = tx_msg_p->dma_handle;
4434d9de298SMichael Speer 			dma_status = ddi_dma_addr_bind_handle(dma_handle, NULL,
4443dec9fcdSqs 			    (caddr_t)b_rptr, len, dma_flags,
4453dec9fcdSqs 			    DDI_DMA_DONTWAIT, NULL,
4463dec9fcdSqs 			    &dma_cookie, &ncookies);
4474d9de298SMichael Speer 			if (dma_status == DDI_DMA_MAPPED) {
4483dec9fcdSqs 				dma_ioaddr = dma_cookie.dmac_laddress;
4493dec9fcdSqs 				len = (int)dma_cookie.dmac_size;
4503dec9fcdSqs 				clen = (uint32_t)dma_cookie.dmac_size;
4513dec9fcdSqs 				HXGE_DEBUG_MSG((hxgep, TX_CTL,
4523dec9fcdSqs 				    "==> hxge_start(12_1): "
4533dec9fcdSqs 				    "USE DVMA: len %d clen %d ngathers %d",
4543dec9fcdSqs 				    len, clen, ngathers));
4553dec9fcdSqs 				hpi_desc_handle.regp = (uint64_t)tx_desc_p;
4563dec9fcdSqs 				while (ncookies > 1) {
4573dec9fcdSqs 					ngathers++;
4583dec9fcdSqs 					/*
4593dec9fcdSqs 					 * this is the fix for multiple
4603dec9fcdSqs 					 * cookies, which are basically
4613dec9fcdSqs 					 * a descriptor entry, we don't set
4623dec9fcdSqs 					 * SOP bit as well as related fields
4633dec9fcdSqs 					 */
4643dec9fcdSqs 
4653dec9fcdSqs 					(void) hpi_txdma_desc_gather_set(
4663dec9fcdSqs 					    hpi_desc_handle, &tx_desc,
4673dec9fcdSqs 					    (ngathers -1), mark_mode,
4683dec9fcdSqs 					    ngathers, dma_ioaddr, clen);
4693dec9fcdSqs 					tx_msg_p->tx_msg_size = clen;
4703dec9fcdSqs 					HXGE_DEBUG_MSG((hxgep, TX_CTL,
4713dec9fcdSqs 					    "==> hxge_start:  DMA "
4723dec9fcdSqs 					    "ncookie %d ngathers %d "
4733dec9fcdSqs 					    "dma_ioaddr $%p len %d"
4743dec9fcdSqs 					    "desc $%p descp $%p (%d)",
4753dec9fcdSqs 					    ncookies, ngathers,
4763dec9fcdSqs 					    dma_ioaddr, clen,
4773dec9fcdSqs 					    *tx_desc_p, tx_desc_p, i));
4783dec9fcdSqs 
4793dec9fcdSqs 					ddi_dma_nextcookie(dma_handle,
4803dec9fcdSqs 					    &dma_cookie);
4813dec9fcdSqs 					dma_ioaddr = dma_cookie.dmac_laddress;
4823dec9fcdSqs 
4833dec9fcdSqs 					len = (int)dma_cookie.dmac_size;
4843dec9fcdSqs 					clen = (uint32_t)dma_cookie.dmac_size;
4853dec9fcdSqs 					HXGE_DEBUG_MSG((hxgep, TX_CTL,
4863dec9fcdSqs 					    "==> hxge_start(12_2): "
4873dec9fcdSqs 					    "USE DVMA: len %d clen %d ",
4883dec9fcdSqs 					    len, clen));
4893dec9fcdSqs 
4903dec9fcdSqs 					i = TXDMA_DESC_NEXT_INDEX(i, 1,
4913dec9fcdSqs 					    tx_ring_p->tx_wrap_mask);
4923dec9fcdSqs 					tx_desc_p = &tx_desc_ring_vp[i];
4933dec9fcdSqs 
4943dec9fcdSqs 					hpi_desc_handle.regp =
495*86ef0a63SRichard Lowe 					    (uint64_t)tx_desc_p;
4963dec9fcdSqs 					tx_msg_p = &tx_msg_ring[i];
4973dec9fcdSqs 					tx_msg_p->flags.dma_type = USE_NONE;
4983dec9fcdSqs 					tx_desc.value = 0;
4993dec9fcdSqs 					ncookies--;
5003dec9fcdSqs 				}
5013dec9fcdSqs 				tdc_stats->tx_ddi_pkts++;
5023dec9fcdSqs 				HXGE_DEBUG_MSG((hxgep, TX_CTL,
5033dec9fcdSqs 				    "==> hxge_start: DMA: ddi packets %d",
5043dec9fcdSqs 				    tdc_stats->tx_ddi_pkts));
5053dec9fcdSqs 			} else {
5063dec9fcdSqs 				HXGE_ERROR_MSG((hxgep, HXGE_ERR_CTL,
5073dec9fcdSqs 				    "dma mapping failed for %d "
5083dec9fcdSqs 				    "bytes addr $%p flags %x (%d)",
5093dec9fcdSqs 				    len, b_rptr, status, status));
5103dec9fcdSqs 				good_packet = B_FALSE;
5113dec9fcdSqs 				tdc_stats->tx_dma_bind_fail++;
5123dec9fcdSqs 				tx_msg_p->flags.dma_type = USE_NONE;
5134d9de298SMichael Speer 				status = 1;
5143dec9fcdSqs 				goto hxge_start_fail2;
5153dec9fcdSqs 			}
5163dec9fcdSqs 		} /* ddi dvma */
5173dec9fcdSqs 
5183dec9fcdSqs 		nmp = nmp->b_cont;
5193dec9fcdSqs hxge_start_control_header_only:
5203dec9fcdSqs 		hpi_desc_handle.regp = (uint64_t)tx_desc_p;
5213dec9fcdSqs 		ngathers++;
5223dec9fcdSqs 
5233dec9fcdSqs 		if (ngathers == 1) {
5243dec9fcdSqs #ifdef	HXGE_DEBUG
5253dec9fcdSqs 			save_desc_p = &sop_tx_desc;
5263dec9fcdSqs #endif
5273dec9fcdSqs 			sop_tx_desc_p = &sop_tx_desc;
5283dec9fcdSqs 			sop_tx_desc_p->value = 0;
5293dec9fcdSqs 			sop_tx_desc_p->bits.tr_len = clen;
530fe930412Sqs 			sop_tx_desc_p->bits.sad = dma_ioaddr >> 32;
531fe930412Sqs 			sop_tx_desc_p->bits.sad_l = dma_ioaddr & 0xffffffff;
5323dec9fcdSqs 		} else {
5333dec9fcdSqs #ifdef	HXGE_DEBUG
5343dec9fcdSqs 			save_desc_p = &tx_desc;
5353dec9fcdSqs #endif
5363dec9fcdSqs 			tmp_desc_p = &tx_desc;
5373dec9fcdSqs 			tmp_desc_p->value = 0;
5383dec9fcdSqs 			tmp_desc_p->bits.tr_len = clen;
539fe930412Sqs 			tmp_desc_p->bits.sad = dma_ioaddr >> 32;
540fe930412Sqs 			tmp_desc_p->bits.sad_l = dma_ioaddr & 0xffffffff;
5413dec9fcdSqs 
5423dec9fcdSqs 			tx_desc_p->value = tmp_desc_p->value;
5433dec9fcdSqs 		}
5443dec9fcdSqs 
5453dec9fcdSqs 		HXGE_DEBUG_MSG((hxgep, TX_CTL,
5463dec9fcdSqs 		    "==> hxge_start(13): Desc_entry %d ngathers %d "
5473dec9fcdSqs 		    "desc_vp $%p tx_desc_p $%p "
5483dec9fcdSqs 		    "len %d clen %d pkt_len %d pack_len %d nmblks %d "
5493dec9fcdSqs 		    "dma_ioaddr (SAD) $%p mark %d",
5503dec9fcdSqs 		    i, ngathers, tx_desc_ring_vp, tx_desc_p,
5513dec9fcdSqs 		    len, clen, pkt_len, pack_len, nmblks,
5523dec9fcdSqs 		    dma_ioaddr, mark_mode));
5533dec9fcdSqs 
5543dec9fcdSqs #ifdef HXGE_DEBUG
5553dec9fcdSqs 		hpi_desc_handle.hxgep = hxgep;
5563dec9fcdSqs 		hpi_desc_handle.function.function = 0;
5573dec9fcdSqs 		hpi_desc_handle.function.instance = hxgep->instance;
5583dec9fcdSqs 		sad = save_desc_p->bits.sad;
559fe930412Sqs 		sad = (sad << 32) | save_desc_p->bits.sad_l;
5603dec9fcdSqs 		xfer_len = save_desc_p->bits.tr_len;
5613dec9fcdSqs 
5623dec9fcdSqs 		HXGE_DEBUG_MSG((hxgep, TX_CTL, "\n\t: value 0x%llx\n"
5633dec9fcdSqs 		    "\t\tsad $%p\ttr_len %d len %d\tnptrs %d\t"
5643dec9fcdSqs 		    "mark %d sop %d\n",
5653dec9fcdSqs 		    save_desc_p->value, sad, save_desc_p->bits.tr_len,
5663dec9fcdSqs 		    xfer_len, save_desc_p->bits.num_ptr,
5673dec9fcdSqs 		    save_desc_p->bits.mark, save_desc_p->bits.sop));
5683dec9fcdSqs 
5693dec9fcdSqs 		hpi_txdma_dump_desc_one(hpi_desc_handle, NULL, i);
5703dec9fcdSqs #endif
5713dec9fcdSqs 
5723dec9fcdSqs 		tx_msg_p->tx_msg_size = clen;
5733dec9fcdSqs 		i = TXDMA_DESC_NEXT_INDEX(i, 1, tx_ring_p->tx_wrap_mask);
5743dec9fcdSqs 		if (ngathers > hxge_tx_max_gathers) {
5753dec9fcdSqs 			good_packet = B_FALSE;
5760dc2366fSVenugopal Iyer 			mac_hcksum_get(mp, &start_offset, &stuff_offset,
5770dc2366fSVenugopal Iyer 			    &end_offset, &value, &cksum_flags);
5783dec9fcdSqs 
5793dec9fcdSqs 			HXGE_DEBUG_MSG((NULL, TX_CTL,
5803dec9fcdSqs 			    "==> hxge_start(14): pull msg - "
5813dec9fcdSqs 			    "len %d pkt_len %d ngathers %d",
5823dec9fcdSqs 			    len, pkt_len, ngathers));
5833dec9fcdSqs 			goto hxge_start_fail2;
5843dec9fcdSqs 		}
5853dec9fcdSqs 	} /* while (nmp) */
5863dec9fcdSqs 
5873dec9fcdSqs 	tx_msg_p->tx_message = mp;
5883dec9fcdSqs 	tx_desc_p = &tx_desc_ring_vp[sop_index];
5893dec9fcdSqs 	hpi_desc_handle.regp = (uint64_t)tx_desc_p;
5903dec9fcdSqs 
5913dec9fcdSqs 	pkthdrp = (p_tx_pkt_hdr_all_t)hdrp;
5923dec9fcdSqs 	pkthdrp->reserved = 0;
5933dec9fcdSqs 	hdrp->value = 0;
5943dec9fcdSqs 	(void) hxge_fill_tx_hdr(mp, B_FALSE, cksum_on,
5953dec9fcdSqs 	    (pkt_len - TX_PKT_HEADER_SIZE), npads, pkthdrp);
596fe930412Sqs 
597fe930412Sqs 	/*
598fe930412Sqs 	 * Hardware header should not be counted as part of the frame
599fe930412Sqs 	 * when determining the frame size
600fe930412Sqs 	 */
601fe930412Sqs 	if ((pkt_len - TX_PKT_HEADER_SIZE) > (STD_FRAME_SIZE - ETHERFCSL)) {
6023dec9fcdSqs 		tdc_stats->tx_jumbo_pkts++;
6033dec9fcdSqs 	}
6043dec9fcdSqs 
6053dec9fcdSqs 	min_len = (hxgep->msg_min + TX_PKT_HEADER_SIZE + (npads * 2));
6063dec9fcdSqs 	if (pkt_len < min_len) {
6073dec9fcdSqs 		/* Assume we use bcopy to premapped buffers */
6083dec9fcdSqs 		kaddr = (caddr_t)DMA_COMMON_VPTR(tx_msg_p->buf_dma);
6093dec9fcdSqs 		HXGE_DEBUG_MSG((NULL, TX_CTL,
6103dec9fcdSqs 		    "==> hxge_start(14-1): < (msg_min + 16)"
6113dec9fcdSqs 		    "len %d pkt_len %d min_len %d bzero %d ngathers %d",
6123dec9fcdSqs 		    len, pkt_len, min_len, (min_len - pkt_len), ngathers));
6133dec9fcdSqs 		bzero((kaddr + pkt_len), (min_len - pkt_len));
6143dec9fcdSqs 		pkt_len = tx_msg_p->tx_msg_size = min_len;
6153dec9fcdSqs 
6163dec9fcdSqs 		sop_tx_desc_p->bits.tr_len = min_len;
6173dec9fcdSqs 
6183dec9fcdSqs 		HXGE_MEM_PIO_WRITE64(hpi_desc_handle, sop_tx_desc_p->value);
6193dec9fcdSqs 		tx_desc_p->value = sop_tx_desc_p->value;
6203dec9fcdSqs 
6213dec9fcdSqs 		HXGE_DEBUG_MSG((NULL, TX_CTL,
6223dec9fcdSqs 		    "==> hxge_start(14-2): < msg_min - "
6233dec9fcdSqs 		    "len %d pkt_len %d min_len %d ngathers %d",
6243dec9fcdSqs 		    len, pkt_len, min_len, ngathers));
6253dec9fcdSqs 	}
6263dec9fcdSqs 
6273dec9fcdSqs 	HXGE_DEBUG_MSG((hxgep, TX_CTL, "==> hxge_start: cksum_flags 0x%x ",
6283dec9fcdSqs 	    cksum_flags));
6293dec9fcdSqs 	if (cksum_flags & HCK_PARTIALCKSUM) {
6303dec9fcdSqs 		HXGE_DEBUG_MSG((hxgep, TX_CTL,
6313dec9fcdSqs 		    "==> hxge_start: cksum_flags 0x%x (partial checksum) ",
6323dec9fcdSqs 		    cksum_flags));
6333dec9fcdSqs 		cksum_on = B_TRUE;
6343dec9fcdSqs 		HXGE_DEBUG_MSG((hxgep, TX_CTL,
6353dec9fcdSqs 		    "==> hxge_start: from IP cksum_flags 0x%x "
6363dec9fcdSqs 		    "(partial checksum) "
6373dec9fcdSqs 		    "start_offset %d stuff_offset %d",
6383dec9fcdSqs 		    cksum_flags, start_offset, stuff_offset));
6393dec9fcdSqs 		tmp_len = (uint64_t)(start_offset >> 1);
6403dec9fcdSqs 		hdrp->value |= (tmp_len << TX_PKT_HEADER_L4START_SHIFT);
6413dec9fcdSqs 		tmp_len = (uint64_t)(stuff_offset >> 1);
6423dec9fcdSqs 		hdrp->value |= (tmp_len << TX_PKT_HEADER_L4STUFF_SHIFT);
6433dec9fcdSqs 
6443dec9fcdSqs 		HXGE_DEBUG_MSG((hxgep, TX_CTL,
6453dec9fcdSqs 		    "==> hxge_start: from IP cksum_flags 0x%x "
6463dec9fcdSqs 		    "(partial checksum) "
6473dec9fcdSqs 		    "after SHIFT start_offset %d stuff_offset %d",
6483dec9fcdSqs 		    cksum_flags, start_offset, stuff_offset));
6493dec9fcdSqs 	}
6503dec9fcdSqs 
6513dec9fcdSqs 	/*
6523dec9fcdSqs 	 * pkt_len already includes 16 + paddings!!
6533dec9fcdSqs 	 * Update the control header length
6543dec9fcdSqs 	 */
6553dec9fcdSqs 
6563dec9fcdSqs 	/*
6573dec9fcdSqs 	 * Note that Hydra is different from Neptune where
6583dec9fcdSqs 	 * tot_xfer_len = (pkt_len - TX_PKT_HEADER_SIZE);
6593dec9fcdSqs 	 */
6603dec9fcdSqs 	tot_xfer_len = pkt_len;
6613dec9fcdSqs 	tmp_len = hdrp->value |
6623dec9fcdSqs 	    (tot_xfer_len << TX_PKT_HEADER_TOT_XFER_LEN_SHIFT);
6633dec9fcdSqs 
6643dec9fcdSqs 	HXGE_DEBUG_MSG((hxgep, TX_CTL,
6653dec9fcdSqs 	    "==> hxge_start(15_x1): setting SOP "
6663dec9fcdSqs 	    "tot_xfer_len 0x%llx (%d) pkt_len %d tmp_len "
6673dec9fcdSqs 	    "0x%llx hdrp->value 0x%llx",
6683dec9fcdSqs 	    tot_xfer_len, tot_xfer_len, pkt_len, tmp_len, hdrp->value));
6693dec9fcdSqs #if defined(_BIG_ENDIAN)
6703dec9fcdSqs 	hdrp->value = ddi_swap64(tmp_len);
6713dec9fcdSqs #else
6723dec9fcdSqs 	hdrp->value = tmp_len;
6733dec9fcdSqs #endif
6743dec9fcdSqs 	HXGE_DEBUG_MSG((hxgep,
6753dec9fcdSqs 	    TX_CTL, "==> hxge_start(15_x2): setting SOP "
6763dec9fcdSqs 	    "after SWAP: tot_xfer_len 0x%llx pkt_len %d "
6773dec9fcdSqs 	    "tmp_len 0x%llx hdrp->value 0x%llx",
6783dec9fcdSqs 	    tot_xfer_len, pkt_len, tmp_len, hdrp->value));
6793dec9fcdSqs 
6803dec9fcdSqs 	HXGE_DEBUG_MSG((hxgep, TX_CTL, "==> hxge_start(15): setting SOP "
6813dec9fcdSqs 	    "wr_index %d tot_xfer_len (%d) pkt_len %d npads %d",
6823dec9fcdSqs 	    sop_index, tot_xfer_len, pkt_len, npads));
6833dec9fcdSqs 
6843dec9fcdSqs 	sop_tx_desc_p->bits.sop = 1;
6853dec9fcdSqs 	sop_tx_desc_p->bits.mark = mark_mode;
6863dec9fcdSqs 	sop_tx_desc_p->bits.num_ptr = ngathers;
6873dec9fcdSqs 
6883dec9fcdSqs 	if (mark_mode)
6893dec9fcdSqs 		tdc_stats->tx_marks++;
6903dec9fcdSqs 
6913dec9fcdSqs 	HXGE_MEM_PIO_WRITE64(hpi_desc_handle, sop_tx_desc_p->value);
6923dec9fcdSqs 	HXGE_DEBUG_MSG((hxgep, TX_CTL, "==> hxge_start(16): set SOP done"));
6933dec9fcdSqs 
6943dec9fcdSqs #ifdef HXGE_DEBUG
6953dec9fcdSqs 	hpi_desc_handle.hxgep = hxgep;
6963dec9fcdSqs 	hpi_desc_handle.function.function = 0;
6973dec9fcdSqs 	hpi_desc_handle.function.instance = hxgep->instance;
6983dec9fcdSqs 
6993dec9fcdSqs 	HXGE_DEBUG_MSG((hxgep, TX_CTL, "\n\t: value 0x%llx\n"
7003dec9fcdSqs 	    "\t\tsad $%p\ttr_len %d len %d\tnptrs %d\tmark %d sop %d\n",
7013dec9fcdSqs 	    save_desc_p->value, sad, save_desc_p->bits.tr_len,
7023dec9fcdSqs 	    xfer_len, save_desc_p->bits.num_ptr, save_desc_p->bits.mark,
7033dec9fcdSqs 	    save_desc_p->bits.sop));
7043dec9fcdSqs 	(void) hpi_txdma_dump_desc_one(hpi_desc_handle, NULL, sop_index);
7053dec9fcdSqs 
7063dec9fcdSqs 	dump_len = (pkt_len > 128) ? 128: pkt_len;
7073dec9fcdSqs 	HXGE_DEBUG_MSG((hxgep, TX_CTL,
7083dec9fcdSqs 	    "==> hxge_start: dump packets(17) (after sop set, len "
7093dec9fcdSqs 	    " (len/dump_len/pkt_len/tot_xfer_len) %d/%d/%d/%d):\n"
7103dec9fcdSqs 	    "ptr $%p: %s", len, dump_len, pkt_len, tot_xfer_len,
7113dec9fcdSqs 	    (char *)hdrp, hxge_dump_packet((char *)hdrp, dump_len)));
7123dec9fcdSqs 	HXGE_DEBUG_MSG((hxgep, TX_CTL,
7133dec9fcdSqs 	    "==> hxge_start(18): TX desc sync: sop_index %d", sop_index));
7143dec9fcdSqs #endif
7153dec9fcdSqs 
7163dec9fcdSqs 	if ((ngathers == 1) || tx_ring_p->wr_index < i) {
7173dec9fcdSqs 		(void) ddi_dma_sync(tx_desc_dma_handle,
7183dec9fcdSqs 		    sop_index * sizeof (tx_desc_t),
7193dec9fcdSqs 		    ngathers * sizeof (tx_desc_t), DDI_DMA_SYNC_FORDEV);
7203dec9fcdSqs 
7213dec9fcdSqs 		HXGE_DEBUG_MSG((hxgep, TX_CTL, "hxge_start(19): sync 1 "
7223dec9fcdSqs 		    "cs_off = 0x%02X cs_s_off = 0x%02X "
7233dec9fcdSqs 		    "pkt_len %d ngathers %d sop_index %d\n",
7243dec9fcdSqs 		    stuff_offset, start_offset,
7253dec9fcdSqs 		    pkt_len, ngathers, sop_index));
7263dec9fcdSqs 	} else { /* more than one descriptor and wrap around */
7273dec9fcdSqs 		uint32_t nsdescs = tx_ring_p->tx_ring_size - sop_index;
7283dec9fcdSqs 		(void) ddi_dma_sync(tx_desc_dma_handle,
7293dec9fcdSqs 		    sop_index * sizeof (tx_desc_t),
7303dec9fcdSqs 		    nsdescs * sizeof (tx_desc_t), DDI_DMA_SYNC_FORDEV);
7313dec9fcdSqs 		HXGE_DEBUG_MSG((hxgep, TX_CTL, "hxge_start(20): sync 1 "
7323dec9fcdSqs 		    "cs_off = 0x%02X cs_s_off = 0x%02X "
7333dec9fcdSqs 		    "pkt_len %d ngathers %d sop_index %d\n",
7343dec9fcdSqs 		    stuff_offset, start_offset, pkt_len, ngathers, sop_index));
7353dec9fcdSqs 
7363dec9fcdSqs 		(void) ddi_dma_sync(tx_desc_dma_handle, 0,
7373dec9fcdSqs 		    (ngathers - nsdescs) * sizeof (tx_desc_t),
7383dec9fcdSqs 		    DDI_DMA_SYNC_FORDEV);
7393dec9fcdSqs 		HXGE_DEBUG_MSG((hxgep, TX_CTL, "hxge_start(21): sync 2 "
7403dec9fcdSqs 		    "cs_off = 0x%02X cs_s_off = 0x%02X "
7413dec9fcdSqs 		    "pkt_len %d ngathers %d sop_index %d\n",
7423dec9fcdSqs 		    stuff_offset, start_offset,
7433dec9fcdSqs 		    pkt_len, ngathers, sop_index));
7443dec9fcdSqs 	}
7453dec9fcdSqs 
7463dec9fcdSqs 	tail_index = tx_ring_p->wr_index;
7473dec9fcdSqs 	tail_wrap = tx_ring_p->wr_index_wrap;
7483dec9fcdSqs 
7493dec9fcdSqs 	tx_ring_p->wr_index = i;
7503dec9fcdSqs 	if (tx_ring_p->wr_index <= tail_index) {
7513dec9fcdSqs 		tx_ring_p->wr_index_wrap = ((tail_wrap == B_TRUE) ?
7523dec9fcdSqs 		    B_FALSE : B_TRUE);
7533dec9fcdSqs 	}
7543dec9fcdSqs 
7553dec9fcdSqs 	tx_ring_p->descs_pending += ngathers;
7563dec9fcdSqs 	HXGE_DEBUG_MSG((hxgep, TX_CTL, "==> hxge_start: TX kick: "
7573dec9fcdSqs 	    "channel %d wr_index %d wrap %d ngathers %d desc_pend %d",
7583dec9fcdSqs 	    tx_ring_p->tdc, tx_ring_p->wr_index, tx_ring_p->wr_index_wrap,
7593dec9fcdSqs 	    ngathers, tx_ring_p->descs_pending));
7603dec9fcdSqs 	HXGE_DEBUG_MSG((hxgep, TX_CTL, "==> hxge_start: TX KICKING: "));
7613dec9fcdSqs 
7623dec9fcdSqs 	kick.value = 0;
7633dec9fcdSqs 	kick.bits.wrap = tx_ring_p->wr_index_wrap;
7643dec9fcdSqs 	kick.bits.tail = (uint16_t)tx_ring_p->wr_index;
7653dec9fcdSqs 
7663dec9fcdSqs 	/* Kick start the Transmit kick register */
7673dec9fcdSqs 	TXDMA_REG_WRITE64(HXGE_DEV_HPI_HANDLE(hxgep),
7683dec9fcdSqs 	    TDC_TDR_KICK, (uint8_t)tx_ring_p->tdc, kick.value);
7693dec9fcdSqs 	tdc_stats->tx_starts++;
7703dec9fcdSqs 	MUTEX_EXIT(&tx_ring_p->lock);
7713dec9fcdSqs 	HXGE_DEBUG_MSG((hxgep, TX_CTL, "<== hxge_start"));
7723dec9fcdSqs 	return (status);
7733dec9fcdSqs 
7743dec9fcdSqs hxge_start_fail2:
7753dec9fcdSqs 	if (good_packet == B_FALSE) {
7763dec9fcdSqs 		cur_index = sop_index;
7773dec9fcdSqs 		HXGE_DEBUG_MSG((hxgep, TX_CTL, "==> hxge_start: clean up"));
7783dec9fcdSqs 		for (i = 0; i < ngathers; i++) {
7793dec9fcdSqs 			tx_desc_p = &tx_desc_ring_vp[cur_index];
7803dec9fcdSqs 			hpi_handle.regp = (uint64_t)tx_desc_p;
7813dec9fcdSqs 			tx_msg_p = &tx_msg_ring[cur_index];
7823dec9fcdSqs 			(void) hpi_txdma_desc_set_zero(hpi_handle, 1);
7833dec9fcdSqs 			if (tx_msg_p->flags.dma_type == USE_DVMA) {
7843dec9fcdSqs 				HXGE_DEBUG_MSG((hxgep, TX_CTL,
7853dec9fcdSqs 				    "tx_desc_p = %X index = %d",
7863dec9fcdSqs 				    tx_desc_p, tx_ring_p->rd_index));
7873dec9fcdSqs 				(void) dvma_unload(tx_msg_p->dvma_handle,
7883dec9fcdSqs 				    0, -1);
7893dec9fcdSqs 				tx_msg_p->dvma_handle = NULL;
7903dec9fcdSqs 				if (tx_ring_p->dvma_wr_index ==
7913dec9fcdSqs 				    tx_ring_p->dvma_wrap_mask)
7923dec9fcdSqs 					tx_ring_p->dvma_wr_index = 0;
7933dec9fcdSqs 				else
7943dec9fcdSqs 					tx_ring_p->dvma_wr_index++;
7953dec9fcdSqs 				tx_ring_p->dvma_pending--;
7963dec9fcdSqs 			} else if (tx_msg_p->flags.dma_type == USE_DMA) {
7973dec9fcdSqs 				if (ddi_dma_unbind_handle(
7983dec9fcdSqs 				    tx_msg_p->dma_handle)) {
7993dec9fcdSqs 					cmn_err(CE_WARN, "hxge_start: "
8003dec9fcdSqs 					    "ddi_dma_unbind_handle failed");
8013dec9fcdSqs 				}
8023dec9fcdSqs 			}
8033dec9fcdSqs 			tx_msg_p->flags.dma_type = USE_NONE;
8043dec9fcdSqs 			cur_index = TXDMA_DESC_NEXT_INDEX(cur_index, 1,
8053dec9fcdSqs 			    tx_ring_p->tx_wrap_mask);
8063dec9fcdSqs 
8073dec9fcdSqs 		}
8083dec9fcdSqs 	}
8093dec9fcdSqs 
8103dec9fcdSqs 	MUTEX_EXIT(&tx_ring_p->lock);
8113dec9fcdSqs 
8123dec9fcdSqs hxge_start_fail1:
8133dec9fcdSqs 	/* Add FMA to check the access handle hxge_hregh */
8143dec9fcdSqs 	HXGE_DEBUG_MSG((hxgep, TX_CTL, "<== hxge_start"));
8153dec9fcdSqs 	return (status);
8163dec9fcdSqs }
817