1 /*
2  * CDDL HEADER START
3  *
4  * The contents of this file are subject to the terms of the
5  * Common Development and Distribution License, Version 1.0 only
6  * (the "License").  You may not use this file except in compliance
7  * with the License.
8  *
9  * You can obtain a copy of the license at usr/src/OPENSOLARIS.LICENSE
10  * or http://www.opensolaris.org/os/licensing.
11  * See the License for the specific language governing permissions
12  * and limitations under the License.
13  *
14  * When distributing Covered Code, include this CDDL HEADER in each
15  * file and include the License file at usr/src/OPENSOLARIS.LICENSE.
16  * If applicable, add the following below this CDDL HEADER, with the
17  * fields enclosed by brackets "[]" replaced with your own identifying
18  * information: Portions Copyright [yyyy] [name of copyright owner]
19  *
20  * CDDL HEADER END
21  */
22 /*
23  * Copyright 2005 Sun Microsystems, Inc.  All rights reserved.
24  * Use is subject to license terms.
25  */
26 
27 #pragma ident	"%Z%%M%	%I%	%E% SMI"
28 
29 /*
30  * hci1394_q.c
31  *    This code decouples some of the OpenHCI async descriptor logic/structures
32  *    from the async processing.  The goal was to combine as much of the
33  *    duplicate code as possible for the different type of async transfers
34  *    without going too overboard.
35  *
36  *    There are two parts to the Q, the descriptor buffer and the data buffer.
37  *    For the most part, data to be transmitted and data which is received go
38  *    in the data buffers.  The information of where to get the data and put
39  *    the data reside in the descriptor buffers. There are exceptions to this.
40  */
41 
42 
43 #include <sys/types.h>
44 #include <sys/conf.h>
45 #include <sys/ddi.h>
46 #include <sys/modctl.h>
47 #include <sys/stat.h>
48 #include <sys/sunddi.h>
49 #include <sys/cmn_err.h>
50 #include <sys/kmem.h>
51 #include <sys/note.h>
52 
53 #include <sys/1394/adapters/hci1394.h>
54 
55 
56 static int hci1394_q_reserve(hci1394_q_buf_t *qbuf, uint_t size,
57     uint32_t *io_addr);
58 static void hci1394_q_unreserve(hci1394_q_buf_t *qbuf);
59 static void hci1394_q_buf_setup(hci1394_q_buf_t *qbuf);
60 static void hci1394_q_reset(hci1394_q_handle_t q_handle);
61 static void hci1394_q_next_buf(hci1394_q_buf_t *qbuf);
62 
63 static void hci1394_q_at_write_OLI(hci1394_q_handle_t q_handle,
64     hci1394_q_buf_t *qbuf, hci1394_q_cmd_t *cmd, hci1394_basic_pkt_t *hdr,
65     uint_t hdrsize);
66 static void hci1394_q_at_write_OMI(hci1394_q_handle_t q_handle,
67     hci1394_q_buf_t *qbuf, hci1394_q_cmd_t *cmd, hci1394_basic_pkt_t *hdr,
68     uint_t hdrsize);
69 static void hci1394_q_at_write_OL(hci1394_q_handle_t q_handle,
70     hci1394_q_buf_t *qbuf, hci1394_q_cmd_t *cmd, uint32_t io_addr,
71     uint_t datasize);
72 static void hci1394_q_at_rep_put8(hci1394_q_buf_t *qbuf, hci1394_q_cmd_t *cmd,
73     uint8_t *data, uint_t datasize);
74 static void hci1394_q_at_copy_from_mblk(hci1394_q_buf_t *qbuf,
75     hci1394_q_cmd_t *cmd, h1394_mblk_t *mblk);
76 
77 static void hci1394_q_ar_write_IM(hci1394_q_handle_t q_handle,
78     hci1394_q_buf_t *qbuf, uint32_t io_addr, uint_t datasize);
79 
80 _NOTE(SCHEME_PROTECTS_DATA("unique", msgb))
81 
82 /*
83  * hci1394_q_init()
84  *    Initialize a Q.  A Q consists of a descriptor buffer and a data buffer and
85  *    can be either an AT or AR Q. hci1394_q_init() returns a handle which
86  *    should be used for the reset of the hci1394_q_* calls.
87  */
88 int
89 hci1394_q_init(hci1394_drvinfo_t *drvinfo,
90     hci1394_ohci_handle_t ohci_handle, hci1394_q_info_t *qinfo,
91     hci1394_q_handle_t *q_handle)
92 {
93 	hci1394_q_buf_t *desc;
94 	hci1394_q_buf_t *data;
95 	hci1394_buf_parms_t parms;
96 	hci1394_q_t *q;
97 	int status;
98 	int index;
99 
100 
101 	ASSERT(drvinfo != NULL);
102 	ASSERT(qinfo != NULL);
103 	ASSERT(q_handle != NULL);
104 	TNF_PROBE_0_DEBUG(hci1394_q_init_enter, HCI1394_TNF_HAL_STACK, "");
105 
106 	/*
107 	 * allocate the memory to track this Q.  Initialize the internal Q
108 	 * structure.
109 	 */
110 	q = kmem_alloc(sizeof (hci1394_q_t), KM_SLEEP);
111 	q->q_drvinfo = drvinfo;
112 	q->q_info = *qinfo;
113 	q->q_ohci = ohci_handle;
114 	mutex_init(&q->q_mutex, NULL, MUTEX_DRIVER, drvinfo->di_iblock_cookie);
115 	desc = &q->q_desc;
116 	data = &q->q_data;
117 
118 	/*
119 	 * Allocate the Descriptor buffer.
120 	 *
121 	 * XXX - Only want 1 cookie for now. Change this to OHCI_MAX_COOKIE
122 	 * after we have tested the multiple cookie code on x86.
123 	 */
124 	parms.bp_length = qinfo->qi_desc_size;
125 	parms.bp_max_cookies = 1;
126 	parms.bp_alignment = 16;
127 	status = hci1394_buf_alloc(drvinfo, &parms, &desc->qb_buf,
128 	    &desc->qb_buf_handle);
129 	if (status != DDI_SUCCESS) {
130 		mutex_destroy(&q->q_mutex);
131 		kmem_free(q, sizeof (hci1394_q_t));
132 		*q_handle = NULL;
133 		TNF_PROBE_0(hci1394_q_init_bae_fail, HCI1394_TNF_HAL_ERROR, "");
134 		TNF_PROBE_0_DEBUG(hci1394_q_init_exit, HCI1394_TNF_HAL_STACK,
135 		    "");
136 		return (DDI_FAILURE);
137 	}
138 
139 	/* Copy in buffer cookies into our local cookie array */
140 	desc->qb_cookie[0] = desc->qb_buf.bi_cookie;
141 	for (index = 1; index < desc->qb_buf.bi_cookie_count; index++) {
142 		ddi_dma_nextcookie(desc->qb_buf.bi_dma_handle,
143 		    &desc->qb_buf.bi_cookie);
144 		desc->qb_cookie[index] = desc->qb_buf.bi_cookie;
145 	}
146 
147 	/*
148 	 * Allocate the Data buffer.
149 	 *
150 	 * XXX - Only want 1 cookie for now. Change this to OHCI_MAX_COOKIE
151 	 * after we have tested the multiple cookie code on x86.
152 	 */
153 	parms.bp_length = qinfo->qi_data_size;
154 	parms.bp_max_cookies = 1;
155 	parms.bp_alignment = 16;
156 	status = hci1394_buf_alloc(drvinfo, &parms, &data->qb_buf,
157 	    &data->qb_buf_handle);
158 	if (status != DDI_SUCCESS) {
159 		/* Free the allocated Descriptor buffer */
160 		hci1394_buf_free(&desc->qb_buf_handle);
161 
162 		mutex_destroy(&q->q_mutex);
163 		kmem_free(q, sizeof (hci1394_q_t));
164 		*q_handle = NULL;
165 		TNF_PROBE_0(hci1394_q_init_baa_fail, HCI1394_TNF_HAL_ERROR, "");
166 		TNF_PROBE_0_DEBUG(hci1394_q_init_exit, HCI1394_TNF_HAL_STACK,
167 		    "");
168 		return (DDI_FAILURE);
169 	}
170 
171 	/*
172 	 * We must have at least 2 ARQ data buffers, If we only have one, we
173 	 * will artificially create 2. We must have 2 so that we always have a
174 	 * descriptor with free data space to write AR data to. When one is
175 	 * empty, it will take us a bit to get a new descriptor back into the
176 	 * chain.
177 	 */
178 	if ((qinfo->qi_mode == HCI1394_ARQ) &&
179 	    (data->qb_buf.bi_cookie_count == 1)) {
180 		data->qb_buf.bi_cookie_count = 2;
181 		data->qb_cookie[0] = data->qb_buf.bi_cookie;
182 		data->qb_cookie[0].dmac_size /= 2;
183 		data->qb_cookie[1] = data->qb_cookie[0];
184 		data->qb_cookie[1].dmac_laddress =
185 		    data->qb_cookie[0].dmac_laddress +
186 		    data->qb_cookie[0].dmac_size;
187 		data->qb_cookie[1].dmac_address =
188 		    data->qb_cookie[0].dmac_address +
189 		    data->qb_cookie[0].dmac_size;
190 
191 	/* We have more than 1 cookie or we are an AT Q */
192 	} else {
193 		/* Copy in buffer cookies into our local cookie array */
194 		data->qb_cookie[0] = data->qb_buf.bi_cookie;
195 		for (index = 1; index < data->qb_buf.bi_cookie_count; index++) {
196 			ddi_dma_nextcookie(data->qb_buf.bi_dma_handle,
197 			    &data->qb_buf.bi_cookie);
198 			data->qb_cookie[index] = data->qb_buf.bi_cookie;
199 		}
200 	}
201 
202 	/* The top and bottom of the Q are only set once */
203 	desc->qb_ptrs.qp_top = desc->qb_buf.bi_kaddr;
204 	desc->qb_ptrs.qp_bottom = desc->qb_buf.bi_kaddr +
205 	    desc->qb_buf.bi_real_length - 1;
206 	data->qb_ptrs.qp_top = data->qb_buf.bi_kaddr;
207 	data->qb_ptrs.qp_bottom = data->qb_buf.bi_kaddr +
208 	    data->qb_buf.bi_real_length - 1;
209 
210 	/*
211 	 * reset the Q pointers to their original settings.  Setup IM
212 	 * descriptors if this is an AR Q.
213 	 */
214 	hci1394_q_reset(q);
215 
216 	/* if this is an AT Q, create a queued list for the AT descriptors */
217 	if (qinfo->qi_mode == HCI1394_ATQ) {
218 		hci1394_tlist_init(drvinfo, NULL, &q->q_queued_list);
219 	}
220 
221 	*q_handle = q;
222 
223 	TNF_PROBE_0_DEBUG(hci1394_q_init_exit, HCI1394_TNF_HAL_STACK, "");
224 
225 	return (DDI_SUCCESS);
226 }
227 
228 
229 /*
230  * hci1394_q_fini()
231  *    Cleanup after a successful hci1394_q_init(). Notice that a pointer to the
232  *    handle is used for the parameter.  fini() will set your handle to NULL
233  *    before returning.
234  */
235 void
236 hci1394_q_fini(hci1394_q_handle_t *q_handle)
237 {
238 	hci1394_q_t *q;
239 
240 	ASSERT(q_handle != NULL);
241 	TNF_PROBE_0_DEBUG(hci1394_q_fini_enter, HCI1394_TNF_HAL_STACK, "");
242 
243 	q = *q_handle;
244 	if (q->q_info.qi_mode == HCI1394_ATQ) {
245 		hci1394_tlist_fini(&q->q_queued_list);
246 	}
247 	mutex_destroy(&q->q_mutex);
248 	hci1394_buf_free(&q->q_desc.qb_buf_handle);
249 	hci1394_buf_free(&q->q_data.qb_buf_handle);
250 	kmem_free(q, sizeof (hci1394_q_t));
251 	*q_handle = NULL;
252 
253 	TNF_PROBE_0_DEBUG(hci1394_q_fini_exit, HCI1394_TNF_HAL_STACK, "");
254 }
255 
256 
257 /*
258  * hci1394_q_buf_setup()
259  *    Initialization of buffer pointers which are present in both the descriptor
260  *    buffer and data buffer (No reason to duplicate the code)
261  */
262 static void
263 hci1394_q_buf_setup(hci1394_q_buf_t *qbuf)
264 {
265 	ASSERT(qbuf != NULL);
266 	TNF_PROBE_0_DEBUG(hci1394_q_buf_setup_enter, HCI1394_TNF_HAL_STACK, "");
267 
268 	/* start with the first cookie */
269 	qbuf->qb_ptrs.qp_current_buf = 0;
270 	qbuf->qb_ptrs.qp_begin = qbuf->qb_ptrs.qp_top;
271 	qbuf->qb_ptrs.qp_end = qbuf->qb_ptrs.qp_begin +
272 	    qbuf->qb_cookie[qbuf->qb_ptrs.qp_current_buf].dmac_size - 1;
273 	qbuf->qb_ptrs.qp_current = qbuf->qb_ptrs.qp_begin;
274 	qbuf->qb_ptrs.qp_offset = 0;
275 
276 	/*
277 	 * The free_buf and free pointer will change everytime an ACK (of some
278 	 * type) is processed.  Free is the last byte in the last cookie.
279 	 */
280 	qbuf->qb_ptrs.qp_free_buf = qbuf->qb_buf.bi_cookie_count - 1;
281 	qbuf->qb_ptrs.qp_free = qbuf->qb_ptrs.qp_bottom;
282 
283 	/*
284 	 * Start with no space to write descriptors.  We first need to call
285 	 * hci1394_q_reserve() before calling hci1394_q_at_write_O*().
286 	 */
287 	qbuf->qb_ptrs.qp_resv_size = 0;
288 
289 	TNF_PROBE_0_DEBUG(hci1394_q_buf_setup_exit, HCI1394_TNF_HAL_STACK, "");
290 }
291 
292 
293 /*
294  * hci1394_q_reset()
295  *    Resets the buffers to an initial state.  This should be called during
296  *    attach and resume.
297  */
298 static void
299 hci1394_q_reset(hci1394_q_handle_t q_handle)
300 {
301 	hci1394_q_buf_t *desc;
302 	hci1394_q_buf_t *data;
303 	int index;
304 
305 	ASSERT(q_handle != NULL);
306 	TNF_PROBE_0_DEBUG(hci1394_q_reset_enter, HCI1394_TNF_HAL_STACK, "");
307 
308 	mutex_enter(&q_handle->q_mutex);
309 	desc = &q_handle->q_desc;
310 	data = &q_handle->q_data;
311 
312 	hci1394_q_buf_setup(desc);
313 	hci1394_q_buf_setup(data);
314 
315 	/* DMA starts off stopped, no previous descriptor to link from */
316 	q_handle->q_dma_running = B_FALSE;
317 	q_handle->q_block_cnt = 0;
318 	q_handle->q_previous = NULL;
319 
320 	/* If this is an AR Q, setup IM's for the data buffers that we have */
321 	if (q_handle->q_info.qi_mode == HCI1394_ARQ) {
322 		/*
323 		 * This points to where to find the first IM descriptor.  Since
324 		 * we just reset the pointers in hci1394_q_buf_setup(), the
325 		 * first IM we write below will be found at the top of the Q.
326 		 */
327 		q_handle->q_head = desc->qb_ptrs.qp_top;
328 
329 		for (index = 0; index < data->qb_buf.bi_cookie_count; index++) {
330 			hci1394_q_ar_write_IM(q_handle, desc,
331 			    data->qb_cookie[index].dmac_address,
332 			    data->qb_cookie[index].dmac_size);
333 		}
334 
335 		/*
336 		 * The space left in the current IM is the size of the buffer.
337 		 * The current buffer is the first buffer added to the AR Q.
338 		 */
339 		q_handle->q_space_left = data->qb_cookie[0].dmac_size;
340 	}
341 
342 	mutex_exit(&q_handle->q_mutex);
343 	TNF_PROBE_0_DEBUG(hci1394_q_reset_exit, HCI1394_TNF_HAL_STACK, "");
344 }
345 
346 
347 /*
348  * hci1394_q_resume()
349  *    This is called during a resume (after a successful suspend). Currently
350  *    we only call reset.  Since this is not a time critical function, we will
351  *    leave this as a separate function to increase readability.
352  */
353 void
354 hci1394_q_resume(hci1394_q_handle_t q_handle)
355 {
356 	ASSERT(q_handle != NULL);
357 	TNF_PROBE_0_DEBUG(hci1394_q_resume_enter, HCI1394_TNF_HAL_STACK, "");
358 	hci1394_q_reset(q_handle);
359 	TNF_PROBE_0_DEBUG(hci1394_q_resume_exit, HCI1394_TNF_HAL_STACK, "");
360 }
361 
362 
363 /*
364  * hci1394_q_stop()
365  *    This call informs us that a DMA engine has been stopped.  It does not
366  *    perform the actual stop. We need to know this so that when we add a
367  *    new descriptor, we do a start instead of a wake.
368  */
369 void
370 hci1394_q_stop(hci1394_q_handle_t q_handle)
371 {
372 	ASSERT(q_handle != NULL);
373 	TNF_PROBE_0_DEBUG(hci1394_q_stop_enter, HCI1394_TNF_HAL_STACK, "");
374 	mutex_enter(&q_handle->q_mutex);
375 	q_handle->q_dma_running = B_FALSE;
376 	mutex_exit(&q_handle->q_mutex);
377 	TNF_PROBE_0_DEBUG(hci1394_q_stop_exit, HCI1394_TNF_HAL_STACK, "");
378 }
379 
380 
381 /*
382  * hci1394_q_reserve()
383  *    Reserve space in the AT descriptor or data buffer. This ensures that we
384  *    can get a contiguous buffer. Descriptors have to be in a contiguous
385  *    buffer. Data does not have to be in a contiguous buffer but we do this to
386  *    reduce complexity. For systems with small page sizes (e.g. x86), this
387  *    could result in inefficient use of the data buffers when sending large
388  *    data blocks (this only applies to non-physical block write ATREQs and
389  *    block read ATRESP). Since it looks like most protocols that use large data
390  *    blocks (like SPB-2), use physical transfers to do this (due to their
391  *    efficiency), this will probably not be a real world problem.  If it turns
392  *    out to be a problem, the options are to force a single cookie for the data
393  *    buffer, allow multiple cookies and have a larger data space, or change the
394  *    data code to use a OMI, OM, OL descriptor sequence (instead of OMI, OL).
395  */
396 static int
397 hci1394_q_reserve(hci1394_q_buf_t *qbuf, uint_t size, uint32_t *io_addr)
398 {
399 	uint_t aligned_size;
400 
401 
402 	ASSERT(qbuf != NULL);
403 	TNF_PROBE_0_DEBUG(hci1394_q_reserve_enter, HCI1394_TNF_HAL_STACK, "");
404 
405 	/* Save backup of pointers in case we have to unreserve */
406 	qbuf->qb_backup_ptrs = qbuf->qb_ptrs;
407 
408 	/*
409 	 * Make sure all alloc's are quadlet aligned. The data doesn't have to
410 	 * be, so we will force it to be.
411 	 */
412 	aligned_size = HCI1394_ALIGN_QUAD(size);
413 
414 	/*
415 	 * if the free pointer is in the current buffer and the free pointer
416 	 * is below the current pointer (i.e. has not wrapped around)
417 	 */
418 	if ((qbuf->qb_ptrs.qp_current_buf == qbuf->qb_ptrs.qp_free_buf) &&
419 	    (qbuf->qb_ptrs.qp_free >= qbuf->qb_ptrs.qp_current)) {
420 		/*
421 		 * The free pointer is in this buffer below the current pointer.
422 		 * Check to see if we have enough free space left.
423 		 */
424 		if ((qbuf->qb_ptrs.qp_current + aligned_size) <=
425 		    qbuf->qb_ptrs.qp_free) {
426 			/* Setup up our reserved size, return the IO address */
427 			qbuf->qb_ptrs.qp_resv_size = aligned_size;
428 			*io_addr = (uint32_t)(qbuf->qb_cookie[
429 			    qbuf->qb_ptrs.qp_current_buf].dmac_address +
430 			    qbuf->qb_ptrs.qp_offset);
431 
432 		/*
433 		 * The free pointer is in this buffer below the current pointer.
434 		 * We do not have enough free space for the alloc. Return
435 		 * failure.
436 		 */
437 		} else {
438 			qbuf->qb_ptrs.qp_resv_size = 0;
439 			TNF_PROBE_0(hci1394_q_reserve_ns_fail,
440 			    HCI1394_TNF_HAL_ERROR, "");
441 			TNF_PROBE_0_DEBUG(hci1394_q_reserve_exit,
442 			    HCI1394_TNF_HAL_STACK, "");
443 			return (DDI_FAILURE);
444 		}
445 
446 	/*
447 	 * If there is not enough room to fit in the current buffer (not
448 	 * including wrap around), we will go to the next buffer and check
449 	 * there. If we only have one buffer (i.e. one cookie), we will end up
450 	 * staying at the current buffer and wrapping the address back to the
451 	 * top.
452 	 */
453 	} else if ((qbuf->qb_ptrs.qp_current + aligned_size) >
454 	    qbuf->qb_ptrs.qp_end) {
455 		/* Go to the next buffer (or the top of ours for one cookie) */
456 		hci1394_q_next_buf(qbuf);
457 
458 		/* If the free pointer is in the new current buffer */
459 		if (qbuf->qb_ptrs.qp_current_buf == qbuf->qb_ptrs.qp_free_buf) {
460 			/*
461 			 * The free pointer is in this buffer. If we do not have
462 			 * enough free space for the alloc. Return failure.
463 			 */
464 			if ((qbuf->qb_ptrs.qp_current + aligned_size) >
465 			    qbuf->qb_ptrs.qp_free) {
466 				qbuf->qb_ptrs.qp_resv_size = 0;
467 				TNF_PROBE_0(hci1394_q_reserve_ns_fail,
468 				    HCI1394_TNF_HAL_ERROR, "");
469 				TNF_PROBE_0_DEBUG(hci1394_q_reserve_exit,
470 				    HCI1394_TNF_HAL_STACK, "");
471 				return (DDI_FAILURE);
472 			/*
473 			 * The free pointer is in this buffer. We have enough
474 			 * free space left.
475 			 */
476 			} else {
477 				/*
478 				 * Setup up our reserved size, return the IO
479 				 * address
480 				 */
481 				qbuf->qb_ptrs.qp_resv_size = aligned_size;
482 				*io_addr = (uint32_t)(qbuf->qb_cookie[
483 				    qbuf->qb_ptrs.qp_current_buf].dmac_address +
484 				    qbuf->qb_ptrs.qp_offset);
485 			}
486 
487 		/*
488 		 * We switched buffers and the free pointer is still in another
489 		 * buffer. We have sufficient space in this buffer for the alloc
490 		 * after changing buffers.
491 		 */
492 		} else {
493 			/* Setup up our reserved size, return the IO address */
494 			qbuf->qb_ptrs.qp_resv_size = aligned_size;
495 			*io_addr = (uint32_t)(qbuf->qb_cookie[
496 			    qbuf->qb_ptrs.qp_current_buf].dmac_address +
497 			    qbuf->qb_ptrs.qp_offset);
498 		}
499 	/*
500 	 * The free pointer is in another buffer. We have sufficient space in
501 	 * this buffer for the alloc.
502 	 */
503 	} else {
504 		/* Setup up our reserved size, return the IO address */
505 		qbuf->qb_ptrs.qp_resv_size = aligned_size;
506 		*io_addr = (uint32_t)(qbuf->qb_cookie[
507 		    qbuf->qb_ptrs.qp_current_buf].dmac_address +
508 		    qbuf->qb_ptrs.qp_offset);
509 	}
510 
511 	TNF_PROBE_0_DEBUG(hci1394_q_reserve_exit, HCI1394_TNF_HAL_STACK, "");
512 
513 	return (DDI_SUCCESS);
514 }
515 
516 /*
517  * hci1394_q_unreserve()
518  *    Set the buffer pointer to what they were before hci1394_reserve().  This
519  *    will be called when we encounter errors during hci1394_q_at*().
520  */
521 static void
522 hci1394_q_unreserve(hci1394_q_buf_t *qbuf)
523 {
524 	ASSERT(qbuf != NULL);
525 	TNF_PROBE_0_DEBUG(hci1394_q_unreserve_enter, HCI1394_TNF_HAL_STACK, "");
526 
527 	/* Go back to pointer setting before the reserve */
528 	qbuf->qb_ptrs = qbuf->qb_backup_ptrs;
529 
530 	TNF_PROBE_0_DEBUG(hci1394_q_unreserve_exit, HCI1394_TNF_HAL_STACK, "");
531 }
532 
533 
534 /*
535  * hci1394_q_next_buf()
536  *    Set our current buffer to the next cookie.  If we only have one cookie, we
537  *    will go back to the top of our buffer.
538  */
539 void
540 hci1394_q_next_buf(hci1394_q_buf_t *qbuf)
541 {
542 	ASSERT(qbuf != NULL);
543 	TNF_PROBE_0_DEBUG(hci1394_q_next_buf_enter, HCI1394_TNF_HAL_STACK, "");
544 
545 	/*
546 	 * go to the next cookie, if we are >= the cookie count, go back to the
547 	 * first cookie.
548 	 */
549 	qbuf->qb_ptrs.qp_current_buf++;
550 	if (qbuf->qb_ptrs.qp_current_buf >= qbuf->qb_buf.bi_cookie_count) {
551 		qbuf->qb_ptrs.qp_current_buf = 0;
552 	}
553 
554 	/* adjust the begin, end, current, and offset pointers */
555 	qbuf->qb_ptrs.qp_begin = qbuf->qb_ptrs.qp_end + 1;
556 	if (qbuf->qb_ptrs.qp_begin > qbuf->qb_ptrs.qp_bottom) {
557 		qbuf->qb_ptrs.qp_begin = qbuf->qb_ptrs.qp_top;
558 	}
559 	qbuf->qb_ptrs.qp_end = qbuf->qb_ptrs.qp_begin +
560 	    qbuf->qb_cookie[qbuf->qb_ptrs.qp_current_buf].dmac_size - 1;
561 	qbuf->qb_ptrs.qp_current = qbuf->qb_ptrs.qp_begin;
562 	qbuf->qb_ptrs.qp_offset = 0;
563 
564 	TNF_PROBE_0_DEBUG(hci1394_q_next_buf_exit, HCI1394_TNF_HAL_STACK, "");
565 }
566 
567 
568 /*
569  * hci1394_q_at()
570  *    Place an AT command that does NOT need the data buffer into the DMA chain.
571  *    Some examples of this are quadlet read/write, PHY packets, ATREQ Block
572  *    Read, and ATRESP block write. result is only valid on failure.
573  */
574 int
575 hci1394_q_at(hci1394_q_handle_t q_handle, hci1394_q_cmd_t *cmd,
576     hci1394_basic_pkt_t *hdr, uint_t hdrsize, int *result)
577 {
578 	int status;
579 	uint32_t ioaddr;
580 
581 
582 	ASSERT(q_handle != NULL);
583 	ASSERT(cmd != NULL);
584 	ASSERT(hdr != NULL);
585 	TNF_PROBE_0_DEBUG(hci1394_q_at_enter, HCI1394_TNF_HAL_STACK, "");
586 
587 	mutex_enter(&q_handle->q_mutex);
588 
589 	/*
590 	 * Check the HAL state and generation when the AT Q is locked.  This
591 	 * will make sure that we get all the commands when we flush the Q's
592 	 * during a reset or shutdown.
593 	 */
594 	if ((hci1394_state(q_handle->q_drvinfo) != HCI1394_NORMAL) ||
595 	    (hci1394_ohci_current_busgen(q_handle->q_ohci) !=
596 	    cmd->qc_generation)) {
597 		*result = H1394_STATUS_INVALID_BUSGEN;
598 		mutex_exit(&q_handle->q_mutex);
599 		TNF_PROBE_0(hci1394_q_at_st_fail, HCI1394_TNF_HAL_ERROR, "");
600 		TNF_PROBE_0_DEBUG(hci1394_q_at_exit, HCI1394_TNF_HAL_STACK,
601 		    "");
602 		return (DDI_FAILURE);
603 	}
604 
605 	/* save away the argument to pass up when this command completes */
606 	cmd->qc_node.tln_addr = cmd;
607 
608 	/* we have not written any 16 byte blocks to the descriptor yet */
609 	q_handle->q_block_cnt = 0;
610 
611 	/* Reserve space for an OLI in the descriptor buffer */
612 	status = hci1394_q_reserve(&q_handle->q_desc,
613 	    sizeof (hci1394_desc_imm_t), &ioaddr);
614 	if (status != DDI_SUCCESS) {
615 		*result = H1394_STATUS_NOMORE_SPACE;
616 		mutex_exit(&q_handle->q_mutex);
617 		TNF_PROBE_0(hci1394_q_at_qre_fail, HCI1394_TNF_HAL_ERROR, "");
618 		TNF_PROBE_0_DEBUG(hci1394_q_at_exit, HCI1394_TNF_HAL_STACK,
619 		    "");
620 		return (DDI_FAILURE);
621 	}
622 
623 	/* write the OLI to the descriptor buffer */
624 	hci1394_q_at_write_OLI(q_handle, &q_handle->q_desc, cmd, hdr, hdrsize);
625 
626 	/* Add the AT command to the queued list */
627 	hci1394_tlist_add(q_handle->q_queued_list, &cmd->qc_node);
628 
629 	mutex_exit(&q_handle->q_mutex);
630 	TNF_PROBE_0_DEBUG(hci1394_q_at_exit, HCI1394_TNF_HAL_STACK, "");
631 
632 	return (DDI_SUCCESS);
633 }
634 
635 
636 /*
637  * XXX - NOTE: POSSIBLE FUTURE OPTIMIZATION
638  *    ATREQ Block read and write's that go through software are not very
639  *    efficient (one of the reasons to use physical space). A copy is forced
640  *    on all block reads due to the design of OpenHCI. Writes do not have this
641  *    same restriction.  This design forces a copy for writes too (we always
642  *    copy into a data buffer before sending). There are many reasons for this
643  *    including complexity reduction.  There is a data size threshold where a
644  *    copy is more expensive than mapping the data buffer address (or worse
645  *    case a big enough difference where it pays to do it). However, we move
646  *    block data around in mblks which means that our data may be scattered
647  *    over many buffers.  This adds to the complexity of mapping and setting
648  *    up the OpenHCI descriptors.
649  *
650  *    If someone really needs a speedup on block write ATREQs, my recommendation
651  *    would be to add an additional command type at the target interface for a
652  *    fast block write.  The target driver would pass a mapped io addr to use.
653  *    A function like "hci1394_q_at_with_ioaddr()" could be created which would
654  *    be almost an exact copy of hci1394_q_at_with_data() without the
655  *    hci1394_q_reserve() and hci1394_q_at_rep_put8() for the data buffer.
656  */
657 
658 
659 /*
660  * hci1394_q_at_with_data()
661  *    Place an AT command that does need the data buffer into the DMA chain.
662  *    The data is passed as a pointer to a kernel virtual address. An example of
663  *    this is the lock operations. result is only valid on failure.
664  */
665 int
666 hci1394_q_at_with_data(hci1394_q_handle_t q_handle, hci1394_q_cmd_t *cmd,
667     hci1394_basic_pkt_t *hdr, uint_t hdrsize, uint8_t *data, uint_t datasize,
668     int *result)
669 {
670 	uint32_t desc_ioaddr;
671 	uint32_t data_ioaddr;
672 	int status;
673 
674 
675 	ASSERT(q_handle != NULL);
676 	ASSERT(cmd != NULL);
677 	ASSERT(hdr != NULL);
678 	ASSERT(data != NULL);
679 	TNF_PROBE_0_DEBUG(hci1394_q_at_with_data_enter, HCI1394_TNF_HAL_STACK,
680 	    "");
681 
682 	mutex_enter(&q_handle->q_mutex);
683 
684 	/*
685 	 * Check the HAL state and generation when the AT Q is locked.  This
686 	 * will make sure that we get all the commands when we flush the Q's
687 	 * during a reset or shutdown.
688 	 */
689 	if ((hci1394_state(q_handle->q_drvinfo) != HCI1394_NORMAL) ||
690 	    (hci1394_ohci_current_busgen(q_handle->q_ohci) !=
691 	    cmd->qc_generation)) {
692 		*result = H1394_STATUS_INVALID_BUSGEN;
693 		mutex_exit(&q_handle->q_mutex);
694 		TNF_PROBE_0_DEBUG(hci1394_q_at_wd_st_fail,
695 		    HCI1394_TNF_HAL_STACK, "");
696 		return (DDI_FAILURE);
697 	}
698 
699 	/* save away the argument to pass up when this command completes */
700 	cmd->qc_node.tln_addr = cmd;
701 
702 	/* we have not written any 16 byte blocks to the descriptor yet */
703 	q_handle->q_block_cnt = 0;
704 
705 	/* Reserve space for an OMI and OL in the descriptor buffer */
706 	status = hci1394_q_reserve(&q_handle->q_desc,
707 	    (sizeof (hci1394_desc_imm_t) + sizeof (hci1394_desc_t)),
708 	    &desc_ioaddr);
709 	if (status != DDI_SUCCESS) {
710 		*result = H1394_STATUS_NOMORE_SPACE;
711 		mutex_exit(&q_handle->q_mutex);
712 		TNF_PROBE_0(hci1394_q_at_wd_qre_fail,
713 		    HCI1394_TNF_HAL_ERROR, "");
714 		TNF_PROBE_0_DEBUG(hci1394_q_at_with_data_exit,
715 		    HCI1394_TNF_HAL_STACK, "");
716 		return (DDI_FAILURE);
717 	}
718 
719 	/* allocate space for data in the data buffer */
720 	status = hci1394_q_reserve(&q_handle->q_data, datasize, &data_ioaddr);
721 	if (status != DDI_SUCCESS) {
722 		*result = H1394_STATUS_NOMORE_SPACE;
723 		hci1394_q_unreserve(&q_handle->q_desc);
724 		mutex_exit(&q_handle->q_mutex);
725 		TNF_PROBE_0(hci1394_q_at_wd_qra_fail,
726 		    HCI1394_TNF_HAL_ERROR, "");
727 		TNF_PROBE_0_DEBUG(hci1394_q_at_with_data_exit,
728 		    HCI1394_TNF_HAL_STACK, "");
729 		return (DDI_FAILURE);
730 	}
731 
732 	/* Copy data into data buffer */
733 	hci1394_q_at_rep_put8(&q_handle->q_data, cmd, data, datasize);
734 
735 	/* write the OMI to the descriptor buffer */
736 	hci1394_q_at_write_OMI(q_handle, &q_handle->q_desc, cmd, hdr, hdrsize);
737 
738 	/* write the OL to the descriptor buffer */
739 	hci1394_q_at_write_OL(q_handle, &q_handle->q_desc, cmd, data_ioaddr,
740 	    datasize);
741 
742 	/* Add the AT command to the queued list */
743 	hci1394_tlist_add(q_handle->q_queued_list, &cmd->qc_node);
744 
745 	mutex_exit(&q_handle->q_mutex);
746 	TNF_PROBE_0_DEBUG(hci1394_q_at_with_data_exit, HCI1394_TNF_HAL_STACK,
747 	    "");
748 
749 	return (DDI_SUCCESS);
750 }
751 
752 
753 /*
754  * hci1394_q_at_with_mblk()
755  *    Place an AT command that does need the data buffer into the DMA chain.
756  *    The data is passed in mblk_t(s). Examples of this are a block write
757  *    ATREQ and a block read ATRESP. The services layer and the hal use a
758  *    private structure (h1394_mblk_t) to keep track of how much of the mblk
759  *    to send since we may have to break the transfer up into smaller blocks.
760  *    (i.e. a 1MByte block write would go out in 2KByte chunks. result is only
761  *    valid on failure.
762  */
763 int
764 hci1394_q_at_with_mblk(hci1394_q_handle_t q_handle, hci1394_q_cmd_t *cmd,
765     hci1394_basic_pkt_t *hdr, uint_t hdrsize, h1394_mblk_t *mblk, int *result)
766 {
767 	uint32_t desc_ioaddr;
768 	uint32_t data_ioaddr;
769 	int status;
770 
771 
772 	ASSERT(q_handle != NULL);
773 	ASSERT(cmd != NULL);
774 	ASSERT(hdr != NULL);
775 	ASSERT(mblk != NULL);
776 	TNF_PROBE_0_DEBUG(hci1394_q_at_with_mblk_enter, HCI1394_TNF_HAL_STACK,
777 	    "");
778 
779 	mutex_enter(&q_handle->q_mutex);
780 
781 	/*
782 	 * Check the HAL state and generation when the AT Q is locked.  This
783 	 * will make sure that we get all the commands when we flush the Q's
784 	 * during a reset or shutdown.
785 	 */
786 	if ((hci1394_state(q_handle->q_drvinfo) != HCI1394_NORMAL) ||
787 	    (hci1394_ohci_current_busgen(q_handle->q_ohci) !=
788 	    cmd->qc_generation)) {
789 		*result = H1394_STATUS_INVALID_BUSGEN;
790 		mutex_exit(&q_handle->q_mutex);
791 		TNF_PROBE_0_DEBUG(hci1394_q_at_wm_st_fail,
792 		    HCI1394_TNF_HAL_STACK, "");
793 		return (DDI_FAILURE);
794 	}
795 
796 	/* save away the argument to pass up when this command completes */
797 	cmd->qc_node.tln_addr = cmd;
798 
799 	/* we have not written any 16 byte blocks to the descriptor yet */
800 	q_handle->q_block_cnt = 0;
801 
802 	/* Reserve space for an OMI and OL in the descriptor buffer */
803 	status = hci1394_q_reserve(&q_handle->q_desc,
804 	    (sizeof (hci1394_desc_imm_t) + sizeof (hci1394_desc_t)),
805 	    &desc_ioaddr);
806 	if (status != DDI_SUCCESS) {
807 		*result = H1394_STATUS_NOMORE_SPACE;
808 		mutex_exit(&q_handle->q_mutex);
809 		TNF_PROBE_0(hci1394_q_at_wm_qre_fail,
810 		    HCI1394_TNF_HAL_ERROR, "");
811 		TNF_PROBE_0_DEBUG(hci1394_q_at_with_mblk_exit,
812 		    HCI1394_TNF_HAL_STACK, "");
813 		return (DDI_FAILURE);
814 	}
815 
816 	/* Reserve space for data in the data buffer */
817 	status = hci1394_q_reserve(&q_handle->q_data, mblk->length,
818 	    &data_ioaddr);
819 	if (status != DDI_SUCCESS) {
820 		*result = H1394_STATUS_NOMORE_SPACE;
821 		hci1394_q_unreserve(&q_handle->q_desc);
822 		mutex_exit(&q_handle->q_mutex);
823 		TNF_PROBE_0(hci1394_q_at_wm_qra_fail,
824 		    HCI1394_TNF_HAL_ERROR, "");
825 		TNF_PROBE_0_DEBUG(hci1394_q_at_with_mblk_exit,
826 		    HCI1394_TNF_HAL_STACK, "");
827 		return (DDI_FAILURE);
828 	}
829 
830 	/* Copy mblk data into data buffer */
831 	hci1394_q_at_copy_from_mblk(&q_handle->q_data, cmd, mblk);
832 
833 	/* write the OMI to the descriptor buffer */
834 	hci1394_q_at_write_OMI(q_handle, &q_handle->q_desc, cmd, hdr, hdrsize);
835 
836 	/* write the OL to the descriptor buffer */
837 	hci1394_q_at_write_OL(q_handle, &q_handle->q_desc, cmd, data_ioaddr,
838 	    mblk->length);
839 
840 	/* Add the AT command to the queued list */
841 	hci1394_tlist_add(q_handle->q_queued_list, &cmd->qc_node);
842 
843 	mutex_exit(&q_handle->q_mutex);
844 	TNF_PROBE_0_DEBUG(hci1394_q_at_with_mblk_exit, HCI1394_TNF_HAL_STACK,
845 	    "");
846 
847 	return (DDI_SUCCESS);
848 }
849 
850 
851 /*
852  * hci1394_q_at_next()
853  *    Return the next completed AT command in cmd.  If flush_q is true, we will
854  *    return the command regardless if it finished or not.  We will flush
855  *    during bus reset processing, shutdown, and detach.
856  */
857 void
858 hci1394_q_at_next(hci1394_q_handle_t q_handle, boolean_t flush_q,
859     hci1394_q_cmd_t **cmd)
860 {
861 	hci1394_q_buf_t *desc;
862 	hci1394_q_buf_t *data;
863 	hci1394_tlist_node_t *node;
864 	uint32_t cmd_status;
865 
866 
867 	ASSERT(q_handle != NULL);
868 	ASSERT(cmd != NULL);
869 	TNF_PROBE_0_DEBUG(hci1394_q_at_next_enter, HCI1394_TNF_HAL_STACK, "");
870 
871 	mutex_enter(&q_handle->q_mutex);
872 
873 	desc = &q_handle->q_desc;
874 	data = &q_handle->q_data;
875 
876 	/* Sync descriptor buffer */
877 	(void) ddi_dma_sync(desc->qb_buf.bi_dma_handle, 0,
878 	    desc->qb_buf.bi_length, DDI_DMA_SYNC_FORKERNEL);
879 
880 	/* Look at the top cmd on the queued list (without removing it) */
881 	hci1394_tlist_peek(q_handle->q_queued_list, &node);
882 	if (node == NULL) {
883 		/* There are no more commands left on the queued list */
884 		*cmd = NULL;
885 		mutex_exit(&q_handle->q_mutex);
886 		TNF_PROBE_0_DEBUG(hci1394_q_at_next_exit, HCI1394_TNF_HAL_STACK,
887 		    "");
888 		return;
889 	}
890 
891 	/*
892 	 * There is a command on the list, read its status and timestamp when
893 	 * it was sent
894 	 */
895 	*cmd = (hci1394_q_cmd_t *)node->tln_addr;
896 	cmd_status = ddi_get32(desc->qb_buf.bi_handle, (*cmd)->qc_status_addr);
897 	(*cmd)->qc_timestamp = cmd_status & DESC_ST_TIMESTAMP_MASK;
898 	cmd_status = HCI1394_DESC_EVT_GET(cmd_status);
899 
900 	/*
901 	 * If we are flushing the q (e.g. due to a bus reset), we will return
902 	 * the command regardless of its completion status. If we are not
903 	 * flushing the Q and we do not have status on the command (e.g. status
904 	 * = 0), we are done with this Q for now.
905 	 */
906 	if (flush_q == B_FALSE) {
907 		if (cmd_status == 0) {
908 			*cmd = NULL;
909 			mutex_exit(&q_handle->q_mutex);
910 			TNF_PROBE_0_DEBUG(hci1394_q_at_next_exit,
911 			    HCI1394_TNF_HAL_STACK, "");
912 			return;
913 		}
914 	}
915 
916 	/*
917 	 * The command completed, remove it from the queued list. There is not
918 	 * a race condition to delete the node in the list here.  This is the
919 	 * only place the node will be deleted so we do not need to check the
920 	 * return status.
921 	 */
922 	(void) hci1394_tlist_delete(q_handle->q_queued_list, node);
923 
924 	/*
925 	 * Free the space used by the command in the descriptor and data
926 	 * buffers.
927 	 */
928 	desc->qb_ptrs.qp_free_buf = (*cmd)->qc_descriptor_buf;
929 	desc->qb_ptrs.qp_free = (*cmd)->qc_descriptor_end;
930 	if ((*cmd)->qc_data_used == B_TRUE) {
931 		data->qb_ptrs.qp_free_buf = (*cmd)->qc_data_buf;
932 		data->qb_ptrs.qp_free = (*cmd)->qc_data_end;
933 	}
934 
935 	/* return command status */
936 	(*cmd)->qc_status = cmd_status;
937 
938 	mutex_exit(&q_handle->q_mutex);
939 	TNF_PROBE_0_DEBUG(hci1394_q_at_next_exit, HCI1394_TNF_HAL_STACK, "");
940 }
941 
942 
943 /*
944  * hci1394_q_at_write_OMI()
945  *    Write an OMI descriptor into the AT descriptor buffer passed in as qbuf.
946  *    Buffer state information is stored in cmd.  Use the hdr and hdr size for
947  *    the additional information attached to an immediate descriptor.
948  */
949 void
950 hci1394_q_at_write_OMI(hci1394_q_handle_t q_handle, hci1394_q_buf_t *qbuf,
951     hci1394_q_cmd_t *cmd, hci1394_basic_pkt_t *hdr, uint_t hdrsize)
952 {
953 	hci1394_desc_imm_t *desc;
954 	uint32_t data;
955 
956 
957 	ASSERT(qbuf != NULL);
958 	ASSERT(cmd != NULL);
959 	ASSERT(hdr != NULL);
960 	ASSERT(MUTEX_HELD(&q_handle->q_mutex));
961 	TNF_PROBE_0_DEBUG(hci1394_q_at_write_OMI_enter, HCI1394_TNF_HAL_STACK,
962 	    "");
963 
964 	/* The only valid "header" sizes for an OMI are 8 bytes or 16 bytes */
965 	ASSERT((hdrsize == 8) || (hdrsize == 16));
966 
967 	/* Make sure enough room for OMI */
968 	ASSERT(qbuf->qb_ptrs.qp_resv_size >= sizeof (hci1394_desc_imm_t));
969 
970 	/* Store the offset of the top of this descriptor block */
971 	qbuf->qb_ptrs.qp_offset = (uint32_t)(qbuf->qb_ptrs.qp_current -
972 	    qbuf->qb_ptrs.qp_begin);
973 
974 	/* Setup OpenHCI OMI Header */
975 	desc = (hci1394_desc_imm_t *)qbuf->qb_ptrs.qp_current;
976 	data = DESC_AT_OMI | (hdrsize & DESC_HDR_REQCOUNT_MASK);
977 	ddi_put32(qbuf->qb_buf.bi_handle, &desc->hdr, data);
978 	ddi_put32(qbuf->qb_buf.bi_handle, &desc->data_addr, 0);
979 	ddi_put32(qbuf->qb_buf.bi_handle, &desc->branch, 0);
980 	ddi_put32(qbuf->qb_buf.bi_handle, &desc->status, cmd->qc_timestamp);
981 
982 	/*
983 	 * Copy in 1394 header. Size is in bytes, convert it to a 32-bit word
984 	 * count.
985 	 */
986 	ddi_rep_put32(qbuf->qb_buf.bi_handle, &hdr->q1, &desc->q1,
987 	    hdrsize >> 2, DDI_DEV_AUTOINCR);
988 
989 	/*
990 	 * We wrote 2 16 byte blocks in the descriptor buffer, update the count
991 	 * accordingly.  Update the reserved size and current pointer.
992 	 */
993 	q_handle->q_block_cnt += 2;
994 	qbuf->qb_ptrs.qp_resv_size -= sizeof (hci1394_desc_imm_t);
995 	qbuf->qb_ptrs.qp_current += sizeof (hci1394_desc_imm_t);
996 
997 	TNF_PROBE_0_DEBUG(hci1394_q_at_write_OMI_exit, HCI1394_TNF_HAL_STACK,
998 	    "");
999 }
1000 
1001 
1002 /*
1003  * hci1394_q_at_write_OLI()
1004  *    Write an OLI descriptor into the AT descriptor buffer passed in as qbuf.
1005  *    Buffer state information is stored in cmd.  Use the hdr and hdr size for
1006  *    the additional information attached to an immediate descriptor.
1007  */
1008 void
1009 hci1394_q_at_write_OLI(hci1394_q_handle_t q_handle, hci1394_q_buf_t *qbuf,
1010     hci1394_q_cmd_t *cmd, hci1394_basic_pkt_t *hdr, uint_t hdrsize)
1011 {
1012 	hci1394_desc_imm_t *desc;
1013 	uint32_t data;
1014 	uint32_t command_ptr;
1015 	uint32_t tcode;
1016 
1017 
1018 	ASSERT(qbuf != NULL);
1019 	ASSERT(cmd != NULL);
1020 	ASSERT(hdr != NULL);
1021 	ASSERT(MUTEX_HELD(&q_handle->q_mutex));
1022 	TNF_PROBE_0_DEBUG(hci1394_q_at_write_OLI_enter, HCI1394_TNF_HAL_STACK,
1023 	    "");
1024 
1025 	/* The only valid "header" sizes for an OLI are 8, 12, 16 bytes */
1026 	ASSERT((hdrsize == 8) || (hdrsize == 12) || (hdrsize == 16));
1027 
1028 	/* make sure enough room for 1 OLI */
1029 	ASSERT(qbuf->qb_ptrs.qp_resv_size >= sizeof (hci1394_desc_imm_t));
1030 
1031 	/* Store the offset of the top of this descriptor block */
1032 	qbuf->qb_ptrs.qp_offset = (uint32_t)(qbuf->qb_ptrs.qp_current -
1033 	    qbuf->qb_ptrs.qp_begin);
1034 
1035 	/* Setup OpenHCI OLI Header */
1036 	desc = (hci1394_desc_imm_t *)qbuf->qb_ptrs.qp_current;
1037 	data = DESC_AT_OLI | (hdrsize & DESC_HDR_REQCOUNT_MASK);
1038 	ddi_put32(qbuf->qb_buf.bi_handle, &desc->hdr, data);
1039 	ddi_put32(qbuf->qb_buf.bi_handle, &desc->data_addr, 0);
1040 	ddi_put32(qbuf->qb_buf.bi_handle, &desc->branch, 0);
1041 	ddi_put32(qbuf->qb_buf.bi_handle, &desc->status, cmd->qc_timestamp);
1042 
1043 	/* Setup 1394 Header */
1044 	tcode = (hdr->q1 & DESC_PKT_TCODE_MASK) >> DESC_PKT_TCODE_SHIFT;
1045 	if ((tcode == IEEE1394_TCODE_WRITE_QUADLET) ||
1046 	    (tcode == IEEE1394_TCODE_READ_QUADLET_RESP)) {
1047 		/*
1048 		 * if the tcode = a quadlet write, move the last quadlet as
1049 		 * 8-bit data.  All data is treated as 8-bit data (even quadlet
1050 		 * reads and writes). Therefore, target drivers MUST take that
1051 		 * into consideration when accessing device registers.
1052 		 */
1053 		ddi_rep_put32(qbuf->qb_buf.bi_handle, &hdr->q1, &desc->q1, 3,
1054 		    DDI_DEV_AUTOINCR);
1055 		ddi_rep_put8(qbuf->qb_buf.bi_handle, (uint8_t *)&hdr->q4,
1056 		    (uint8_t *)&desc->q4, 4, DDI_DEV_AUTOINCR);
1057 	} else {
1058 		ddi_rep_put32(qbuf->qb_buf.bi_handle, &hdr->q1, &desc->q1,
1059 		    hdrsize >> 2, DDI_DEV_AUTOINCR);
1060 	}
1061 
1062 	/*
1063 	 * We wrote 2 16 byte blocks in the descriptor buffer, update the count
1064 	 * accordingly.
1065 	 */
1066 	q_handle->q_block_cnt += 2;
1067 
1068 	/*
1069 	 * Sync buffer in case DMA engine currently running. This must be done
1070 	 * before writing the command pointer in the previous descriptor.
1071 	 */
1072 	(void) ddi_dma_sync(qbuf->qb_buf.bi_dma_handle, 0,
1073 	    qbuf->qb_buf.bi_length, DDI_DMA_SYNC_FORDEV);
1074 
1075 	/* save away the status address for quick access in at_next() */
1076 	cmd->qc_status_addr = &desc->status;
1077 
1078 	/*
1079 	 * Setup the command pointer.  This tells the HW where to get the
1080 	 * descriptor we just setup.  This includes the IO address along with
1081 	 * a 4 bit 16 byte block count
1082 	 */
1083 	command_ptr = (uint32_t)((qbuf->qb_cookie[qbuf->qb_ptrs.qp_current_buf
1084 	    ].dmac_address + qbuf->qb_ptrs.qp_offset) | (q_handle->q_block_cnt &
1085 	    DESC_Z_MASK));
1086 
1087 	/*
1088 	 * if we previously setup a descriptor, add this new descriptor into
1089 	 * the previous descriptor's "next" pointer.
1090 	 */
1091 	if (q_handle->q_previous != NULL) {
1092 		ddi_put32(qbuf->qb_buf.bi_handle, &q_handle->q_previous->branch,
1093 		    command_ptr);
1094 		/* Sync buffer again, this gets the command pointer */
1095 		(void) ddi_dma_sync(qbuf->qb_buf.bi_dma_handle, 0,
1096 		    qbuf->qb_buf.bi_length, DDI_DMA_SYNC_FORDEV);
1097 	}
1098 
1099 	/*
1100 	 * this is now the previous descriptor.  Update the current pointer,
1101 	 * clear the block count and reserved size since this is the end of
1102 	 * this command.
1103 	 */
1104 	q_handle->q_previous = (hci1394_desc_t *)desc;
1105 	qbuf->qb_ptrs.qp_current += sizeof (hci1394_desc_imm_t);
1106 	q_handle->q_block_cnt = 0;
1107 	qbuf->qb_ptrs.qp_resv_size = 0;
1108 
1109 	/* save away cleanup info when we are done with the command */
1110 	cmd->qc_descriptor_buf = qbuf->qb_ptrs.qp_current_buf;
1111 	cmd->qc_descriptor_end = qbuf->qb_ptrs.qp_current - 1;
1112 
1113 	/* If the DMA is not running, start it */
1114 	if (q_handle->q_dma_running == B_FALSE) {
1115 		q_handle->q_info.qi_start(q_handle->q_info.qi_callback_arg,
1116 		    command_ptr);
1117 		q_handle->q_dma_running = B_TRUE;
1118 	/* the DMA is running, wake it up */
1119 	} else {
1120 		q_handle->q_info.qi_wake(q_handle->q_info.qi_callback_arg);
1121 	}
1122 
1123 	TNF_PROBE_0_DEBUG(hci1394_q_at_write_OLI_exit, HCI1394_TNF_HAL_STACK,
1124 	    "");
1125 }
1126 
1127 
1128 /*
1129  * hci1394_q_at_write_OL()
1130  *    Write an OL descriptor into the AT descriptor buffer passed in as qbuf.
1131  *    Buffer state information is stored in cmd.  The IO address of the data
1132  *    buffer is passed in io_addr.  Size is the size of the data to be
1133  *    transferred.
1134  */
1135 void
1136 hci1394_q_at_write_OL(hci1394_q_handle_t q_handle, hci1394_q_buf_t *qbuf,
1137     hci1394_q_cmd_t *cmd, uint32_t io_addr, uint_t size)
1138 {
1139 	hci1394_desc_t *desc;
1140 	uint32_t data;
1141 	uint32_t command_ptr;
1142 
1143 
1144 	ASSERT(q_handle != NULL);
1145 	ASSERT(qbuf != NULL);
1146 	ASSERT(cmd != NULL);
1147 	ASSERT(MUTEX_HELD(&q_handle->q_mutex));
1148 	TNF_PROBE_0_DEBUG(hci1394_q_at_write_OL_enter, HCI1394_TNF_HAL_STACK,
1149 	    "");
1150 
1151 	/* make sure enough room for OL */
1152 	ASSERT(qbuf->qb_ptrs.qp_resv_size >= sizeof (hci1394_desc_t));
1153 
1154 	/* Setup OpenHCI OL Header */
1155 	desc = (hci1394_desc_t *)qbuf->qb_ptrs.qp_current;
1156 	data = DESC_AT_OL | (size & DESC_HDR_REQCOUNT_MASK);
1157 	ddi_put32(qbuf->qb_buf.bi_handle, &desc->hdr, data);
1158 	ddi_put32(qbuf->qb_buf.bi_handle, &desc->data_addr, io_addr);
1159 	ddi_put32(qbuf->qb_buf.bi_handle, &desc->branch, 0);
1160 	ddi_put32(qbuf->qb_buf.bi_handle, &desc->status, 0);
1161 
1162 	/*
1163 	 * We wrote 1 16 byte block in the descriptor buffer, update the count
1164 	 * accordingly.
1165 	 */
1166 	q_handle->q_block_cnt++;
1167 
1168 	/*
1169 	 * Sync buffer in case DMA engine currently running. This must be done
1170 	 * before writing the command pointer in the previous descriptor.
1171 	 */
1172 	(void) ddi_dma_sync(qbuf->qb_buf.bi_dma_handle, 0,
1173 	    qbuf->qb_buf.bi_length, DDI_DMA_SYNC_FORDEV);
1174 
1175 	/* save away the status address for quick access in at_next() */
1176 	cmd->qc_status_addr = &desc->status;
1177 
1178 	/*
1179 	 * Setup the command pointer.  This tells the HW where to get the
1180 	 * descriptor we just setup.  This includes the IO address along with
1181 	 * a 4 bit 16 byte block count
1182 	 */
1183 	command_ptr = (uint32_t)((qbuf->qb_cookie[qbuf->qb_ptrs.qp_current_buf
1184 	    ].dmac_address + qbuf->qb_ptrs.qp_offset) | (q_handle->q_block_cnt &
1185 	    DESC_Z_MASK));
1186 
1187 	/*
1188 	 * if we previously setup a descriptor, add this new descriptor into
1189 	 * the previous descriptor's "next" pointer.
1190 	 */
1191 	if (q_handle->q_previous != NULL) {
1192 		ddi_put32(qbuf->qb_buf.bi_handle, &q_handle->q_previous->branch,
1193 		    command_ptr);
1194 		/* Sync buffer again, this gets the command pointer */
1195 		(void) ddi_dma_sync(qbuf->qb_buf.bi_dma_handle, 0,
1196 		    qbuf->qb_buf.bi_length, DDI_DMA_SYNC_FORDEV);
1197 	}
1198 
1199 	/*
1200 	 * this is now the previous descriptor.  Update the current pointer,
1201 	 * clear the block count and reserved size since this is the end of
1202 	 * this command.
1203 	 */
1204 	q_handle->q_previous = desc;
1205 	qbuf->qb_ptrs.qp_current += sizeof (hci1394_desc_t);
1206 	q_handle->q_block_cnt = 0;
1207 	qbuf->qb_ptrs.qp_resv_size = 0;
1208 
1209 	/* save away cleanup info when we are done with the command */
1210 	cmd->qc_descriptor_buf = qbuf->qb_ptrs.qp_current_buf;
1211 	cmd->qc_descriptor_end = qbuf->qb_ptrs.qp_current - 1;
1212 
1213 	/* If the DMA is not running, start it */
1214 	if (q_handle->q_dma_running == B_FALSE) {
1215 		q_handle->q_info.qi_start(q_handle->q_info.qi_callback_arg,
1216 		    command_ptr);
1217 		q_handle->q_dma_running = B_TRUE;
1218 	/* the DMA is running, wake it up */
1219 	} else {
1220 		q_handle->q_info.qi_wake(q_handle->q_info.qi_callback_arg);
1221 	}
1222 
1223 	TNF_PROBE_0_DEBUG(hci1394_q_at_write_OL_exit, HCI1394_TNF_HAL_STACK,
1224 	    "");
1225 }
1226 
1227 
1228 /*
1229  * hci1394_q_at_rep_put8()
1230  *    Copy a byte stream from a kernel virtual address (data) to a IO mapped
1231  *    data buffer (qbuf).  Copy datasize bytes.  State information for the
1232  *    data buffer is kept in cmd.
1233  */
1234 void
1235 hci1394_q_at_rep_put8(hci1394_q_buf_t *qbuf, hci1394_q_cmd_t *cmd,
1236     uint8_t *data, uint_t datasize)
1237 {
1238 	ASSERT(qbuf != NULL);
1239 	ASSERT(cmd != NULL);
1240 	ASSERT(data != NULL);
1241 	TNF_PROBE_0_DEBUG(hci1394_q_at_rep_put8_enter, HCI1394_TNF_HAL_STACK,
1242 	    "");
1243 
1244 	/* Make sure enough room for data */
1245 	ASSERT(qbuf->qb_ptrs.qp_resv_size >= datasize);
1246 
1247 	/* Copy in data into the data buffer */
1248 	ddi_rep_put8(qbuf->qb_buf.bi_handle, data,
1249 	    (uint8_t *)qbuf->qb_ptrs.qp_current, datasize, DDI_DEV_AUTOINCR);
1250 
1251 	/* Update the current pointer, offset, and reserved size */
1252 	qbuf->qb_ptrs.qp_current += datasize;
1253 	qbuf->qb_ptrs.qp_offset = (uint32_t)(qbuf->qb_ptrs.qp_current -
1254 	    qbuf->qb_ptrs.qp_begin);
1255 	qbuf->qb_ptrs.qp_resv_size -= datasize;
1256 
1257 	/* save away cleanup info when we are done with the command */
1258 	cmd->qc_data_used = B_TRUE;
1259 	cmd->qc_data_buf = qbuf->qb_ptrs.qp_current_buf;
1260 	cmd->qc_data_end = qbuf->qb_ptrs.qp_current - 1;
1261 
1262 	/* Sync data buffer */
1263 	(void) ddi_dma_sync(qbuf->qb_buf.bi_dma_handle, 0,
1264 	    qbuf->qb_buf.bi_length, DDI_DMA_SYNC_FORDEV);
1265 
1266 	TNF_PROBE_0_DEBUG(hci1394_q_at_rep_put8_exit, HCI1394_TNF_HAL_STACK,
1267 	    "");
1268 }
1269 
1270 
1271 /*
1272  * hci1394_q_at_copy_from_mblk()
1273  *    Copy a byte stream from a mblk(s) to a IO mapped data buffer (qbuf).
1274  *    Copy mblk->length bytes. The services layer and the hal use a private
1275  *    structure (h1394_mblk_t) to keep track of how much of the mblk to send
1276  *    since we may have to break the transfer up into smaller blocks. (i.e. a
1277  *    1MByte block write would go out in 2KByte chunks. State information for
1278  *    the data buffer is kept in cmd.
1279  */
1280 static void
1281 hci1394_q_at_copy_from_mblk(hci1394_q_buf_t *qbuf, hci1394_q_cmd_t *cmd,
1282     h1394_mblk_t *mblk)
1283 {
1284 	uint_t bytes_left;
1285 	uint_t length;
1286 
1287 
1288 	ASSERT(qbuf != NULL);
1289 	ASSERT(cmd != NULL);
1290 	ASSERT(mblk != NULL);
1291 	TNF_PROBE_0_DEBUG(hci1394_q_at_copy_from_mblk_enter,
1292 	    HCI1394_TNF_HAL_STACK, "");
1293 
1294 	/* We return these variables to the Services Layer when we are done */
1295 	mblk->next_offset = mblk->curr_offset;
1296 	mblk->next_mblk = mblk->curr_mblk;
1297 	bytes_left = mblk->length;
1298 
1299 	/* do while there are bytes left to copy */
1300 	do {
1301 		/*
1302 		 * If the entire data portion of the current block transfer is
1303 		 * contained within a single mblk.
1304 		 */
1305 		if ((mblk->next_offset + bytes_left) <=
1306 		    (mblk->next_mblk->b_wptr)) {
1307 			/* Copy the data into the data Q */
1308 			hci1394_q_at_rep_put8(qbuf, cmd,
1309 			    (uint8_t *)mblk->next_offset, bytes_left);
1310 
1311 			/* increment the mblk offset */
1312 			mblk->next_offset += bytes_left;
1313 
1314 			/* we have no more bytes to put into the buffer */
1315 			bytes_left = 0;
1316 
1317 			/*
1318 			 * If our offset is at the end of data in this mblk, go
1319 			 * to the next mblk.
1320 			 */
1321 			if (mblk->next_offset >= mblk->next_mblk->b_wptr) {
1322 				mblk->next_mblk = mblk->next_mblk->b_cont;
1323 				if (mblk->next_mblk != NULL) {
1324 					mblk->next_offset =
1325 					    mblk->next_mblk->b_rptr;
1326 				}
1327 			}
1328 
1329 		/*
1330 		 * The data portion of the current block transfer is spread
1331 		 * across two or more mblk's
1332 		 */
1333 		} else {
1334 			/*
1335 			 * Figure out how much data is in this mblk.
1336 			 */
1337 			length = mblk->next_mblk->b_wptr - mblk->next_offset;
1338 
1339 			/* Copy the data into the atreq data Q */
1340 			hci1394_q_at_rep_put8(qbuf, cmd,
1341 			    (uint8_t *)mblk->next_offset, length);
1342 
1343 			/* update the bytes left count, go to the next mblk */
1344 			bytes_left = bytes_left - length;
1345 			mblk->next_mblk = mblk->next_mblk->b_cont;
1346 			ASSERT(mblk->next_mblk != NULL);
1347 			mblk->next_offset = mblk->next_mblk->b_rptr;
1348 		}
1349 	} while (bytes_left > 0);
1350 
1351 	TNF_PROBE_0_DEBUG(hci1394_q_at_copy_from_mblk_exit,
1352 	    HCI1394_TNF_HAL_STACK, "");
1353 }
1354 
1355 
1356 /*
1357  * hci1394_q_ar_next()
1358  *    Return an address to the next received AR packet.  If there are no more
1359  *    AR packets in the buffer, q_addr will be set to NULL.
1360  */
1361 void
1362 hci1394_q_ar_next(hci1394_q_handle_t q_handle, uint32_t **q_addr)
1363 {
1364 	hci1394_desc_t *desc;
1365 	hci1394_q_buf_t *descb;
1366 	hci1394_q_buf_t *datab;
1367 	uint32_t residual_count;
1368 
1369 
1370 	ASSERT(q_handle != NULL);
1371 	ASSERT(q_addr != NULL);
1372 	TNF_PROBE_0_DEBUG(hci1394_q_ar_next_enter, HCI1394_TNF_HAL_STACK, "");
1373 
1374 	descb = &q_handle->q_desc;
1375 	datab = &q_handle->q_data;
1376 
1377 	/* Sync Descriptor buffer */
1378 	(void) ddi_dma_sync(descb->qb_buf.bi_dma_handle, 0,
1379 	    descb->qb_buf.bi_length, DDI_DMA_SYNC_FORKERNEL);
1380 
1381 	/*
1382 	 * Check residual in current IM count vs q_space_left to see if we have
1383 	 * received any more responses
1384 	 */
1385 	desc = (hci1394_desc_t *)q_handle->q_head;
1386 	residual_count = ddi_get32(descb->qb_buf.bi_handle, &desc->status);
1387 	residual_count &= DESC_ST_RESCOUNT_MASK;
1388 	if (residual_count >= q_handle->q_space_left) {
1389 		/* No new packets received */
1390 		*q_addr = NULL;
1391 		TNF_PROBE_0_DEBUG(hci1394_q_ar_next_exit,
1392 		    HCI1394_TNF_HAL_STACK, "");
1393 		return;
1394 	}
1395 
1396 	/* Sync Data Q */
1397 	(void) ddi_dma_sync(datab->qb_buf.bi_dma_handle, 0,
1398 	    datab->qb_buf.bi_length, DDI_DMA_SYNC_FORKERNEL);
1399 
1400 	/*
1401 	 * We have a new packet, return the address of the start of the
1402 	 * packet.
1403 	 */
1404 	*q_addr = (uint32_t *)datab->qb_ptrs.qp_current;
1405 
1406 	TNF_PROBE_0_DEBUG(hci1394_q_ar_next_exit, HCI1394_TNF_HAL_STACK, "");
1407 }
1408 
1409 
1410 /*
1411  * hci1394_q_ar_free()
1412  *    Free the space used by the AR packet at the top of the data buffer. AR
1413  *    packets are processed in the order that they are received.  This will
1414  *    free the oldest received packet which has not yet been freed.  size is
1415  *    how much space the packet takes up.
1416  */
1417 void
1418 hci1394_q_ar_free(hci1394_q_handle_t q_handle, uint_t size)
1419 {
1420 	hci1394_q_buf_t *descb;
1421 	hci1394_q_buf_t *datab;
1422 
1423 
1424 	ASSERT(q_handle != NULL);
1425 	TNF_PROBE_0_DEBUG(hci1394_q_ar_free_enter, HCI1394_TNF_HAL_STACK, "");
1426 
1427 	descb = &q_handle->q_desc;
1428 	datab = &q_handle->q_data;
1429 
1430 	/*
1431 	 * Packet is in multiple buffers. Theoretically a buffer could be broken
1432 	 * in more than two buffers for an ARRESP.  Since the buffers should be
1433 	 * in at least 4K increments this will not happen since the max packet
1434 	 * size is 2KBytes.
1435 	 */
1436 	if ((datab->qb_ptrs.qp_current + size) > datab->qb_ptrs.qp_end) {
1437 		/* Add IM descriptor for used buffer back into Q */
1438 		hci1394_q_ar_write_IM(q_handle, descb,
1439 		    datab->qb_cookie[datab->qb_ptrs.qp_current_buf
1440 		    ].dmac_address,
1441 		    datab->qb_cookie[datab->qb_ptrs.qp_current_buf].dmac_size);
1442 
1443 		/* Go to the next buffer */
1444 		hci1394_q_next_buf(datab);
1445 
1446 		/* Update next buffers pointers for partial packet */
1447 		size -= q_handle->q_space_left;
1448 		datab->qb_ptrs.qp_current += size;
1449 		q_handle->q_space_left =
1450 		    datab->qb_cookie[datab->qb_ptrs.qp_current_buf].dmac_size -
1451 		    size;
1452 
1453 		/* Change the head pointer to the next IM descriptor */
1454 		q_handle->q_head += sizeof (hci1394_desc_t);
1455 		if ((q_handle->q_head + sizeof (hci1394_desc_t)) >
1456 		    (descb->qb_ptrs.qp_bottom + 1)) {
1457 			q_handle->q_head = descb->qb_ptrs.qp_top;
1458 		}
1459 
1460 	/* Packet is only in one buffer */
1461 	} else {
1462 		q_handle->q_space_left -= size;
1463 		datab->qb_ptrs.qp_current += size;
1464 	}
1465 
1466 	TNF_PROBE_0_DEBUG(hci1394_q_ar_free_exit, HCI1394_TNF_HAL_STACK, "");
1467 }
1468 
1469 
1470 /*
1471  * hci1394_q_ar_get32()
1472  *    Read a quadlet of data regardless if it is in the current buffer or has
1473  *    wrapped to the top buffer.  If the address passed to this routine is
1474  *    passed the bottom of the data buffer, this routine will automatically
1475  *    wrap back to the top of the Q and look in the correct offset from the
1476  *    top. Copy the data into the kernel virtual address provided.
1477  */
1478 uint32_t
1479 hci1394_q_ar_get32(hci1394_q_handle_t q_handle, uint32_t *addr)
1480 {
1481 	hci1394_q_buf_t *data;
1482 	uintptr_t new_addr;
1483 	uint32_t data32;
1484 
1485 
1486 	ASSERT(q_handle != NULL);
1487 	ASSERT(addr != NULL);
1488 	TNF_PROBE_0_DEBUG(hci1394_q_get32_enter, HCI1394_TNF_HAL_STACK, "");
1489 
1490 	data = &q_handle->q_data;
1491 
1492 	/*
1493 	 * if the data has wrapped to the top of the buffer, adjust the address.
1494 	 */
1495 	if ((uintptr_t)addr > (uintptr_t)data->qb_ptrs.qp_bottom) {
1496 		new_addr = (uintptr_t)data->qb_ptrs.qp_top + ((uintptr_t)addr -
1497 		    ((uintptr_t)data->qb_ptrs.qp_bottom + (uintptr_t)1));
1498 		data32 = ddi_get32(data->qb_buf.bi_handle,
1499 		    (uint32_t *)new_addr);
1500 
1501 	/* data is before end of buffer */
1502 	} else {
1503 		data32 = ddi_get32(data->qb_buf.bi_handle, addr);
1504 	}
1505 
1506 	TNF_PROBE_0_DEBUG(hci1394_q_get32_exit, HCI1394_TNF_HAL_STACK, "");
1507 
1508 	return (data32);
1509 }
1510 
1511 
1512 /*
1513  * hci1394_q_ar_rep_get8()
1514  *    Read a byte stream of data regardless if it is contiguous or has partially
1515  *    or fully wrapped to the top buffer.  If the address passed to this routine
1516  *    is passed the bottom of the data buffer, or address + size is past the
1517  *    bottom of the data buffer. this routine will automatically wrap back to
1518  *    the top of the Q and look in the correct offset from the top. Copy the
1519  *    data into the kernel virtual address provided.
1520  */
1521 void
1522 hci1394_q_ar_rep_get8(hci1394_q_handle_t q_handle, uint8_t *dest,
1523     uint8_t *q_addr, uint_t size)
1524 {
1525 	hci1394_q_buf_t *data;
1526 	uintptr_t new_addr;
1527 	uint_t new_size;
1528 	uintptr_t new_dest;
1529 
1530 
1531 	ASSERT(q_handle != NULL);
1532 	ASSERT(dest != NULL);
1533 	ASSERT(q_addr != NULL);
1534 	TNF_PROBE_0_DEBUG(hci1394_q_ar_rep_get8_enter, HCI1394_TNF_HAL_STACK,
1535 	    "");
1536 
1537 	data = &q_handle->q_data;
1538 
1539 	/*
1540 	 * There are three cases:
1541 	 *   1) All of the data has wrapped.
1542 	 *   2) Some of the data has not wrapped and some has wrapped.
1543 	 *   3) None of the data has wrapped.
1544 	 */
1545 
1546 	/* All of the data has wrapped, just adjust the starting address */
1547 	if ((uintptr_t)q_addr > (uintptr_t)data->qb_ptrs.qp_bottom) {
1548 		new_addr = (uintptr_t)data->qb_ptrs.qp_top +
1549 		    ((uintptr_t)q_addr - ((uintptr_t)data->qb_ptrs.qp_bottom +
1550 		    (uintptr_t)1));
1551 		ddi_rep_get8(data->qb_buf.bi_handle, dest, (uint8_t *)new_addr,
1552 		    size, DDI_DEV_AUTOINCR);
1553 
1554 	/*
1555 	 * Some of the data has wrapped. Copy the data that hasn't wrapped,
1556 	 * adjust the address, then copy the rest.
1557 	 */
1558 	} else if (((uintptr_t)q_addr + (uintptr_t)size) >
1559 	    ((uintptr_t)data->qb_ptrs.qp_bottom + (uintptr_t)1)) {
1560 		/* Copy first half */
1561 		new_size = (uint_t)(((uintptr_t)data->qb_ptrs.qp_bottom +
1562 		    (uintptr_t)1) - (uintptr_t)q_addr);
1563 		ddi_rep_get8(data->qb_buf.bi_handle, dest, q_addr, new_size,
1564 		    DDI_DEV_AUTOINCR);
1565 
1566 		/* copy second half */
1567 		new_dest = (uintptr_t)dest + (uintptr_t)new_size;
1568 		new_size = size - new_size;
1569 		new_addr = (uintptr_t)data->qb_ptrs.qp_top;
1570 		ddi_rep_get8(data->qb_buf.bi_handle, (uint8_t *)new_dest,
1571 		    (uint8_t *)new_addr, new_size, DDI_DEV_AUTOINCR);
1572 
1573 	/* None of the data has wrapped */
1574 	} else {
1575 		ddi_rep_get8(data->qb_buf.bi_handle, dest, q_addr, size,
1576 		    DDI_DEV_AUTOINCR);
1577 	}
1578 
1579 	TNF_PROBE_0_DEBUG(hci1394_q_ar_rep_get8_exit, HCI1394_TNF_HAL_STACK,
1580 	    "");
1581 }
1582 
1583 
1584 /*
1585  * hci1394_q_ar_copy_to_mblk()
1586  *    Read a byte stream of data regardless if it is contiguous or has partially
1587  *    or fully wrapped to the top buffer.  If the address passed to this routine
1588  *    is passed the bottom of the data buffer, or address + size is passed the
1589  *    bottom of the data buffer. this routine will automatically wrap back to
1590  *    the top of the Q and look in the correct offset from the top. Copy the
1591  *    data into the mblk provided. The services layer and the hal use a private
1592  *    structure (h1394_mblk_t) to keep track of how much of the mblk to receive
1593  *    into since we may have to break the transfer up into smaller blocks.
1594  *    (i.e. a 1MByte block read would go out in 2KByte requests.
1595  */
1596 void
1597 hci1394_q_ar_copy_to_mblk(hci1394_q_handle_t q_handle, uint8_t *addr,
1598     h1394_mblk_t *mblk)
1599 {
1600 	uint8_t *new_addr;
1601 	uint_t bytes_left;
1602 	uint_t length;
1603 
1604 
1605 	ASSERT(q_handle != NULL);
1606 	ASSERT(addr != NULL);
1607 	ASSERT(mblk != NULL);
1608 	TNF_PROBE_0_DEBUG(hci1394_q_copy_to_mblk_enter,
1609 	    HCI1394_TNF_HAL_STACK, "");
1610 
1611 	/* We return these variables to the Services Layer when we are done */
1612 	mblk->next_offset = mblk->curr_offset;
1613 	mblk->next_mblk = mblk->curr_mblk;
1614 	bytes_left = mblk->length;
1615 
1616 	/* the address we copy from will change as we change mblks */
1617 	new_addr = addr;
1618 
1619 	/* do while there are bytes left to copy */
1620 	do {
1621 		/*
1622 		 * If the entire data portion of the current block transfer is
1623 		 * contained within a single mblk.
1624 		 */
1625 		if ((mblk->next_offset + bytes_left) <=
1626 		    (mblk->next_mblk->b_datap->db_lim)) {
1627 			/* Copy the data into the mblk */
1628 			hci1394_q_ar_rep_get8(q_handle,
1629 			    (uint8_t *)mblk->next_offset, new_addr, bytes_left);
1630 
1631 			/* increment the offset */
1632 			mblk->next_offset += bytes_left;
1633 			mblk->next_mblk->b_wptr = mblk->next_offset;
1634 
1635 			/* we have no more bytes to put into the buffer */
1636 			bytes_left = 0;
1637 
1638 			/*
1639 			 * If our offset is at the end of data in this mblk, go
1640 			 * to the next mblk.
1641 			 */
1642 			if (mblk->next_offset >=
1643 			    mblk->next_mblk->b_datap->db_lim) {
1644 				mblk->next_mblk = mblk->next_mblk->b_cont;
1645 				if (mblk->next_mblk != NULL) {
1646 					mblk->next_offset =
1647 					    mblk->next_mblk->b_wptr;
1648 				}
1649 			}
1650 
1651 		/*
1652 		 * The data portion of the current block transfer is spread
1653 		 * across two or more mblk's
1654 		 */
1655 		} else {
1656 			/* Figure out how much data is in this mblk */
1657 			length = mblk->next_mblk->b_datap->db_lim -
1658 			    mblk->next_offset;
1659 
1660 			/* Copy the data into the mblk */
1661 			hci1394_q_ar_rep_get8(q_handle,
1662 			    (uint8_t *)mblk->next_offset, new_addr, length);
1663 			mblk->next_mblk->b_wptr =
1664 			    mblk->next_mblk->b_datap->db_lim;
1665 
1666 			/*
1667 			 * update the bytes left and address to copy from, go
1668 			 * to the next mblk.
1669 			 */
1670 			bytes_left = bytes_left - length;
1671 			new_addr = (uint8_t *)((uintptr_t)new_addr +
1672 			    (uintptr_t)length);
1673 			mblk->next_mblk = mblk->next_mblk->b_cont;
1674 			ASSERT(mblk->next_mblk != NULL);
1675 			mblk->next_offset = mblk->next_mblk->b_wptr;
1676 		}
1677 	} while (bytes_left > 0);
1678 
1679 	TNF_PROBE_0_DEBUG(hci1394_q_copy_to_mblk_exit,
1680 	    HCI1394_TNF_HAL_STACK, "");
1681 }
1682 
1683 
1684 /*
1685  * hci1394_q_ar_write_IM()
1686  *    Write an IM descriptor into the AR descriptor buffer passed in as qbuf.
1687  *    The IO address of the data buffer is passed in io_addr.  datasize is the
1688  *    size of the data data buffer to receive into.
1689  */
1690 void
1691 hci1394_q_ar_write_IM(hci1394_q_handle_t q_handle, hci1394_q_buf_t *qbuf,
1692     uint32_t io_addr, uint_t datasize)
1693 {
1694 	hci1394_desc_t *desc;
1695 	uint32_t data;
1696 	uint32_t command_ptr;
1697 
1698 
1699 	ASSERT(q_handle != NULL);
1700 	ASSERT(qbuf != NULL);
1701 	TNF_PROBE_0_DEBUG(hci1394_q_ar_write_IM_enter, HCI1394_TNF_HAL_STACK,
1702 	    "");
1703 
1704 	/* Make sure enough room for IM */
1705 	if ((qbuf->qb_ptrs.qp_current + sizeof (hci1394_desc_t)) >
1706 	    (qbuf->qb_ptrs.qp_bottom + 1)) {
1707 		hci1394_q_next_buf(qbuf);
1708 	} else {
1709 		/* Store the offset of the top of this descriptor block */
1710 		qbuf->qb_ptrs.qp_offset = (uint32_t)(qbuf->qb_ptrs.qp_current -
1711 		    qbuf->qb_ptrs.qp_begin);
1712 	}
1713 
1714 	/* Setup OpenHCI IM Header */
1715 	desc = (hci1394_desc_t *)qbuf->qb_ptrs.qp_current;
1716 	data = DESC_AR_IM | (datasize & DESC_HDR_REQCOUNT_MASK);
1717 	ddi_put32(qbuf->qb_buf.bi_handle, &desc->hdr, data);
1718 	ddi_put32(qbuf->qb_buf.bi_handle, &desc->data_addr, io_addr);
1719 	ddi_put32(qbuf->qb_buf.bi_handle, &desc->branch, 0);
1720 	ddi_put32(qbuf->qb_buf.bi_handle, &desc->status, datasize &
1721 	    DESC_ST_RESCOUNT_MASK);
1722 
1723 	/*
1724 	 * Sync buffer in case DMA engine currently running. This must be done
1725 	 * before writing the command pointer in the previous descriptor.
1726 	 */
1727 	(void) ddi_dma_sync(qbuf->qb_buf.bi_dma_handle, 0,
1728 	    qbuf->qb_buf.bi_length, DDI_DMA_SYNC_FORDEV);
1729 
1730 	/*
1731 	 * Setup the command pointer.  This tells the HW where to get the
1732 	 * descriptor we just setup.  This includes the IO address along with
1733 	 * a 4 bit 16 byte block count.  We only wrote 1 16 byte block.
1734 	 */
1735 	command_ptr = (uint32_t)((qbuf->qb_cookie[qbuf->qb_ptrs.qp_current_buf
1736 	    ].dmac_address + qbuf->qb_ptrs.qp_offset) | 1);
1737 
1738 	/*
1739 	 * if we previously setup a descriptor, add this new descriptor into
1740 	 * the previous descriptor's "next" pointer.
1741 	 */
1742 	if (q_handle->q_previous != NULL) {
1743 		ddi_put32(qbuf->qb_buf.bi_handle,
1744 		    &q_handle->q_previous->branch, command_ptr);
1745 		/* Sync buffer again, this gets the command pointer */
1746 		(void) ddi_dma_sync(qbuf->qb_buf.bi_dma_handle, 0,
1747 		    qbuf->qb_buf.bi_length, DDI_DMA_SYNC_FORDEV);
1748 	}
1749 
1750 	/* this is the new previous descriptor.  Update the current pointer */
1751 	q_handle->q_previous = desc;
1752 	qbuf->qb_ptrs.qp_current += sizeof (hci1394_desc_t);
1753 
1754 	/* If the DMA is not running, start it */
1755 	if (q_handle->q_dma_running == B_FALSE) {
1756 		q_handle->q_info.qi_start(q_handle->q_info.qi_callback_arg,
1757 		    command_ptr);
1758 		q_handle->q_dma_running = B_TRUE;
1759 	/* the DMA is running, wake it up */
1760 	} else {
1761 		q_handle->q_info.qi_wake(q_handle->q_info.qi_callback_arg);
1762 	}
1763 
1764 	TNF_PROBE_0_DEBUG(hci1394_q_ar_write_IM_exit, HCI1394_TNF_HAL_STACK,
1765 	    "");
1766 }
1767