/illumos-gate/usr/src/uts/common/fs/nfs/ |
H A D | nfs4_srv.c | 3564 uio.uio_iov = &iov; in rfs4_op_read() 3929 struct iovec iov; in rfs4_op_readlink() local 4007 iov.iov_base = data; in rfs4_op_readlink() 4008 iov.iov_len = MAXPATHLEN; in rfs4_op_readlink() 4009 uio.uio_iov = &iov; in rfs4_op_readlink() 5776 struct iovec iov[MAX_IOVECS]; in rfs4_op_write() local 5889 iovp = iov; in rfs4_op_write() 5896 iovp = iov; in rfs4_op_write() 5901 iovp = iov; in rfs4_op_write() 5925 if (iovp != iov) in rfs4_op_write() [all …]
|
H A D | nfs_subr.c | 5233 struct iovec iov; in do_xattr_exists_check() local 5243 uio.uio_iov = &iov; in do_xattr_exists_check() 5250 iov.iov_base = dbuf; in do_xattr_exists_check() 5251 iov.iov_len = dlen; in do_xattr_exists_check()
|
H A D | nfs_srv.c | 617 struct iovec iov; in rfs_readlink() local 697 uio.uio_iov = &iov; in rfs_readlink() 768 struct iovec iov; in rfs_read() local 928 uio.uio_iov = &iov; in rfs_read() 1154 uio.uio_iov = iov; in rfs_write_sync() 1190 iovp = iov; in rfs_write_sync() 1226 if (iovp != iov) in rfs_write_sync() 1644 niovp = iov; in rfs_write() 1723 if (niovp != iov) in rfs_write() 2651 struct iovec iov; in rfs_readdir() local [all …]
|
H A D | nfs_server.c | 3326 rfs_read_alloc(uint_t len, struct iovec **iov, int *iovcnt) in rfs_read_alloc() argument 3337 *iov = iovarr; in rfs_read_alloc()
|
H A D | nfs4_state.c | 466 struct iovec iov[3]; in rfs4_ss_getstate() local 526 iov[0].iov_len = sizeof (int); in rfs4_ss_getstate() 532 uio.uio_iov = iov; in rfs4_ss_getstate() 572 iov[0].iov_len = id_len; in rfs4_ss_getstate() 574 uio.uio_iov = iov; in rfs4_ss_getstate() 637 struct iovec iov; in rfs4_ss_oldstate() local 659 iov.iov_base = dirt; in rfs4_ss_oldstate() 661 uio.uio_iov = &iov; in rfs4_ss_oldstate() 998 struct iovec iov[4]; in rfs4_ss_clid_write_one() local 1044 iov[0].iov_len = sizeof (int); in rfs4_ss_clid_write_one() [all …]
|
/illumos-gate/exception_lists/ |
H A D | cstyle | 1206 usr/src/cmd/bhyve/iov.[ch]
|
/illumos-gate/usr/src/uts/common/fs/smbsrv/ |
H A D | smb2_durable.c | 874 iovec_t iov; in smb2_dh_read_nvlist() local 900 iov.iov_base = fbuf; in smb2_dh_read_nvlist() 901 iov.iov_len = flen; in smb2_dh_read_nvlist() 902 uio.uio_iov = &iov; in smb2_dh_read_nvlist() 1139 iovec_t iov; in smb2_dh_update_nvfile() local 1175 iov.iov_base = (void *) buf; in smb2_dh_update_nvfile() 1176 iov.iov_len = buflen; in smb2_dh_update_nvfile() 1177 uio.uio_iov = &iov; in smb2_dh_update_nvfile()
|
/illumos-gate/usr/src/lib/libvmmapi/common/ |
H A D | vmmapi.c | 1820 iov[i].iov_base = 0; in vm_copy_setup() 1821 iov[i].iov_len = 0; in vm_copy_setup() 1837 iov->iov_base = va; in vm_copy_setup() 1838 iov->iov_len = n; in vm_copy_setup() 1839 iov++; in vm_copy_setup() 1867 assert(iov->iov_len); in vm_copyin() 1869 src = iov->iov_base; in vm_copyin() 1872 iov++; in vm_copyin() 1887 assert(iov->iov_len); in vm_copyout() 1889 dst = iov->iov_base; in vm_copyout() [all …]
|
H A D | vmmapi.h | 306 uint64_t gla, size_t len, int prot, struct iovec *iov, int iovcnt, 310 void vm_copy_teardown(struct iovec *iov, int iovcnt);
|
/illumos-gate/usr/src/cmd/bhyve/ |
H A D | virtio.c | 217 _vq_record(int i, struct vring_desc *vd, struct vmctx *ctx, struct iovec *iov, in _vq_record() argument 222 iov[i].iov_base = paddr_guest2host(ctx, vd->addr, vd->len); in _vq_record() 223 iov[i].iov_len = vd->len; in _vq_record() 266 vq_getchain(struct vqueue_info *vq, struct iovec *iov, int niov, in vq_getchain() argument 328 _vq_record(i, vdir, ctx, iov, niov, &req); in vq_getchain() 365 _vq_record(i, vp, ctx, iov, niov, &req); in vq_getchain()
|
H A D | virtio.h | 429 int vq_getchain(struct vqueue_info *vq, struct iovec *iov, int niov,
|
H A D | task_switch.c | 221 struct iovec iov[2]; in desc_table_rw() local 232 sizeof(*desc), doread ? PROT_READ : PROT_WRITE, iov, nitems(iov), in desc_table_rw() 238 vm_copyin(iov, desc, sizeof(*desc)); in desc_table_rw() 240 vm_copyout(desc, iov, sizeof(*desc)); in desc_table_rw() 449 uint32_t eip, struct tss32 *tss, struct iovec *iov) in tss32_save() argument 477 vm_copyout(tss, iov, sizeof(struct tss32)); in tss32_save() 494 uint16_t ot_sel, struct tss32 *tss, struct iovec *iov, int *faultptr) in tss32_restore() argument 573 vm_copyout(tss, iov, sizeof(*tss)); in tss32_restore() 784 struct iovec iov[2]; in push_errcode() local 838 iov, nitems(iov), faultptr); in push_errcode() [all …]
|
H A D | pci_virtio_scsi.c | 580 struct iovec iov[VTSCSI_MAXSEG]; in pci_vtscsi_controlq_notify() local 589 n = vq_getchain(vq, iov, VTSCSI_MAXSEG, &req); in pci_vtscsi_controlq_notify() 592 bufsize = iov_to_buf(iov, n, &buf); in pci_vtscsi_controlq_notify() 594 buf_to_iov((uint8_t *)buf + bufsize - iolen, iolen, iov, n, in pci_vtscsi_controlq_notify() 618 struct iovec iov[VTSCSI_MAXSEG]; in pci_vtscsi_requestq_notify() local 626 n = vq_getchain(vq, iov, VTSCSI_MAXSEG, &vireq); in pci_vtscsi_requestq_notify() 634 memcpy(req->vsr_iov_in, iov, in pci_vtscsi_requestq_notify() 636 memcpy(req->vsr_iov_out, iov + vireq.readable, in pci_vtscsi_requestq_notify()
|
H A D | pci_virtio_viona.c | 297 struct iovec *iov, size_t niov) in pci_viona_control_rx() argument 307 v = *(uint8_t *)iov[0].iov_base; in pci_viona_control_rx() 342 iov->iov_len - sizeof (table->vncm_entries)) { in pci_viona_control_mac_dump() 343 DPRINTF(" Bad table size %u", iov->iov_len); in pci_viona_control_mac_dump() 355 struct iovec *iov, size_t niov) in pci_viona_control_mac() argument 384 pci_viona_control_mac_dump("UNICAST", &iov[0]); in pci_viona_control_mac() 412 struct iovec iov[VIONA_CTLQ_MAXSEGS + 1]; in pci_viona_control() local 414 struct iovec *siov = iov; in pci_viona_control() 453 ackp = (uint8_t *)iov[req.readable].iov_base; in pci_viona_control() 455 siov = &iov[1]; in pci_viona_control() [all …]
|
H A D | pci_virtio_input.c | 131 struct iovec iov; member 192 struct iovec iov; in pci_vtinput_notify_statusq() local 194 const int n = vq_getchain(vq, &iov, 1, &req); in pci_vtinput_notify_statusq() 202 memcpy(&event, iov.iov_base, sizeof(event)); in pci_vtinput_notify_statusq() 527 struct iovec iov; in vtinput_eventqueue_send_events() local 529 const int n = vq_getchain(vq, &iov, 1, &req); in vtinput_eventqueue_send_events() 542 if (iov.iov_len < sizeof(struct vtinput_event)) { in vtinput_eventqueue_send_events() 544 iov.iov_len)); in vtinput_eventqueue_send_events() 551 queue->events[i].iov = iov; in vtinput_eventqueue_send_events() 561 memcpy(event.iov.iov_base, &event.event, in vtinput_eventqueue_send_events()
|
H A D | pci_virtio_net.c | 196 if (iov[0].iov_len < hlen) { in iov_trim_hdr() 204 iov[0].iov_len -= hlen; in iov_trim_hdr() 205 if (iov[0].iov_len == 0) { in iov_trim_hdr() 214 riov = &iov[1]; in iov_trim_hdr() 216 iov[0].iov_base = (void *)((uintptr_t)iov[0].iov_base + hlen); in iov_trim_hdr() 217 riov = &iov[0]; in iov_trim_hdr() 234 struct iovec iov[VTNET_MAXSEGS + 1]; in pci_vtnet_rx() local 274 riov = iov; in pci_vtnet_rx() 338 riov = iov; in pci_vtnet_rx() 448 struct iovec iov[VTNET_MAXSEGS + 1]; in pci_vtnet_proctx() local [all …]
|
H A D | pci_virtio_rnd.c | 109 struct iovec iov; in pci_vtrnd_notify() local 122 n = vq_getchain(vq, &iov, 1, &req); in pci_vtrnd_notify() 125 len = read(sc->vrsc_fd, iov.iov_base, iov.iov_len); in pci_vtrnd_notify()
|
H A D | pci_virtio_9p.c | 158 pci_vt9p_get_buffer(struct l9p_request *req, struct iovec *iov, size_t *niov, in pci_vt9p_get_buffer() argument 164 memcpy(iov, preq->vsr_iov + preq->vsr_respidx, in pci_vt9p_get_buffer() 171 pci_vt9p_send(struct l9p_request *req, const struct iovec *iov __unused, in pci_vt9p_send() 188 pci_vt9p_drop(struct l9p_request *req, const struct iovec *iov __unused, in pci_vt9p_drop() 204 struct iovec iov[VT9P_MAX_IOV]; in pci_vt9p_notify() local 213 n = vq_getchain(vq, iov, VT9P_MAX_IOV, &req); in pci_vt9p_notify() 225 preq->vsr_iov = iov; in pci_vt9p_notify() 231 "len=%zu\r\n", i, iov[i].iov_base, in pci_vt9p_notify() 232 iov[i].iov_len)); in pci_vt9p_notify() 235 l9p_connection_recv(sc->vsc_conn, iov, preq->vsr_respidx, preq); in pci_vt9p_notify()
|
H A D | pci_virtio_block.c | 290 struct iovec iov[BLOCKIF_IOV_MAX + 2]; in pci_vtblk_proc() local 293 n = vq_getchain(vq, iov, BLOCKIF_IOV_MAX + 2, &req); in pci_vtblk_proc() 308 vbh = (struct virtio_blk_hdr *)iov[0].iov_base; in pci_vtblk_proc() 312 io->io_status = (uint8_t *)iov[--n].iov_base; in pci_vtblk_proc() 314 assert(iov[n].iov_len == 1); in pci_vtblk_proc() 334 iolen += iov[i].iov_len; in pci_vtblk_proc() 355 if (iov[1].iov_len != sizeof (*discard)) { in pci_vtblk_proc() 362 iov[1].iov_base; in pci_vtblk_proc() 396 memset(iov[1].iov_base, 0, iov[1].iov_len); in pci_vtblk_proc() 397 strncpy(iov[1].iov_base, sc->vbsc_ident, in pci_vtblk_proc() [all …]
|
H A D | pci_virtio_console.c | 436 struct iovec iov; in pci_vtcon_sock_rx() local 461 n = vq_getchain(vq, &iov, 1, &req); in pci_vtcon_sock_rx() 463 len = readv(sock->vss_conn_fd, &iov, n); in pci_vtcon_sock_rx() 487 struct iovec *iov, int niov) in pci_vtcon_sock_tx() argument 503 iov[i].iov_len); in pci_vtcon_sock_tx() 517 struct iovec *iov, int niov) in pci_vtcon_control_tx() argument 598 struct iovec iov; in pci_vtcon_control_send() local 606 n = vq_getchain(vq, &iov, 1, &req); in pci_vtcon_control_send() 611 memcpy((uint8_t *)iov.iov_base + in pci_vtcon_control_send() 624 struct iovec iov[1]; in pci_vtcon_notify_tx() local [all …]
|
H A D | pci_nvme.c | 2583 struct iovec *iov = req->io_req.br_iov; in pci_nvme_dealloc_sm() local 2586 iov += req->prev_gpaddr; in pci_nvme_dealloc_sm() 2589 req->io_req.br_offset = (off_t)iov->iov_base; in pci_nvme_dealloc_sm() 2590 req->io_req.br_resid = iov->iov_len; in pci_nvme_dealloc_sm() 2692 struct iovec *iov = req->io_req.br_iov; in nvme_opc_dataset_mgmt() local 2705 iov[dr].iov_base = (void *)offset; in nvme_opc_dataset_mgmt() 2706 iov[dr].iov_len = bytes; in nvme_opc_dataset_mgmt()
|
H A D | pci_e82545.c | 1001 off -= iov->iov_len; in e82545_iov_checksum() 1002 iov++; in e82545_iov_checksum() 1015 iov++; in e82545_iov_checksum() 1085 struct iovec *iov; in e82545_transmit() local 1102 iov = &iovb[2]; in e82545_transmit() 1328 iov->iov_base = (uint8_t *)iov->iov_base + now; in e82545_transmit() 1342 iov->iov_len -= now; in e82545_transmit() 1344 iov++; in e82545_transmit() 1348 iov--; in e82545_transmit() 1351 iov->iov_base = hdr; in e82545_transmit() [all …]
|
H A D | net_backends.c | 377 iov, iovcnt, 0); in tap_recv() 801 count_iov(iov, iovcnt))); in netmap_send() 858 iov_frag_buf = iov->iov_base; in netmap_recv() 859 iov_frag_size = iov->iov_len; in netmap_recv() 888 iov++; in netmap_recv() 896 iov_frag_buf = iov->iov_base; in netmap_recv() 1105 len = iov[0].iov_len; in be_dlpi_send() 1182 iov, iovcnt, 0); in be_dlpi_recv() 1190 len = iov[0].iov_len; in be_dlpi_recv() 1497 struct iovec iov; in netbe_rx_discard() local [all …]
|
H A D | net_backends.h | 46 ssize_t netbe_send(net_backend_t *be, const struct iovec *iov, int iovcnt); 48 ssize_t netbe_recv(net_backend_t *be, const struct iovec *iov, int iovcnt);
|
H A D | iov.c | 72 count_iov(const struct iovec *iov, int niov) in count_iov() argument 78 total += iov[i].iov_len; in count_iov() 90 size_t toseek = MIN(length - done, iov[i].iov_len); in truncate_iov() 93 if (toseek <= iov[i].iov_len) { in truncate_iov() 94 iov[i].iov_len = toseek; in truncate_iov() 107 total = count_iov(iov, niov); in iov_to_buf() 113 memcpy((uint8_t *)*buf + ptr, iov[i].iov_base, iov[i].iov_len); in iov_to_buf() 114 ptr += iov[i].iov_len; in iov_to_buf() 136 seek_iov(iov, niov, diov, &ndiov, seek); in buf_to_iov() 137 iov = diov; in buf_to_iov() [all …]
|