Lines Matching refs:scmap

5898 		void    *scmap;  in cluster_push_err()  local
5901 scmap = wbp->cl_scmap; in cluster_push_err()
5908 …retval = sparse_cluster_push(wbp, &scmap, vp, ubc_getsize(vp), PUSH_ALL, flags, callback, callback… in cluster_push_err()
5937 vfs_drt_control(&scmap, 0); /* emit stats and free this memory. Dirty pages stay intact. */ in cluster_push_err()
5938 scmap = NULL; in cluster_push_err()
5941 wbp->cl_scmap = scmap; in cluster_push_err()
6464 sparse_cluster_push(struct cl_writebehind *wbp, void **scmap, vnode_t vp, off_t EOF, int push_flag, in sparse_cluster_push() argument
6473 …KERNEL_DEBUG((FSDBG_CODE(DBG_FSRW, 79)) | DBG_FUNC_START, kdebug_vnode(vp), (*scmap), 0, push_flag… in sparse_cluster_push()
6476 vfs_drt_control(scmap, 1); in sparse_cluster_push()
6479 l_scmap = *scmap; in sparse_cluster_push()
6484 if (vfs_drt_get_cluster(scmap, &offset, &length) != KERN_SUCCESS) { in sparse_cluster_push()
6508 if (*scmap != l_scmap) { in sparse_cluster_push()
6514 if (vfs_drt_mark_pages(scmap, offset, length, NULL) != KERN_SUCCESS) { in sparse_cluster_push()
6525 KERNEL_DEBUG((FSDBG_CODE(DBG_FSRW, 79)) | DBG_FUNC_END, kdebug_vnode(vp), (*scmap), error, 0, 0); in sparse_cluster_push()
6535 sparse_cluster_add(struct cl_writebehind *wbp, void **scmap, vnode_t vp, struct cl_extent *cl, off_… in sparse_cluster_add() argument
6544 …KERNEL_DEBUG((FSDBG_CODE(DBG_FSRW, 80)) | DBG_FUNC_START, (*scmap), 0, cl->b_addr, (int)cl->e_addr… in sparse_cluster_add()
6549 while (vfs_drt_mark_pages(scmap, offset, length, &new_dirty) != KERN_SUCCESS) { in sparse_cluster_add()
6556 if (vfs_get_scmap_push_behavior_internal(scmap, &push_flag)) { in sparse_cluster_add()
6560 …error = sparse_cluster_push(wbp, scmap, vp, EOF, push_flag, 0, callback, callback_arg, vm_initiate… in sparse_cluster_add()
6569 KERNEL_DEBUG((FSDBG_CODE(DBG_FSRW, 80)) | DBG_FUNC_END, kdebug_vnode(vp), (*scmap), error, 0, 0); in sparse_cluster_add()