Lines Matching refs:ringbuffer

43 	mach_vm_reclaim_ring_t ringbuffer = NULL;  in ringbuffer_init()  local
46 kern_return_t kr = mach_vm_reclaim_ring_allocate(&ringbuffer, len, max_len); in ringbuffer_init()
48 return ringbuffer; in ringbuffer_init()
55 mach_vm_reclaim_ring_t ringbuffer = ringbuffer_init(); variable
56 T_ASSERT_NOTNULL(ringbuffer, "ringbuffer is allocated");
57 T_EXPECT_EQ(os_atomic_load(&ringbuffer->head, relaxed), 0ull, "head is zeroed");
58 T_EXPECT_EQ(os_atomic_load(&ringbuffer->busy, relaxed), 0ull, "busy is zeroed");
59 T_EXPECT_EQ(os_atomic_load(&ringbuffer->tail, relaxed), 0ull, "tail is zeroed");
62 T_ASSERT_EQ((size_t)ringbuffer->len, expected_len, "length is set correctly");
63 for (unsigned i = 0; i < ringbuffer->len; i++) {
64 mach_vm_reclaim_entry_t entry = &ringbuffer->entries[i];
75 mach_vm_reclaim_ring_t ringbuffer; variable
76 kern_return_t kr = mach_vm_reclaim_ring_allocate(&ringbuffer, 1, 1);
81 try_cancel(mach_vm_reclaim_ring_t ringbuffer, mach_vm_reclaim_id_t id, mach_vm_address_t addr, mach… in try_cancel() argument
86 kr = mach_vm_reclaim_try_cancel(ringbuffer, id, addr, size, behavior, &state, &update_accounting); in try_cancel()
89 kern_return_t tmp_kr = mach_vm_reclaim_update_kernel_accounting(ringbuffer); in try_cancel()
99 allocate_and_defer_free(size_t size, mach_vm_reclaim_ring_t ringbuffer, in allocate_and_defer_free() argument
110 …kr = mach_vm_reclaim_try_enter(ringbuffer, *addr, size, behavior, &id, &should_update_kernel_accou… in allocate_and_defer_free()
112 kr = mach_vm_reclaim_update_kernel_accounting(ringbuffer); in allocate_and_defer_free()
119 allocate_and_defer_deallocate(size_t size, mach_vm_reclaim_ring_t ringbuffer, unsigned char val, ma… in allocate_and_defer_deallocate() argument
121 return allocate_and_defer_free(size, ringbuffer, val, VM_RECLAIM_DEALLOCATE, addr); in allocate_and_defer_deallocate()
130 mach_vm_reclaim_ring_t ringbuffer = ringbuffer_init(); variable
132 mach_vm_reclaim_id_t idx = allocate_and_defer_deallocate(kAllocationSize, ringbuffer, 1, &addr);
134 mach_vm_reclaim_ring_flush(ringbuffer, 1);
238 mach_vm_reclaim_ring_t ringbuffer = ringbuffer_init(); in reuse_reclaimed_entry() local
240 …mach_vm_reclaim_id_t idx = allocate_and_defer_free(kAllocationSize, ringbuffer, kValue, behavior, … in reuse_reclaimed_entry()
242 kr = mach_vm_reclaim_ring_flush(ringbuffer, 10); in reuse_reclaimed_entry()
244 bool usable = try_cancel(ringbuffer, idx, addr, kAllocationSize, behavior); in reuse_reclaimed_entry()
299 mach_vm_reclaim_ring_t ringbuffer = ringbuffer_init(); in allocate_and_suspend() local
303 kr = mach_vm_reclaim_ring_capacity(ringbuffer, &capacity); in allocate_and_suspend()
310 …mach_vm_reclaim_id_t idx = allocate_and_defer_deallocate(kAllocationSize, ringbuffer, (unsigned ch… in allocate_and_suspend()
318 …kr = mach_vm_reclaim_try_enter(ringbuffer, addr, kAllocationSize, VM_RECLAIM_DEALLOCATE, &id, &sho… in allocate_and_suspend()
325 kr = mach_vm_deallocate(mach_task_self(), (mach_vm_address_t)ringbuffer, buffer_size); in allocate_and_suspend()
503 mach_vm_reclaim_ring_t ringbuffer = ringbuffer_init(); variable
510 …mach_vm_reclaim_id_t idx = allocate_and_defer_deallocate(kAllocationSize, ringbuffer, (unsigned ch…
524 mach_vm_reclaim_ring_t ringbuffer = ringbuffer_init(); variable
526 mach_vm_reclaim_id_t idx = allocate_and_defer_deallocate(kAllocationSize, ringbuffer, 1, &addr);
529 kr = mach_vm_reclaim_ring_capacity(ringbuffer, &capacity);
534 kr = mach_vm_deallocate(mach_task_self(), (mach_vm_address_t)ringbuffer, buffer_size);
537 mach_vm_reclaim_ring_flush(ringbuffer, 10);
562 mach_vm_reclaim_ring_t ringbuffer = ringbuffer_init(); variable
564 mach_vm_reclaim_id_t idx = allocate_and_defer_deallocate(kAllocationSize, ringbuffer, 1, &addr);
567 …kr = mach_vm_reclaim_try_enter(ringbuffer, addr, kAllocationSize, VM_RECLAIM_DEALLOCATE, &idx, &sh…
571 mach_vm_reclaim_ring_flush(ringbuffer, 2);
721 mach_vm_reclaim_ring_t ringbuffer; variable
722 kern_return_t kr = mach_vm_reclaim_ring_allocate(&ringbuffer, 1, 1);
736 mach_vm_reclaim_ring_t ringbuffer; variable
738 kern_return_t kr = mach_vm_reclaim_ring_allocate(&ringbuffer, 1, 1);
766 mach_vm_reclaim_ring_t ringbuffer; variable
775 kr = mach_vm_reclaim_ring_allocate(&ringbuffer, initial_len, max_len);
780 kr = mach_vm_reclaim_ring_capacity(ringbuffer, &old_capacity);
784 ids[i] = allocate_and_defer_deallocate(vm_page_size, ringbuffer, 'A', &addrs[i]);
787 id_tmp = allocate_and_defer_deallocate(vm_page_size, ringbuffer, 'X', &addr_tmp);
789 uint64_t initial_tail = os_atomic_load(&ringbuffer->tail, relaxed);
793 kr = mach_vm_reclaim_ring_resize(ringbuffer, max_len);
797 T_EXPECT_EQ(os_atomic_load(&ringbuffer->head, relaxed), initial_tail, "head is incremented");
798 T_EXPECT_EQ(os_atomic_load(&ringbuffer->busy, relaxed), initial_tail, "busy is incremented");
799 T_EXPECT_EQ(os_atomic_load(&ringbuffer->tail, relaxed), initial_tail, "tail is preserved");
802 kr = mach_vm_reclaim_ring_capacity(ringbuffer, &new_capacity);
810 kr = mach_vm_reclaim_query_state(ringbuffer, ids[i], VM_RECLAIM_DEALLOCATE, &state);
813 bool usable = try_cancel(ringbuffer, ids[i], addrs[i], vm_page_size, VM_RECLAIM_DEALLOCATE);
819 ids[i] = allocate_and_defer_deallocate(vm_page_size, ringbuffer, 'B', &addrs[i]);
822 id_tmp = allocate_and_defer_deallocate(vm_page_size, ringbuffer, 'X', &addr_tmp);
826 bool usable = try_cancel(ringbuffer, ids[i], addrs[i], vm_page_size, VM_RECLAIM_DEALLOCATE);