xref: /xnu-10063.121.3/osfmk/vm/vm_compressor.h (revision 2c2f96dc2b9a4408a43d3150ae9c105355ca3daa)
1 /*
2  * Copyright (c) 2000-2016 Apple Inc. All rights reserved.
3  *
4  * @APPLE_OSREFERENCE_LICENSE_HEADER_START@
5  *
6  * This file contains Original Code and/or Modifications of Original Code
7  * as defined in and that are subject to the Apple Public Source License
8  * Version 2.0 (the 'License'). You may not use this file except in
9  * compliance with the License. The rights granted to you under the License
10  * may not be used to create, or enable the creation or redistribution of,
11  * unlawful or unlicensed copies of an Apple operating system, or to
12  * circumvent, violate, or enable the circumvention or violation of, any
13  * terms of an Apple operating system software license agreement.
14  *
15  * Please obtain a copy of the License at
16  * http://www.opensource.apple.com/apsl/ and read it before using this file.
17  *
18  * The Original Code and all software distributed under the License are
19  * distributed on an 'AS IS' basis, WITHOUT WARRANTY OF ANY KIND, EITHER
20  * EXPRESS OR IMPLIED, AND APPLE HEREBY DISCLAIMS ALL SUCH WARRANTIES,
21  * INCLUDING WITHOUT LIMITATION, ANY WARRANTIES OF MERCHANTABILITY,
22  * FITNESS FOR A PARTICULAR PURPOSE, QUIET ENJOYMENT OR NON-INFRINGEMENT.
23  * Please see the License for the specific language governing rights and
24  * limitations under the License.
25  *
26  * @APPLE_OSREFERENCE_LICENSE_HEADER_END@
27  */
28 #ifndef _VM_VM_COMPRESSOR_H_
29 #define _VM_VM_COMPRESSOR_H_
30 
31 #include <vm/vm_compressor_pager.h>
32 #include <vm/vm_kern.h>
33 #include <vm/vm_page.h>
34 #include <vm/vm_protos.h>
35 #include <vm/WKdm_new.h>
36 #include <vm/vm_object.h>
37 #include <vm/vm_map.h>
38 #include <machine/pmap.h>
39 #include <kern/locks.h>
40 
41 #include <sys/kdebug.h>
42 
43 #if defined(__arm64__)
44 #include <arm64/proc_reg.h>
45 #endif
46 
47 #define C_SEG_OFFSET_BITS       16
48 
49 #define C_SEG_MAX_POPULATE_SIZE (4 * PAGE_SIZE)
50 
51 #if defined(__arm64__) && (DEVELOPMENT || DEBUG)
52 
53 #if defined(XNU_PLATFORM_WatchOS)
54 #define VALIDATE_C_SEGMENTS (1)
55 #endif
56 #endif /* defined(__arm64__) && (DEVELOPMENT || DEBUG) */
57 
58 
59 #if DEBUG || COMPRESSOR_INTEGRITY_CHECKS
60 #define ENABLE_SWAP_CHECKS 1
61 #define ENABLE_COMPRESSOR_CHECKS 1
62 #define POPCOUNT_THE_COMPRESSED_DATA (1)
63 #else
64 #define ENABLE_SWAP_CHECKS 0
65 #define ENABLE_COMPRESSOR_CHECKS 0
66 #endif
67 
68 #define CHECKSUM_THE_SWAP               ENABLE_SWAP_CHECKS              /* Debug swap data */
69 #define CHECKSUM_THE_DATA               ENABLE_COMPRESSOR_CHECKS        /* Debug compressor/decompressor data */
70 #define CHECKSUM_THE_COMPRESSED_DATA    ENABLE_COMPRESSOR_CHECKS        /* Debug compressor/decompressor compressed data */
71 
72 #ifndef VALIDATE_C_SEGMENTS
73 #define VALIDATE_C_SEGMENTS             ENABLE_COMPRESSOR_CHECKS        /* Debug compaction */
74 #endif
75 
76 #define RECORD_THE_COMPRESSED_DATA      0
77 
78 /*
79  * The c_slot structure embeds a packed pointer to a c_slot_mapping
80  * (32bits) which we ideally want to span as much VA space as possible
81  * to not limit zalloc in how it sets itself up.
82  */
83 #if !defined(__LP64__)                  /* no packing */
84 #define C_SLOT_PACKED_PTR_BITS          32
85 #define C_SLOT_PACKED_PTR_SHIFT         0
86 #define C_SLOT_PACKED_PTR_BASE          0
87 
88 #define C_SLOT_C_SIZE_BITS              12
89 #define C_SLOT_C_CODEC_BITS             1
90 #define C_SLOT_C_POPCOUNT_BITS          0
91 #define C_SLOT_C_PADDING_BITS           3
92 
93 #elif defined(__arm64__)                /* 32G from the heap start */
94 #define C_SLOT_PACKED_PTR_BITS          33
95 #define C_SLOT_PACKED_PTR_SHIFT         2
96 #define C_SLOT_PACKED_PTR_BASE          ((uintptr_t)KERNEL_PMAP_HEAP_RANGE_START)
97 
98 #define C_SLOT_C_SIZE_BITS              14
99 #define C_SLOT_C_CODEC_BITS             1
100 #define C_SLOT_C_POPCOUNT_BITS          0
101 #define C_SLOT_C_PADDING_BITS           0
102 
103 #elif defined(__x86_64__)               /* 256G from the heap start */
104 #define C_SLOT_PACKED_PTR_BITS          36
105 #define C_SLOT_PACKED_PTR_SHIFT         2
106 #define C_SLOT_PACKED_PTR_BASE          ((uintptr_t)KERNEL_PMAP_HEAP_RANGE_START)
107 
108 #define C_SLOT_C_SIZE_BITS              12
109 #define C_SLOT_C_CODEC_BITS             0 /* not used */
110 #define C_SLOT_C_POPCOUNT_BITS          0
111 #define C_SLOT_C_PADDING_BITS           0
112 
113 #else
114 #error vm_compressor parameters undefined for this architecture
115 #endif
116 
117 /*
118  * Popcounts needs to represent both 0 and full which requires
119  * (8 ^ C_SLOT_C_SIZE_BITS) + 1 values and (C_SLOT_C_SIZE_BITS + 4) bits.
120  *
121  * We us the (2 * (8 ^ C_SLOT_C_SIZE_BITS) - 1) value to mean "unknown".
122  */
123 #define C_SLOT_NO_POPCOUNT              ((16u << C_SLOT_C_SIZE_BITS) - 1)
124 
125 static_assert((C_SEG_OFFSET_BITS + C_SLOT_C_SIZE_BITS +
126     C_SLOT_C_CODEC_BITS + C_SLOT_C_POPCOUNT_BITS +
127     C_SLOT_C_PADDING_BITS + C_SLOT_PACKED_PTR_BITS) % 32 == 0);
128 
129 struct c_slot {
130 	uint64_t        c_offset:C_SEG_OFFSET_BITS __kernel_ptr_semantics;
131 	uint64_t        c_size:C_SLOT_C_SIZE_BITS;
132 #if C_SLOT_C_CODEC_BITS
133 	uint64_t        c_codec:C_SLOT_C_CODEC_BITS;
134 #endif
135 #if C_SLOT_C_POPCOUNT_BITS
136 	/*
137 	 * This value may not agree with c_pop_cdata, as it may be the
138 	 * population count of the uncompressed data.
139 	 *
140 	 * This value must be C_SLOT_NO_POPCOUNT when the compression algorithm
141 	 * cannot provide it.
142 	 */
143 	uint32_t        c_inline_popcount:C_SLOT_C_POPCOUNT_BITS;
144 #endif
145 #if C_SLOT_C_PADDING_BITS
146 	uint64_t        c_padding:C_SLOT_C_PADDING_BITS;
147 #endif
148 	uint64_t        c_packed_ptr:C_SLOT_PACKED_PTR_BITS __kernel_ptr_semantics;
149 
150 	/* debugging fields, typically not present on release kernels */
151 #if CHECKSUM_THE_DATA
152 	unsigned int    c_hash_data;
153 #endif
154 #if CHECKSUM_THE_COMPRESSED_DATA
155 	unsigned int    c_hash_compressed_data;
156 #endif
157 #if POPCOUNT_THE_COMPRESSED_DATA
158 	unsigned int    c_pop_cdata;
159 #endif
160 } __attribute__((packed, aligned(4)));
161 
162 #define C_IS_EMPTY              0
163 #define C_IS_FREE               1
164 #define C_IS_FILLING            2
165 #define C_ON_AGE_Q              3
166 #define C_ON_SWAPOUT_Q          4
167 #define C_ON_SWAPPEDOUT_Q       5
168 #define C_ON_SWAPPEDOUTSPARSE_Q 6
169 #define C_ON_SWAPPEDIN_Q        7
170 #define C_ON_MAJORCOMPACT_Q     8
171 #define C_ON_BAD_Q              9
172 #define C_ON_SWAPIO_Q          10
173 
174 
175 struct c_segment {
176 	lck_mtx_t       c_lock;
177 	queue_chain_t   c_age_list;
178 	queue_chain_t   c_list;
179 
180 #if CONFIG_FREEZE
181 	queue_chain_t   c_task_list_next_cseg;
182 	task_t          c_task_owner;
183 #endif /* CONFIG_FREEZE */
184 
185 #define C_SEG_MAX_LIMIT         (UINT_MAX)       /* this needs to track the size of c_mysegno */
186 	uint32_t        c_mysegno;
187 
188 	uint32_t        c_creation_ts;
189 	uint64_t        c_generation_id;
190 
191 	int32_t         c_bytes_used;
192 	int32_t         c_bytes_unused;
193 	uint32_t        c_slots_used;
194 
195 	uint16_t        c_firstemptyslot;
196 	uint16_t        c_nextslot;
197 	uint32_t        c_nextoffset;
198 	uint32_t        c_populated_offset;
199 
200 	union {
201 		int32_t *c_buffer;
202 		uint64_t c_swap_handle;
203 	} c_store;
204 
205 #if     VALIDATE_C_SEGMENTS
206 	uint32_t        c_was_minor_compacted;
207 	uint32_t        c_was_major_compacted;
208 	uint32_t        c_was_major_donor;
209 #endif
210 #if CHECKSUM_THE_SWAP
211 	unsigned int    cseg_hash;
212 	unsigned int    cseg_swap_size;
213 #endif /* CHECKSUM_THE_SWAP */
214 
215 	thread_t        c_busy_for_thread;
216 	uint32_t        c_agedin_ts;
217 	uint32_t        c_swappedin_ts;
218 	bool            c_swappedin;
219 	/*
220 	 * Do not pull c_swappedin above into the bitfield below.
221 	 * We update it without always taking the segment
222 	 * lock and rely on the segment being busy instead.
223 	 * The bitfield needs the segment lock. So updating
224 	 * this state, if in the bitfield, without the lock
225 	 * will race with the updates to the other fields and
226 	 * result in a mess.
227 	 */
228 	uint32_t        c_busy:1,
229 	    c_busy_swapping:1,
230 	    c_wanted:1,
231 	    c_on_minorcompact_q:1,              /* can also be on the age_q, the majorcompact_q or the swappedin_q */
232 
233 	    c_state:4,                          /* what state is the segment in which dictates which q to find it on */
234 	    c_overage_swap:1,
235 	    c_has_donated_pages:1,
236 #if CONFIG_FREEZE
237 	    c_has_freezer_pages:1,
238 	    c_reserved:21;
239 #else /* CONFIG_FREEZE */
240 	c_reserved:22;
241 #endif /* CONFIG_FREEZE */
242 
243 	int             c_slot_var_array_len;
244 	struct  c_slot  *c_slot_var_array;
245 	struct  c_slot  c_slot_fixed_array[0];
246 };
247 
248 
249 struct  c_slot_mapping {
250 #if !CONFIG_TRACK_UNMODIFIED_ANON_PAGES
251 	uint32_t        s_cseg:22,      /* segment number + 1 */
252 	    s_cindx:10;                 /* index in the segment */
253 #else /* !CONFIG_TRACK_UNMODIFIED_ANON_PAGES */
254 	uint32_t        s_cseg:21,      /* segment number + 1 */
255 	    s_cindx:10,                 /* index in the segment */
256 	    s_uncompressed:1;           /* This bit indicates that the page resides uncompressed in a swapfile.
257 	                                 * This can happen in 2 ways:-
258 	                                 * 1) Page used to be in the compressor, got decompressed, was not
259 	                                 * modified, and so was pushed uncompressed to a different swapfile on disk.
260 	                                 * 2) Page was in its uncompressed form in a swapfile on disk. It got swapped in
261 	                                 * but was not modified. As we are about to reclaim it, we notice that this bit
262 	                                 * is set in its current slot. And so we can safely toss this clean anonymous page
263 	                                 * because its copy exists on disk.
264 	                                 */
265 #endif /* !CONFIG_TRACK_UNMODIFIED_ANON_PAGES */
266 };
267 #define C_SLOT_MAX_INDEX        (1 << 10)
268 
269 typedef struct c_slot_mapping *c_slot_mapping_t;
270 
271 
272 extern  int             c_seg_fixed_array_len;
273 extern  vm_offset_t     c_buffers;
274 extern int64_t c_segment_compressed_bytes;
275 
276 #define C_SEG_BUFFER_ADDRESS(c_segno)   ((c_buffers + ((uint64_t)c_segno * (uint64_t)c_seg_allocsize)))
277 
278 #define C_SEG_SLOT_FROM_INDEX(cseg, index)      (index < c_seg_fixed_array_len ? &(cseg->c_slot_fixed_array[index]) : &(cseg->c_slot_var_array[index - c_seg_fixed_array_len]))
279 
280 #define C_SEG_OFFSET_TO_BYTES(off)      ((off) * (int) sizeof(int32_t))
281 #define C_SEG_BYTES_TO_OFFSET(bytes)    ((bytes) / (int) sizeof(int32_t))
282 
283 #define C_SEG_UNUSED_BYTES(cseg)        (cseg->c_bytes_unused + (C_SEG_OFFSET_TO_BYTES(cseg->c_populated_offset - cseg->c_nextoffset)))
284 
285 #ifndef __PLATFORM_WKDM_ALIGNMENT_MASK__
286 #define C_SEG_OFFSET_ALIGNMENT_MASK     0x3ULL
287 #define C_SEG_OFFSET_ALIGNMENT_BOUNDARY 0x4
288 #else
289 #define C_SEG_OFFSET_ALIGNMENT_MASK     __PLATFORM_WKDM_ALIGNMENT_MASK__
290 #define C_SEG_OFFSET_ALIGNMENT_BOUNDARY __PLATFORM_WKDM_ALIGNMENT_BOUNDARY__
291 #endif
292 
293 #define C_SEG_SHOULD_MINORCOMPACT_NOW(cseg)     ((C_SEG_UNUSED_BYTES(cseg) >= (c_seg_bufsize / 4)) ? 1 : 0)
294 
295 /*
296  * the decsion to force a c_seg to be major compacted is based on 2 criteria
297  * 1) is the c_seg buffer almost empty (i.e. we have a chance to merge it with another c_seg)
298  * 2) are there at least a minimum number of slots unoccupied so that we have a chance
299  *    of combining this c_seg with another one.
300  */
301 #define C_SEG_SHOULD_MAJORCOMPACT_NOW(cseg)                                                                                     \
302 	((((cseg->c_bytes_unused + (c_seg_bufsize - C_SEG_OFFSET_TO_BYTES(c_seg->c_nextoffset))) >= (c_seg_bufsize / 8)) &&     \
303 	  ((C_SLOT_MAX_INDEX - cseg->c_slots_used) > (c_seg_bufsize / PAGE_SIZE))) \
304 	? 1 : 0)
305 
306 #define C_SEG_ONDISK_IS_SPARSE(cseg)    ((cseg->c_bytes_used < cseg->c_bytes_unused) ? 1 : 0)
307 #define C_SEG_IS_ONDISK(cseg)           ((cseg->c_state == C_ON_SWAPPEDOUT_Q || cseg->c_state == C_ON_SWAPPEDOUTSPARSE_Q))
308 #define C_SEG_IS_ON_DISK_OR_SOQ(cseg)   ((cseg->c_state == C_ON_SWAPPEDOUT_Q || \
309 	                                  cseg->c_state == C_ON_SWAPPEDOUTSPARSE_Q || \
310 	                                  cseg->c_state == C_ON_SWAPOUT_Q || \
311 	                                  cseg->c_state == C_ON_SWAPIO_Q))
312 
313 
314 #define C_SEG_WAKEUP_DONE(cseg)                         \
315 	MACRO_BEGIN                                     \
316 	assert((cseg)->c_busy);                         \
317 	(cseg)->c_busy = 0;                             \
318 	assert((cseg)->c_busy_for_thread != NULL);      \
319 	(cseg)->c_busy_for_thread = NULL;               \
320 	if ((cseg)->c_wanted) {                         \
321 	        (cseg)->c_wanted = 0;                   \
322 	        thread_wakeup((event_t) (cseg));        \
323 	}                                               \
324 	MACRO_END
325 
326 #define C_SEG_BUSY(cseg)                                \
327 	MACRO_BEGIN                                     \
328 	assert((cseg)->c_busy == 0);                    \
329 	(cseg)->c_busy = 1;                             \
330 	assert((cseg)->c_busy_for_thread == NULL);      \
331 	(cseg)->c_busy_for_thread = current_thread();   \
332 	MACRO_END
333 
334 
335 extern vm_map_t compressor_map;
336 
337 #if DEVELOPMENT || DEBUG
338 extern boolean_t write_protect_c_segs;
339 extern int vm_compressor_test_seg_wp;
340 
341 #define C_SEG_MAKE_WRITEABLE(cseg)                      \
342 	MACRO_BEGIN                                     \
343 	if (write_protect_c_segs) {                     \
344 	        vm_map_protect(compressor_map,                  \
345 	                       (vm_map_offset_t)cseg->c_store.c_buffer,         \
346 	                       (vm_map_offset_t)&cseg->c_store.c_buffer[C_SEG_BYTES_TO_OFFSET(c_seg_allocsize)],\
347 	                       VM_PROT_READ | VM_PROT_WRITE,    \
348 	                       0);                              \
349 	}                               \
350 	MACRO_END
351 
352 #define C_SEG_WRITE_PROTECT(cseg)                       \
353 	MACRO_BEGIN                                     \
354 	if (write_protect_c_segs) {                     \
355 	        vm_map_protect(compressor_map,                  \
356 	                       (vm_map_offset_t)cseg->c_store.c_buffer,         \
357 	                       (vm_map_offset_t)&cseg->c_store.c_buffer[C_SEG_BYTES_TO_OFFSET(c_seg_allocsize)],\
358 	                       VM_PROT_READ,                    \
359 	                       0);                              \
360 	}                                                       \
361 	if (vm_compressor_test_seg_wp) {                                \
362 	        volatile uint32_t vmtstmp = *(volatile uint32_t *)cseg->c_store.c_buffer; \
363 	        *(volatile uint32_t *)cseg->c_store.c_buffer = 0xDEADABCD; \
364 	        (void) vmtstmp;                                         \
365 	}                                                               \
366 	MACRO_END
367 #endif
368 
369 typedef struct c_segment *c_segment_t;
370 typedef struct c_slot   *c_slot_t;
371 
372 uint64_t vm_compressor_total_compressions(void);
373 void vm_wake_compactor_swapper(void);
374 void vm_run_compactor(void);
375 void vm_thrashing_jetsam_done(void);
376 void vm_consider_waking_compactor_swapper(void);
377 void vm_consider_swapping(void);
378 void vm_compressor_flush(void);
379 void c_seg_free(c_segment_t);
380 bool vm_compressor_is_thrashing(void);
381 bool vm_compressor_needs_to_swap(bool wake_memorystatus_thread);
382 void c_seg_free_locked(c_segment_t);
383 void c_seg_insert_into_age_q(c_segment_t);
384 void c_seg_need_delayed_compaction(c_segment_t, boolean_t);
385 void c_seg_update_task_owner(c_segment_t, task_t);
386 
387 void vm_decompressor_lock(void);
388 void vm_decompressor_unlock(void);
389 
390 void vm_compressor_delay_trim(void);
391 void vm_compressor_do_warmup(void);
392 void vm_compressor_record_warmup_start(void);
393 void vm_compressor_record_warmup_end(void);
394 
395 int                     vm_wants_task_throttled(task_t);
396 
397 extern void             vm_compaction_swapper_do_init(void);
398 extern void             vm_compressor_swap_init(void);
399 extern lck_rw_t         c_master_lock;
400 
401 #if ENCRYPTED_SWAP
402 extern void             vm_swap_decrypt(c_segment_t);
403 #endif /* ENCRYPTED_SWAP */
404 
405 extern int              vm_swap_low_on_space(void);
406 extern int              vm_swap_out_of_space(void);
407 extern kern_return_t    vm_swap_get(c_segment_t, uint64_t, uint64_t);
408 extern void             vm_swap_free(uint64_t);
409 extern void             vm_swap_consider_defragmenting(int);
410 
411 extern void             c_seg_swapin_requeue(c_segment_t, boolean_t, boolean_t, boolean_t);
412 extern int              c_seg_swapin(c_segment_t, boolean_t, boolean_t);
413 extern void             c_seg_wait_on_busy(c_segment_t);
414 extern void             c_seg_trim_tail(c_segment_t);
415 extern void             c_seg_switch_state(c_segment_t, int, boolean_t);
416 
417 extern boolean_t        fastwake_recording_in_progress;
418 extern int              compaction_swapper_inited;
419 extern int              compaction_swapper_running;
420 extern uint64_t         vm_swap_put_failures;
421 
422 extern int              c_overage_swapped_count;
423 extern int              c_overage_swapped_limit;
424 
425 extern queue_head_t     c_minor_list_head;
426 extern queue_head_t     c_age_list_head;
427 extern queue_head_t     c_major_list_head;
428 extern queue_head_t     c_early_swapout_list_head;
429 extern queue_head_t     c_regular_swapout_list_head;
430 extern queue_head_t     c_late_swapout_list_head;
431 extern queue_head_t     c_swappedout_list_head;
432 extern queue_head_t     c_swappedout_sparse_list_head;
433 
434 extern uint32_t         c_age_count;
435 extern uint32_t         c_early_swapout_count, c_regular_swapout_count, c_late_swapout_count;
436 extern uint32_t         c_swappedout_count;
437 extern uint32_t         c_swappedout_sparse_count;
438 
439 extern int64_t          compressor_bytes_used;
440 extern uint64_t         first_c_segment_to_warm_generation_id;
441 extern uint64_t         last_c_segment_to_warm_generation_id;
442 extern boolean_t        hibernate_flushing;
443 extern boolean_t        hibernate_no_swapspace;
444 extern boolean_t        hibernate_in_progress_with_pinned_swap;
445 extern boolean_t        hibernate_flush_timed_out;
446 extern uint32_t         swapout_target_age;
447 
448 extern void c_seg_insert_into_q(queue_head_t *, c_segment_t);
449 
450 extern uint32_t vm_compressor_minorcompact_threshold_divisor;
451 extern uint32_t vm_compressor_majorcompact_threshold_divisor;
452 extern uint32_t vm_compressor_unthrottle_threshold_divisor;
453 extern uint32_t vm_compressor_catchup_threshold_divisor;
454 
455 extern uint32_t vm_compressor_minorcompact_threshold_divisor_overridden;
456 extern uint32_t vm_compressor_majorcompact_threshold_divisor_overridden;
457 extern uint32_t vm_compressor_unthrottle_threshold_divisor_overridden;
458 extern uint32_t vm_compressor_catchup_threshold_divisor_overridden;
459 
460 extern uint64_t vm_compressor_compute_elapsed_msecs(clock_sec_t, clock_nsec_t, clock_sec_t, clock_nsec_t);
461 
462 extern void kdp_compressor_busy_find_owner(event64_t wait_event, thread_waitinfo_t *waitinfo);
463 
464 #define PAGE_REPLACEMENT_DISALLOWED(enable)     (enable == TRUE ? lck_rw_lock_shared(&c_master_lock) : lck_rw_done(&c_master_lock))
465 #define PAGE_REPLACEMENT_ALLOWED(enable)        (enable == TRUE ? lck_rw_lock_exclusive(&c_master_lock) : lck_rw_done(&c_master_lock))
466 
467 
468 #define AVAILABLE_NON_COMPRESSED_MEMORY         (vm_page_active_count + vm_page_inactive_count + vm_page_free_count + vm_page_speculative_count)
469 #define AVAILABLE_MEMORY                        (AVAILABLE_NON_COMPRESSED_MEMORY + VM_PAGE_COMPRESSOR_COUNT)
470 
471 /*
472  * TODO, there may be a minor optimisation opportunity to replace these divisions
473  * with multiplies and shifts
474  *
475  * By multiplying by 10, the divisors can have more precision w/o resorting to floating point... a divisor specified as 25 is in reality a divide by 2.5
476  * By multiplying by 9, you get a number ~11% smaller which allows us to have another limit point derived from the same base
477  * By multiplying by 11, you get a number ~10% bigger which allows us to generate a reset limit derived from the same base which is useful for hysteresis
478  */
479 
480 #define VM_PAGE_COMPRESSOR_COMPACT_THRESHOLD            (((AVAILABLE_MEMORY) * 10) / (vm_compressor_minorcompact_threshold_divisor ? vm_compressor_minorcompact_threshold_divisor : 10))
481 #define VM_PAGE_COMPRESSOR_SWAP_THRESHOLD               (((AVAILABLE_MEMORY) * 10) / (vm_compressor_majorcompact_threshold_divisor ? vm_compressor_majorcompact_threshold_divisor : 10))
482 
483 #define VM_PAGE_COMPRESSOR_SWAP_UNTHROTTLE_THRESHOLD    (((AVAILABLE_MEMORY) * 10) / (vm_compressor_unthrottle_threshold_divisor ? vm_compressor_unthrottle_threshold_divisor : 10))
484 #define VM_PAGE_COMPRESSOR_SWAP_RETHROTTLE_THRESHOLD    (((AVAILABLE_MEMORY) * 11) / (vm_compressor_unthrottle_threshold_divisor ? vm_compressor_unthrottle_threshold_divisor : 11))
485 
486 #define VM_PAGE_COMPRESSOR_SWAP_HAS_CAUGHTUP_THRESHOLD  (((AVAILABLE_MEMORY) * 11) / (vm_compressor_catchup_threshold_divisor ? vm_compressor_catchup_threshold_divisor : 11))
487 #define VM_PAGE_COMPRESSOR_SWAP_CATCHUP_THRESHOLD       (((AVAILABLE_MEMORY) * 10) / (vm_compressor_catchup_threshold_divisor ? vm_compressor_catchup_threshold_divisor : 10))
488 #define VM_PAGE_COMPRESSOR_HARD_THROTTLE_THRESHOLD      (((AVAILABLE_MEMORY) * 9) / (vm_compressor_catchup_threshold_divisor ? vm_compressor_catchup_threshold_divisor : 9))
489 
490 #if !XNU_TARGET_OS_OSX
491 #define AVAILABLE_NON_COMPRESSED_MIN                    20000
492 #define COMPRESSOR_NEEDS_TO_SWAP()              (((AVAILABLE_NON_COMPRESSED_MEMORY < VM_PAGE_COMPRESSOR_SWAP_THRESHOLD) || \
493 	                                          (AVAILABLE_NON_COMPRESSED_MEMORY < AVAILABLE_NON_COMPRESSED_MIN)) ? 1 : 0)
494 #else /* !XNU_TARGET_OS_OSX */
495 #define COMPRESSOR_NEEDS_TO_SWAP()              ((AVAILABLE_NON_COMPRESSED_MEMORY < VM_PAGE_COMPRESSOR_SWAP_THRESHOLD) ? 1 : 0)
496 #endif /* !XNU_TARGET_OS_OSX */
497 
498 #define HARD_THROTTLE_LIMIT_REACHED()           ((AVAILABLE_NON_COMPRESSED_MEMORY < VM_PAGE_COMPRESSOR_HARD_THROTTLE_THRESHOLD) ? 1 : 0)
499 #define SWAPPER_NEEDS_TO_UNTHROTTLE()           ((AVAILABLE_NON_COMPRESSED_MEMORY < VM_PAGE_COMPRESSOR_SWAP_UNTHROTTLE_THRESHOLD) ? 1 : 0)
500 #define SWAPPER_NEEDS_TO_RETHROTTLE()           ((AVAILABLE_NON_COMPRESSED_MEMORY > VM_PAGE_COMPRESSOR_SWAP_RETHROTTLE_THRESHOLD) ? 1 : 0)
501 #define SWAPPER_NEEDS_TO_CATCHUP()              ((AVAILABLE_NON_COMPRESSED_MEMORY < VM_PAGE_COMPRESSOR_SWAP_CATCHUP_THRESHOLD) ? 1 : 0)
502 #define SWAPPER_HAS_CAUGHTUP()                  ((AVAILABLE_NON_COMPRESSED_MEMORY > VM_PAGE_COMPRESSOR_SWAP_HAS_CAUGHTUP_THRESHOLD) ? 1 : 0)
503 #define COMPRESSOR_NEEDS_TO_MINOR_COMPACT()     ((AVAILABLE_NON_COMPRESSED_MEMORY < VM_PAGE_COMPRESSOR_COMPACT_THRESHOLD) ? 1 : 0)
504 
505 
506 #if !XNU_TARGET_OS_OSX
507 #define COMPRESSOR_FREE_RESERVED_LIMIT          28
508 #else /* !XNU_TARGET_OS_OSX */
509 #define COMPRESSOR_FREE_RESERVED_LIMIT          128
510 #endif /* !XNU_TARGET_OS_OSX */
511 
512 uint32_t vm_compressor_get_encode_scratch_size(void) __pure2;
513 uint32_t vm_compressor_get_decode_scratch_size(void) __pure2;
514 
515 #define COMPRESSOR_SCRATCH_BUF_SIZE vm_compressor_get_encode_scratch_size()
516 
517 #if RECORD_THE_COMPRESSED_DATA
518 extern void      c_compressed_record_init(void);
519 extern void      c_compressed_record_write(char *, int);
520 #endif
521 
522 extern lck_mtx_t c_list_lock_storage;
523 #define          c_list_lock (&c_list_lock_storage)
524 
525 #if DEVELOPMENT || DEBUG
526 extern uint32_t vm_ktrace_enabled;
527 
528 #define VMKDBG(x, ...)          \
529 MACRO_BEGIN                     \
530 if (vm_ktrace_enabled) {        \
531 	KDBG(x, ## __VA_ARGS__);\
532 }                               \
533 MACRO_END
534 
535 #if DEVELOPMENT || DEBUG
536 extern bool compressor_running_perf_test;
537 extern uint64_t compressor_perf_test_pages_processed;
538 #endif /* DEVELOPMENT || DEBUG */
539 #endif
540 
541 #endif /* _VM_VM_COMPRESSOR_H_ */
542