| /* SPDX-License-Identifier: GPL-2.0 */ | 
 |  | 
 | #ifndef BTRFS_MISC_H | 
 | #define BTRFS_MISC_H | 
 |  | 
 | #include <linux/sched.h> | 
 | #include <linux/wait.h> | 
 | #include <asm/div64.h> | 
 |  | 
 | #define in_range(b, first, len) ((b) >= (first) && (b) < (first) + (len)) | 
 |  | 
 | static inline void cond_wake_up(struct wait_queue_head *wq) | 
 | { | 
 | 	/* | 
 | 	 * This implies a full smp_mb barrier, see comments for | 
 | 	 * waitqueue_active why. | 
 | 	 */ | 
 | 	if (wq_has_sleeper(wq)) | 
 | 		wake_up(wq); | 
 | } | 
 |  | 
 | static inline void cond_wake_up_nomb(struct wait_queue_head *wq) | 
 | { | 
 | 	/* | 
 | 	 * Special case for conditional wakeup where the barrier required for | 
 | 	 * waitqueue_active is implied by some of the preceding code. Eg. one | 
 | 	 * of such atomic operations (atomic_dec_and_return, ...), or a | 
 | 	 * unlock/lock sequence, etc. | 
 | 	 */ | 
 | 	if (waitqueue_active(wq)) | 
 | 		wake_up(wq); | 
 | } | 
 |  | 
 | static inline u64 div_factor(u64 num, int factor) | 
 | { | 
 | 	if (factor == 10) | 
 | 		return num; | 
 | 	num *= factor; | 
 | 	return div_u64(num, 10); | 
 | } | 
 |  | 
 | static inline u64 div_factor_fine(u64 num, int factor) | 
 | { | 
 | 	if (factor == 100) | 
 | 		return num; | 
 | 	num *= factor; | 
 | 	return div_u64(num, 100); | 
 | } | 
 |  | 
 | #endif |