/* * Copyright (c) 2024 Intel Corporation * * SPDX-License-Identifier: Apache-2.0 */ #ifndef ZEPHYR_KERNEL_INCLUDE_PRIORITY_Q_H_ #define ZEPHYR_KERNEL_INCLUDE_PRIORITY_Q_H_ #include #include /* Dumb Scheduling */ #if defined(CONFIG_SCHED_SIMPLE) #define _priq_run_init z_priq_simple_init #define _priq_run_add z_priq_simple_add #define _priq_run_remove z_priq_simple_remove #define _priq_run_yield z_priq_simple_yield # if defined(CONFIG_SCHED_CPU_MASK) # define _priq_run_best z_priq_simple_mask_best # else # define _priq_run_best z_priq_simple_best # endif /* CONFIG_SCHED_CPU_MASK */ /* Scalable Scheduling */ #elif defined(CONFIG_SCHED_SCALABLE) #define _priq_run_init z_priq_rb_init #define _priq_run_add z_priq_rb_add #define _priq_run_remove z_priq_rb_remove #define _priq_run_yield z_priq_rb_yield #define _priq_run_best z_priq_rb_best /* Multi Queue Scheduling */ #elif defined(CONFIG_SCHED_MULTIQ) #define _priq_run_init z_priq_mq_init #define _priq_run_add z_priq_mq_add #define _priq_run_remove z_priq_mq_remove #define _priq_run_yield z_priq_mq_yield #define _priq_run_best z_priq_mq_best #endif /* Scalable Wait Queue */ #if defined(CONFIG_WAITQ_SCALABLE) #define _priq_wait_add z_priq_rb_add #define _priq_wait_remove z_priq_rb_remove #define _priq_wait_best z_priq_rb_best /* Dumb Wait Queue */ #elif defined(CONFIG_WAITQ_SIMPLE) #define _priq_wait_add z_priq_simple_add #define _priq_wait_remove z_priq_simple_remove #define _priq_wait_best z_priq_simple_best #endif #if defined(CONFIG_64BIT) #define NBITS 64 #define TRAILING_ZEROS u64_count_trailing_zeros #else #define NBITS 32 #define TRAILING_ZEROS u32_count_trailing_zeros #endif /* CONFIG_64BIT */ static ALWAYS_INLINE void z_priq_simple_init(sys_dlist_t *pq) { sys_dlist_init(pq); } /* * Return value same as e.g. memcmp * > 0 -> thread 1 priority > thread 2 priority * = 0 -> thread 1 priority == thread 2 priority * < 0 -> thread 1 priority < thread 2 priority * Do not rely on the actual value returned aside from the above. * (Again, like memcmp.) */ static ALWAYS_INLINE int32_t z_sched_prio_cmp(struct k_thread *thread_1, struct k_thread *thread_2) { /* `prio` is <32b, so the below cannot overflow. */ int32_t b1 = thread_1->base.prio; int32_t b2 = thread_2->base.prio; if (b1 != b2) { return b2 - b1; } #ifdef CONFIG_SCHED_DEADLINE /* If we assume all deadlines live within the same "half" of * the 32 bit modulus space (this is a documented API rule), * then the latest deadline in the queue minus the earliest is * guaranteed to be (2's complement) non-negative. We can * leverage that to compare the values without having to check * the current time. */ uint32_t d1 = thread_1->base.prio_deadline; uint32_t d2 = thread_2->base.prio_deadline; if (d1 != d2) { /* Sooner deadline means higher effective priority. * Doing the calculation with unsigned types and casting * to signed isn't perfect, but at least reduces this * from UB on overflow to impdef. */ return (int32_t)(d2 - d1); } #endif /* CONFIG_SCHED_DEADLINE */ return 0; } static ALWAYS_INLINE void z_priq_simple_add(sys_dlist_t *pq, struct k_thread *thread) { struct k_thread *t; SYS_DLIST_FOR_EACH_CONTAINER(pq, t, base.qnode_dlist) { if (z_sched_prio_cmp(thread, t) > 0) { sys_dlist_insert(&t->base.qnode_dlist, &thread->base.qnode_dlist); return; } } sys_dlist_append(pq, &thread->base.qnode_dlist); } static ALWAYS_INLINE void z_priq_simple_remove(sys_dlist_t *pq, struct k_thread *thread) { ARG_UNUSED(pq); sys_dlist_remove(&thread->base.qnode_dlist); } static ALWAYS_INLINE void z_priq_simple_yield(sys_dlist_t *pq) { #ifndef CONFIG_SMP sys_dnode_t *n; n = sys_dlist_peek_next_no_check(pq, &_current->base.qnode_dlist); sys_dlist_dequeue(&_current->base.qnode_dlist); struct k_thread *t; /* * As it is possible that the current thread was not at the head of * the run queue, start searching from the present position for where * to re-insert it. */ while (n != NULL) { t = CONTAINER_OF(n, struct k_thread, base.qnode_dlist); if (z_sched_prio_cmp(_current, t) > 0) { sys_dlist_insert(&t->base.qnode_dlist, &_current->base.qnode_dlist); return; } n = sys_dlist_peek_next_no_check(pq, n); } sys_dlist_append(pq, &_current->base.qnode_dlist); #endif } static ALWAYS_INLINE struct k_thread *z_priq_simple_best(sys_dlist_t *pq) { struct k_thread *thread = NULL; sys_dnode_t *n = sys_dlist_peek_head(pq); if (n != NULL) { thread = CONTAINER_OF(n, struct k_thread, base.qnode_dlist); } return thread; } #ifdef CONFIG_SCHED_CPU_MASK static ALWAYS_INLINE struct k_thread *z_priq_simple_mask_best(sys_dlist_t *pq) { /* With masks enabled we need to be prepared to walk the list * looking for one we can run */ struct k_thread *thread; SYS_DLIST_FOR_EACH_CONTAINER(pq, thread, base.qnode_dlist) { if ((thread->base.cpu_mask & BIT(_current_cpu->id)) != 0) { return thread; } } return NULL; } #endif /* CONFIG_SCHED_CPU_MASK */ #if defined(CONFIG_SCHED_SCALABLE) || defined(CONFIG_WAITQ_SCALABLE) static ALWAYS_INLINE void z_priq_rb_init(struct _priq_rb *pq) { bool z_priq_rb_lessthan(struct rbnode *a, struct rbnode *b); *pq = (struct _priq_rb) { .tree = { .lessthan_fn = z_priq_rb_lessthan, } }; } static ALWAYS_INLINE void z_priq_rb_add(struct _priq_rb *pq, struct k_thread *thread) { struct k_thread *t; thread->base.order_key = pq->next_order_key; ++pq->next_order_key; /* Renumber at wraparound. This is tiny code, and in practice * will almost never be hit on real systems. BUT on very * long-running systems where a priq never completely empties * AND that contains very large numbers of threads, it can be * a latency glitch to loop over all the threads like this. */ if (!pq->next_order_key) { RB_FOR_EACH_CONTAINER(&pq->tree, t, base.qnode_rb) { t->base.order_key = pq->next_order_key; ++pq->next_order_key; } } rb_insert(&pq->tree, &thread->base.qnode_rb); } static ALWAYS_INLINE void z_priq_rb_remove(struct _priq_rb *pq, struct k_thread *thread) { rb_remove(&pq->tree, &thread->base.qnode_rb); if (!pq->tree.root) { pq->next_order_key = 0; } } static ALWAYS_INLINE void z_priq_rb_yield(struct _priq_rb *pq) { #ifndef CONFIG_SMP z_priq_rb_remove(pq, _current); z_priq_rb_add(pq, _current); #endif } static ALWAYS_INLINE struct k_thread *z_priq_rb_best(struct _priq_rb *pq) { struct k_thread *thread = NULL; struct rbnode *n = rb_get_min(&pq->tree); if (n != NULL) { thread = CONTAINER_OF(n, struct k_thread, base.qnode_rb); } return thread; } #endif struct prio_info { uint8_t offset_prio; uint8_t idx; uint8_t bit; }; static ALWAYS_INLINE struct prio_info get_prio_info(int8_t old_prio) { struct prio_info ret; ret.offset_prio = old_prio - K_HIGHEST_THREAD_PRIO; ret.idx = ret.offset_prio / NBITS; ret.bit = ret.offset_prio % NBITS; return ret; } static ALWAYS_INLINE unsigned int z_priq_mq_best_queue_index(struct _priq_mq *pq) { unsigned int i = 0; do { if (likely(pq->bitmask[i])) { return i * NBITS + TRAILING_ZEROS(pq->bitmask[i]); } i++; } while (i < PRIQ_BITMAP_SIZE); return K_NUM_THREAD_PRIO - 1; } static ALWAYS_INLINE void z_priq_mq_init(struct _priq_mq *q) { for (size_t i = 0; i < ARRAY_SIZE(q->queues); i++) { sys_dlist_init(&q->queues[i]); } #ifndef CONFIG_SMP q->cached_queue_index = K_NUM_THREAD_PRIO - 1; #endif } static ALWAYS_INLINE void z_priq_mq_add(struct _priq_mq *pq, struct k_thread *thread) { struct prio_info pos = get_prio_info(thread->base.prio); sys_dlist_append(&pq->queues[pos.offset_prio], &thread->base.qnode_dlist); pq->bitmask[pos.idx] |= BIT(pos.bit); #ifndef CONFIG_SMP if (pos.offset_prio < pq->cached_queue_index) { pq->cached_queue_index = pos.offset_prio; } #endif } static ALWAYS_INLINE void z_priq_mq_remove(struct _priq_mq *pq, struct k_thread *thread) { struct prio_info pos = get_prio_info(thread->base.prio); sys_dlist_dequeue(&thread->base.qnode_dlist); if (unlikely(sys_dlist_is_empty(&pq->queues[pos.offset_prio]))) { pq->bitmask[pos.idx] &= ~BIT(pos.bit); #ifndef CONFIG_SMP pq->cached_queue_index = z_priq_mq_best_queue_index(pq); #endif } } static ALWAYS_INLINE void z_priq_mq_yield(struct _priq_mq *pq) { #ifndef CONFIG_SMP struct prio_info pos = get_prio_info(_current->base.prio); sys_dlist_dequeue(&_current->base.qnode_dlist); sys_dlist_append(&pq->queues[pos.offset_prio], &_current->base.qnode_dlist); #endif } static ALWAYS_INLINE struct k_thread *z_priq_mq_best(struct _priq_mq *pq) { #ifdef CONFIG_SMP unsigned int index = z_priq_mq_best_queue_index(pq); #else unsigned int index = pq->cached_queue_index; #endif sys_dnode_t *n = sys_dlist_peek_head(&pq->queues[index]); if (likely(n != NULL)) { return CONTAINER_OF(n, struct k_thread, base.qnode_dlist); } return NULL; } #endif /* ZEPHYR_KERNEL_INCLUDE_PRIORITY_Q_H_ */