Lines Matching +defs:val +defs:se
292 static inline u64 calc_delta_fair(u64 delta, struct sched_entity *se) in calc_delta_fair()
309 #define for_each_sched_entity(se) \ argument
412 is_same_group(struct sched_entity *se, struct sched_entity *pse) in is_same_group()
420 static inline struct sched_entity *parent_entity(const struct sched_entity *se) in parent_entity()
426 find_matching_se(struct sched_entity **se, struct sched_entity **pse) in find_matching_se()
467 static int se_is_idle(struct sched_entity *se) in se_is_idle()
476 #define for_each_sched_entity(se) \ argument
495 static inline struct sched_entity *parent_entity(struct sched_entity *se) in parent_entity()
501 find_matching_se(struct sched_entity **se, struct sched_entity **pse) in find_matching_se()
515 static int se_is_idle(struct sched_entity *se) in se_is_idle()
557 static inline s64 entity_key(struct cfs_rq *cfs_rq, struct sched_entity *se) in entity_key()
624 avg_vruntime_add(struct cfs_rq *cfs_rq, struct sched_entity *se) in avg_vruntime_add()
634 avg_vruntime_sub(struct cfs_rq *cfs_rq, struct sched_entity *se) in avg_vruntime_sub()
695 static void update_entity_lag(struct cfs_rq *cfs_rq, struct sched_entity *se) in update_entity_lag()
740 int entity_eligible(struct cfs_rq *cfs_rq, struct sched_entity *se) in entity_eligible()
761 struct sched_entity *se = __pick_root_entity(cfs_rq); in update_min_vruntime() local
805 static inline void __min_vruntime_update(struct sched_entity *se, struct rb_node *node) in __min_vruntime_update()
814 static inline void __min_slice_update(struct sched_entity *se, struct rb_node *node) in __min_slice_update()
826 static inline bool min_vruntime_update(struct sched_entity *se, bool exit) in min_vruntime_update()
850 static void __enqueue_entity(struct cfs_rq *cfs_rq, struct sched_entity *se) in __enqueue_entity()
859 static void __dequeue_entity(struct cfs_rq *cfs_rq, struct sched_entity *se) in __dequeue_entity()
890 static inline void set_protect_slice(struct sched_entity *se) in set_protect_slice()
895 static inline bool protect_slice(struct sched_entity *se) in protect_slice()
900 static inline void cancel_protect_slice(struct sched_entity *se) in cancel_protect_slice()
928 struct sched_entity *se = __pick_first_entity(cfs_rq); in pick_eevdf() local
1021 static bool update_deadline(struct cfs_rq *cfs_rq, struct sched_entity *se) in update_deadline()
1053 void init_entity_runnable_average(struct sched_entity *se) in init_entity_runnable_average()
1100 struct sched_entity *se = &p->se; in post_init_entity_util_avg() local
1137 void init_entity_runnable_average(struct sched_entity *se) in init_entity_runnable_average()
1190 struct sched_entity *pse, struct sched_entity *se) in do_preempt_short()
1282 update_stats_wait_start_fair(struct cfs_rq *cfs_rq, struct sched_entity *se) in update_stats_wait_start_fair()
1299 update_stats_wait_end_fair(struct cfs_rq *cfs_rq, struct sched_entity *se) in update_stats_wait_end_fair()
1325 update_stats_enqueue_sleeper_fair(struct cfs_rq *cfs_rq, struct sched_entity *se) in update_stats_enqueue_sleeper_fair()
1345 update_stats_enqueue_fair(struct cfs_rq *cfs_rq, struct sched_entity *se, int flags) in update_stats_enqueue_fair()
1362 update_stats_dequeue_fair(struct cfs_rq *cfs_rq, struct sched_entity *se, int flags) in update_stats_dequeue_fair()
1394 update_stats_curr_start(struct cfs_rq *cfs_rq, struct sched_entity *se) in update_stats_curr_start()
3688 account_entity_enqueue(struct cfs_rq *cfs_rq, struct sched_entity *se) in account_entity_enqueue()
3703 account_entity_dequeue(struct cfs_rq *cfs_rq, struct sched_entity *se) in account_entity_dequeue()
3765 enqueue_load_avg(struct cfs_rq *cfs_rq, struct sched_entity *se) in enqueue_load_avg()
3772 dequeue_load_avg(struct cfs_rq *cfs_rq, struct sched_entity *se) in dequeue_load_avg()
3782 enqueue_load_avg(struct cfs_rq *cfs_rq, struct sched_entity *se) { } in enqueue_load_avg()
3784 dequeue_load_avg(struct cfs_rq *cfs_rq, struct sched_entity *se) { } in dequeue_load_avg()
3789 static void reweight_entity(struct cfs_rq *cfs_rq, struct sched_entity *se, in reweight_entity()
3845 struct sched_entity *se = &p->se; in reweight_task_fair() local
3969 static void update_cfs_group(struct sched_entity *se) in update_cfs_group()
3994 static inline void update_cfs_group(struct sched_entity *se) in update_cfs_group()
4187 void set_task_rq_fair(struct sched_entity *se, in set_task_rq_fair()
4281 update_tg_cfs_util(struct cfs_rq *cfs_rq, struct sched_entity *se, struct cfs_rq *gcfs_rq) in update_tg_cfs_util()
4313 update_tg_cfs_runnable(struct cfs_rq *cfs_rq, struct sched_entity *se, struct cfs_rq *gcfs_rq) in update_tg_cfs_runnable()
4343 update_tg_cfs_load(struct cfs_rq *cfs_rq, struct sched_entity *se, struct cfs_rq *gcfs_rq) in update_tg_cfs_load()
4417 static inline int propagate_entity_load_avg(struct sched_entity *se) in propagate_entity_load_avg()
4448 static inline bool skip_blocked_update(struct sched_entity *se) in skip_blocked_update()
4480 static inline int propagate_entity_load_avg(struct sched_entity *se) in propagate_entity_load_avg()
4490 static inline void migrate_se_pelt_lag(struct sched_entity *se) in migrate_se_pelt_lag()
4569 static void migrate_se_pelt_lag(struct sched_entity *se) {} in migrate_se_pelt_lag()
4659 static void attach_entity_load_avg(struct cfs_rq *cfs_rq, struct sched_entity *se) in attach_entity_load_avg()
4714 static void detach_entity_load_avg(struct cfs_rq *cfs_rq, struct sched_entity *se) in detach_entity_load_avg()
4745 static inline void update_load_avg(struct cfs_rq *cfs_rq, struct sched_entity *se, int flags) in update_load_avg()
4791 static void sync_entity_load_avg(struct sched_entity *se) in sync_entity_load_avg()
4804 static void remove_entity_load_avg(struct sched_entity *se) in remove_entity_load_avg()
5156 static inline void update_load_avg(struct cfs_rq *cfs_rq, struct sched_entity *se, int not_used1) in update_load_avg()
5161 static inline void remove_entity_load_avg(struct sched_entity *se) {} in remove_entity_load_avg()
5164 attach_entity_load_avg(struct cfs_rq *cfs_rq, struct sched_entity *se) {} in attach_entity_load_avg()
5166 detach_entity_load_avg(struct cfs_rq *cfs_rq, struct sched_entity *se) {} in detach_entity_load_avg()
5188 struct sched_entity *se = &p->se; in __setparam_fair() local
5203 place_entity(struct cfs_rq *cfs_rq, struct sched_entity *se, int flags) in place_entity()
5317 enqueue_entity(struct cfs_rq *cfs_rq, struct sched_entity *se, int flags) in enqueue_entity()
5384 static void __clear_buddies_next(struct sched_entity *se) in __clear_buddies_next()
5395 static void clear_buddies(struct cfs_rq *cfs_rq, struct sched_entity *se) in clear_buddies()
5403 static void set_delayed(struct sched_entity *se) in set_delayed()
5424 static void clear_delayed(struct sched_entity *se) in clear_delayed()
5446 static inline void finish_delayed_dequeue_entity(struct sched_entity *se) in finish_delayed_dequeue_entity()
5454 dequeue_entity(struct cfs_rq *cfs_rq, struct sched_entity *se, int flags) in dequeue_entity()
5535 set_next_entity(struct cfs_rq *cfs_rq, struct sched_entity *se) in set_next_entity()
5587 struct sched_entity *se; in pick_next_entity() local
5887 struct sched_entity *se; in throttle_cfs_rq() local
5993 struct sched_entity *se; in unthrottle_cfs_rq() local
6790 struct sched_entity *se = &p->se; in hrtick_start_fair() local
6894 requeue_delayed_entity(struct sched_entity *se) in requeue_delayed_entity()
6933 struct sched_entity *se = &p->se; in enqueue_task_fair() local
7067 static int dequeue_entities(struct rq *rq, struct sched_entity *se, int flags) in dequeue_entities()
7565 static inline void set_idle_cores(int cpu, int val) in set_idle_cores()
7672 static inline void set_idle_cores(int cpu, int val) in set_idle_cores()
8655 struct sched_entity *se = &p->se; in migrate_task_rq_fair() local
8681 struct sched_entity *se = &p->se; in task_dead_fair() local
8740 static void set_next_buddy(struct sched_entity *se) in set_next_buddy()
8757 struct sched_entity *se = &donor->se, *pse = &p->se; in check_preempt_wakeup_fair() local
8847 struct sched_entity *se; in pick_task_fair() local
8878 struct sched_entity *se; in pick_next_task_fair() local
8991 struct sched_entity *se = &prev->se; in put_prev_task_fair() local
9007 struct sched_entity *se = &curr->se; in yield_task_fair() local
9034 struct sched_entity *se = &p->se; in yield_to_task_fair() local
9802 struct sched_entity *se; in __update_blocked_fair() local
9842 struct sched_entity *se = cfs_rq->tg->se[cpu_of(rq)]; in update_cfs_rq_h_load() local
12975 __entity_slice_used(struct sched_entity *se, int min_nr_tasks) in __entity_slice_used()
13011 static void se_fi_update(const struct sched_entity *se, unsigned int fi_seq, in se_fi_update()
13029 struct sched_entity *se = &p->se; in task_vruntime_update() local
13111 struct sched_entity *se = &curr->se; in task_tick_fair() local
13167 static void propagate_entity_cfs_rq(struct sched_entity *se) in propagate_entity_cfs_rq()
13193 static void propagate_entity_cfs_rq(struct sched_entity *se) { } in propagate_entity_cfs_rq()
13196 static void detach_entity_cfs_rq(struct sched_entity *se) in detach_entity_cfs_rq()
13218 static void attach_entity_cfs_rq(struct sched_entity *se) in attach_entity_cfs_rq()
13231 struct sched_entity *se = &p->se; in detach_task_cfs_rq() local
13238 struct sched_entity *se = &p->se; in attach_task_cfs_rq() local
13271 struct sched_entity *se = &p->se; in __set_next_task_fair() local
13302 struct sched_entity *se = &p->se; in set_next_task_fair() local
13361 struct sched_entity *se; in alloc_fair_sched_group() local
13402 struct sched_entity *se; in online_fair_sched_group() local
13426 struct sched_entity *se = tg->se[cpu]; in unregister_fair_sched_group() local
13453 struct sched_entity *se, int cpu, in init_tg_cfs_entry()
13505 struct sched_entity *se = tg->se[i]; in __sched_group_set_shares() local
13556 struct sched_entity *se = tg->se[i]; in sched_group_set_idle() local
13605 struct sched_entity *se = &task->se; in get_rr_interval_fair() local