• Home
  • Raw
  • Download

Lines Matching refs:entry

120 	struct sched_entry *entry, *n;  in taprio_free_sched_cb()  local
122 list_for_each_entry_safe(entry, n, &sched->entries, list) { in taprio_free_sched_cb()
123 list_del(&entry->list); in taprio_free_sched_cb()
124 kfree(entry); in taprio_free_sched_cb()
158 struct sched_entry *entry, in get_interval_end_time() argument
165 intv_end = ktime_add_ns(intv_start, entry->interval); in get_interval_end_time()
198 struct sched_entry *entry = NULL, *entry_found = NULL; in find_entry_to_transmit() local
219 list_for_each_entry(entry, &sched->entries, list) { in find_entry_to_transmit()
221 curr_intv_end = get_interval_end_time(sched, admin, entry, in find_entry_to_transmit()
227 if (!(entry->gate_mask & BIT(tc)) || in find_entry_to_transmit()
228 packet_transmit_time > entry->interval) in find_entry_to_transmit()
231 txtime = entry->next_txtime; in find_entry_to_transmit()
238 entry_found = entry; in find_entry_to_transmit()
247 entry_found = entry; in find_entry_to_transmit()
254 entry_found = entry; in find_entry_to_transmit()
269 struct sched_entry *entry; in is_valid_interval() local
275 entry = find_entry_to_transmit(skb, sch, sched, admin, skb->tstamp, in is_valid_interval()
279 return entry; in is_valid_interval()
352 struct sched_entry *entry; in get_packet_txtime() local
379 entry = find_entry_to_transmit(skb, sch, sched, admin, in get_packet_txtime()
383 if (!entry) { in get_packet_txtime()
388 txtime = entry->next_txtime; in get_packet_txtime()
406 entry->next_txtime = ktime_add(interval_start, sched->cycle_time); in get_packet_txtime()
409 entry->next_txtime = transmit_end_time; in get_packet_txtime()
498 struct sched_entry *entry; in taprio_peek_soft() local
504 entry = rcu_dereference(q->current_entry); in taprio_peek_soft()
505 gate_mask = entry ? entry->gate_mask : TAPRIO_ALL_GATES_OPEN; in taprio_peek_soft()
552 static void taprio_set_budget(struct taprio_sched *q, struct sched_entry *entry) in taprio_set_budget() argument
554 atomic_set(&entry->budget, in taprio_set_budget()
555 div64_u64((u64)entry->interval * 1000, in taprio_set_budget()
564 struct sched_entry *entry; in taprio_dequeue_soft() local
569 entry = rcu_dereference(q->current_entry); in taprio_dequeue_soft()
575 gate_mask = entry ? entry->gate_mask : TAPRIO_ALL_GATES_OPEN; in taprio_dequeue_soft()
617 ktime_after(guard, entry->close_time)) { in taprio_dequeue_soft()
624 atomic_sub_return(len, &entry->budget) < 0) { in taprio_dequeue_soft()
662 const struct sched_entry *entry) in should_restart_cycle() argument
664 if (list_is_last(&entry->list, &oper->entries)) in should_restart_cycle()
667 if (ktime_compare(entry->close_time, oper->cycle_close_time) == 0) in should_restart_cycle()
712 struct sched_entry *entry, *next; in advance_sched() local
717 entry = rcu_dereference_protected(q->current_entry, in advance_sched()
733 if (unlikely(!entry || entry->close_time == oper->base_time)) { in advance_sched()
740 if (should_restart_cycle(oper, entry)) { in advance_sched()
746 next = list_next_entry(entry, list); in advance_sched()
749 close_time = ktime_add_ns(entry->close_time, next->interval); in advance_sched()
804 struct sched_entry *entry, in fill_sched_entry() argument
811 entry->command = nla_get_u8( in fill_sched_entry()
815 entry->gate_mask = nla_get_u32( in fill_sched_entry()
830 entry->interval = interval; in fill_sched_entry()
836 struct sched_entry *entry, int index, in parse_sched_entry() argument
849 entry->index = index; in parse_sched_entry()
851 return fill_sched_entry(q, tb, entry, extack); in parse_sched_entry()
866 struct sched_entry *entry; in parse_sched_list() local
873 entry = kzalloc(sizeof(*entry), GFP_KERNEL); in parse_sched_list()
874 if (!entry) { in parse_sched_list()
879 err = parse_sched_entry(q, n, entry, i, extack); in parse_sched_list()
881 kfree(entry); in parse_sched_list()
885 list_add_tail(&entry->list, &sched->entries); in parse_sched_list()
921 struct sched_entry *entry; in parse_taprio_schedule() local
924 list_for_each_entry(entry, &new->entries, list) in parse_taprio_schedule()
925 cycle = ktime_add_ns(cycle, entry->interval); in parse_taprio_schedule()
1140 struct sched_entry *entry; in setup_txtime() local
1143 list_for_each_entry(entry, &sched->entries, list) { in setup_txtime()
1144 entry->next_txtime = ktime_add_ns(base, interval); in setup_txtime()
1145 interval += entry->interval; in setup_txtime()
1242 struct sched_entry *entry; in taprio_sched_to_offload() local
1249 list_for_each_entry(entry, &sched->entries, list) { in taprio_sched_to_offload()
1252 e->command = entry->command; in taprio_sched_to_offload()
1253 e->interval = entry->interval; in taprio_sched_to_offload()
1254 e->gate_mask = tc_map_to_queue_mask(dev, entry->gate_mask); in taprio_sched_to_offload()
1827 const struct sched_entry *entry) in dump_entry() argument
1835 if (nla_put_u32(msg, TCA_TAPRIO_SCHED_ENTRY_INDEX, entry->index)) in dump_entry()
1838 if (nla_put_u8(msg, TCA_TAPRIO_SCHED_ENTRY_CMD, entry->command)) in dump_entry()
1842 entry->gate_mask)) in dump_entry()
1846 entry->interval)) in dump_entry()
1860 struct sched_entry *entry; in dump_schedule() local
1879 list_for_each_entry(entry, &root->entries, list) { in dump_schedule()
1880 if (dump_entry(msg, entry) < 0) in dump_schedule()