• Home
  • Raw
  • Download

Lines Matching refs:cpu

2454 				 int cpu, u16 offset, u16 count)  in remove_xps_queue_cpu()  argument
2460 for (tci = cpu * num_tc; num_tc--; tci++) { in remove_xps_queue_cpu()
4237 int cpu = smp_processor_id(); /* ok because BHs are off */ in __dev_queue_xmit() local
4242 if (READ_ONCE(txq->xmit_lock_owner) != cpu) { in __dev_queue_xmit()
4251 HARD_TX_LOCK(dev, txq, cpu); in __dev_queue_xmit()
4440 rflow->cpu = next_cpu; in set_rps_cpu()
4456 int cpu = -1; in get_rps_cpu() local
4504 tcpu = rflow->cpu; in get_rps_cpu()
4527 cpu = tcpu; in get_rps_cpu()
4537 cpu = tcpu; in get_rps_cpu()
4543 return cpu; in get_rps_cpu()
4566 unsigned int cpu; in rps_may_expire_flow() local
4572 cpu = READ_ONCE(rflow->cpu); in rps_may_expire_flow()
4573 if (rflow->filter == filter_id && cpu < nr_cpu_ids && in rps_may_expire_flow()
4574 ((int)(per_cpu(softnet_data, cpu).input_queue_head - in rps_may_expire_flow()
4662 static int enqueue_to_backlog(struct sk_buff *skb, int cpu, in enqueue_to_backlog() argument
4669 sd = &per_cpu(softnet_data, cpu); in enqueue_to_backlog()
4880 int cpu, rc; in generic_xdp_tx() local
4883 cpu = smp_processor_id(); in generic_xdp_tx()
4884 HARD_TX_LOCK(dev, txq, cpu); in generic_xdp_tx()
4940 int cpu; in netif_rx_internal() local
4945 cpu = get_rps_cpu(skb->dev, skb, &rflow); in netif_rx_internal()
4946 if (cpu < 0) in netif_rx_internal()
4947 cpu = smp_processor_id(); in netif_rx_internal()
4949 ret = enqueue_to_backlog(skb, cpu, &rflow->last_qtail); in netif_rx_internal()
5683 int cpu = get_rps_cpu(skb->dev, skb, &rflow); in netif_receive_skb_internal() local
5685 if (cpu >= 0) { in netif_receive_skb_internal()
5686 ret = enqueue_to_backlog(skb, cpu, &rflow->last_qtail); in netif_receive_skb_internal()
5716 int cpu = get_rps_cpu(skb->dev, skb, &rflow); in netif_receive_skb_list_internal() local
5718 if (cpu >= 0) { in netif_receive_skb_list_internal()
5721 enqueue_to_backlog(skb, cpu, &rflow->last_qtail); in netif_receive_skb_list_internal()
5816 static bool flush_required(int cpu) in flush_required() argument
5819 struct softnet_data *sd = &per_cpu(softnet_data, cpu); in flush_required()
5846 unsigned int cpu; in flush_all_backlogs() local
5857 for_each_online_cpu(cpu) { in flush_all_backlogs()
5858 if (flush_required(cpu)) { in flush_all_backlogs()
5859 queue_work_on(cpu, system_highpri_wq, in flush_all_backlogs()
5860 per_cpu_ptr(&flush_works, cpu)); in flush_all_backlogs()
5861 cpumask_set_cpu(cpu, &flush_cpus); in flush_all_backlogs()
5869 for_each_cpu(cpu, &flush_cpus) in flush_all_backlogs()
5870 flush_work(per_cpu_ptr(&flush_works, cpu)); in flush_all_backlogs()
6426 if (cpu_online(remsd->cpu)) in net_rps_send_ipi()
6427 smp_call_function_single_async(remsd->cpu, &remsd->csd); in net_rps_send_ipi()
10730 int cpu; in dev_fetch_sw_netstats() local
10732 for_each_possible_cpu(cpu) { in dev_fetch_sw_netstats()
10737 stats = per_cpu_ptr(netstats, cpu); in dev_fetch_sw_netstats()
11312 unsigned int cpu; in dev_cpu_dead() local
11316 cpu = smp_processor_id(); in dev_cpu_dead()
11317 sd = &per_cpu(softnet_data, cpu); in dev_cpu_dead()
11690 sd->cpu = i; in net_dev_init()