/arch/sparc/kernel/ |
D | ldc.c | 177 do { if (lp->cfg.debug & LDC_DEBUG_##TYPE) \ 178 printk(KERN_INFO PFX "ID[%lu] " f, lp->id, ## a); \ 181 #define LDC_ABORT(lp) ldc_abort((lp), __func__) argument 210 static unsigned long rx_advance(struct ldc_channel *lp, unsigned long off) in rx_advance() argument 212 return __advance(off, lp->rx_num_entries); in rx_advance() 215 static unsigned long tx_advance(struct ldc_channel *lp, unsigned long off) in tx_advance() argument 217 return __advance(off, lp->tx_num_entries); in tx_advance() 220 static struct ldc_packet *handshake_get_tx_packet(struct ldc_channel *lp, in handshake_get_tx_packet() argument 226 t = tx_advance(lp, lp->tx_tail); in handshake_get_tx_packet() 227 if (t == lp->tx_head) in handshake_get_tx_packet() [all …]
|
D | ds.c | 203 struct ldc_channel *lp; member 244 static int __ds_send(struct ldc_channel *lp, void *data, int len) in __ds_send() argument 250 err = ldc_write(lp, data, len); in __ds_send() 259 static int ds_send(struct ldc_channel *lp, void *data, int len) in ds_send() argument 265 err = __ds_send(lp, data, len); in ds_send() 284 struct ldc_channel *lp = dp->lp; in md_update_data() local 305 ds_send(lp, &pkt, sizeof(pkt)); in md_update_data() 323 struct ldc_channel *lp = dp->lp; in domain_shutdown_data() local 344 ds_send(lp, &pkt, sizeof(pkt)); in domain_shutdown_data() 363 struct ldc_channel *lp = dp->lp; in domain_panic_data() local [all …]
|
D | viohs.c | 24 err = ldc_write(vio->lp, data, len); in vio_ldc_send() 130 ldc_disconnect(vio->lp); in vio_link_state_change() 171 ldc_disconnect(vio->lp); in process_unknown() 739 struct ldc_channel *lp; in vio_ldc_alloc() local 744 lp = ldc_alloc(vio->vdev->channel_id, &cfg, event_arg, vio->name); in vio_ldc_alloc() 745 if (IS_ERR(lp)) in vio_ldc_alloc() 746 return PTR_ERR(lp); in vio_ldc_alloc() 748 vio->lp = lp; in vio_ldc_alloc() 756 ldc_free(vio->lp); in vio_ldc_free() 757 vio->lp = NULL; in vio_ldc_free() [all …]
|
/arch/xtensa/platforms/iss/ |
D | network.c | 85 int (*open)(struct iss_net_private *lp); 86 void (*close)(struct iss_net_private *lp); 87 int (*read)(struct iss_net_private *lp, struct sk_buff **skb); 88 int (*write)(struct iss_net_private *lp, struct sk_buff **skb); 90 int (*poll)(struct iss_net_private *lp); 160 static int tuntap_open(struct iss_net_private *lp) in tuntap_open() argument 163 char *dev_name = lp->tp.info.tuntap.dev_name; in tuntap_open() 170 lp->dev->name, fd, errno); in tuntap_open() 181 lp->dev->name, dev_name, err, errno); in tuntap_open() 186 lp->tp.info.tuntap.fd = fd; in tuntap_open() [all …]
|
/arch/sparc/include/asm/ |
D | ldc.h | 63 void ldc_free(struct ldc_channel *lp); 66 int ldc_bind(struct ldc_channel *lp); 67 void ldc_unbind(struct ldc_channel *lp); 74 int ldc_connect(struct ldc_channel *lp); 75 int ldc_disconnect(struct ldc_channel *lp); 77 int ldc_state(struct ldc_channel *lp); 78 void ldc_set_state(struct ldc_channel *lp, u8 state); 79 int ldc_mode(struct ldc_channel *lp); 80 void __ldc_print(struct ldc_channel *lp, const char *caller); 81 int ldc_rx_reset(struct ldc_channel *lp); [all …]
|
D | spinlock_32.h | 81 register arch_rwlock_t *lp asm("g1"); in __arch_read_lock() 82 lp = rw; in __arch_read_lock() 88 : "r" (lp) in __arch_read_lock() 101 register arch_rwlock_t *lp asm("g1"); in __arch_read_unlock() 102 lp = rw; in __arch_read_unlock() 108 : "r" (lp) in __arch_read_unlock() 121 register arch_rwlock_t *lp asm("g1"); in arch_write_lock() 122 lp = rw; in arch_write_lock() 128 : "r" (lp) in arch_write_lock() 130 *(volatile __u32 *)&lp->lock = ~0U; in arch_write_lock() [all …]
|
/arch/s390/include/asm/ |
D | spinlock.h | 53 static inline int arch_spin_is_locked(arch_spinlock_t *lp) in arch_spin_is_locked() argument 55 return READ_ONCE(lp->lock) != 0; in arch_spin_is_locked() 58 static inline int arch_spin_trylock_once(arch_spinlock_t *lp) in arch_spin_trylock_once() argument 61 return likely(__atomic_cmpxchg_bool(&lp->lock, 0, SPINLOCK_LOCKVAL)); in arch_spin_trylock_once() 64 static inline void arch_spin_lock(arch_spinlock_t *lp) in arch_spin_lock() argument 66 if (!arch_spin_trylock_once(lp)) in arch_spin_lock() 67 arch_spin_lock_wait(lp); in arch_spin_lock() 70 static inline void arch_spin_lock_flags(arch_spinlock_t *lp, in arch_spin_lock_flags() argument 73 if (!arch_spin_trylock_once(lp)) in arch_spin_lock_flags() 74 arch_spin_lock_wait(lp); in arch_spin_lock_flags() [all …]
|
/arch/um/drivers/ |
D | net_kern.c | 75 struct uml_net_private *lp = netdev_priv(dev); in uml_net_rx() local 80 skb = dev_alloc_skb(lp->max_packet); in uml_net_rx() 84 (*lp->read)(lp->fd, drop_skb, lp); in uml_net_rx() 90 skb_put(skb, lp->max_packet); in uml_net_rx() 92 pkt_len = (*lp->read)(lp->fd, skb, lp); in uml_net_rx() 96 skb->protocol = (*lp->protocol)(skb); in uml_net_rx() 110 struct uml_net_private *lp = in uml_dev_close() local 112 dev_close(lp->dev); in uml_dev_close() 118 struct uml_net_private *lp = netdev_priv(dev); in uml_net_interrupt() local 124 spin_lock(&lp->lock); in uml_net_interrupt() [all …]
|
D | slip_kern.c | 47 static int slip_read(int fd, struct sk_buff *skb, struct uml_net_private *lp) in slip_read() argument 50 (struct slip_data *) &lp->user); in slip_read() 53 static int slip_write(int fd, struct sk_buff *skb, struct uml_net_private *lp) in slip_write() argument 56 (struct slip_data *) &lp->user); in slip_write()
|
D | slirp_kern.c | 52 static int slirp_read(int fd, struct sk_buff *skb, struct uml_net_private *lp) in slirp_read() argument 55 (struct slirp_data *) &lp->user); in slirp_read() 58 static int slirp_write(int fd, struct sk_buff *skb, struct uml_net_private *lp) in slirp_write() argument 61 (struct slirp_data *) &lp->user); in slirp_write()
|
D | vde_kern.c | 39 static int vde_read(int fd, struct sk_buff *skb, struct uml_net_private *lp) in vde_read() argument 41 struct vde_data *pri = (struct vde_data *) &lp->user; in vde_read() 51 static int vde_write(int fd, struct sk_buff *skb, struct uml_net_private *lp) in vde_write() argument 53 struct vde_data *pri = (struct vde_data *) &lp->user; in vde_write()
|
D | daemon_kern.c | 42 static int daemon_read(int fd, struct sk_buff *skb, struct uml_net_private *lp) in daemon_read() argument 48 static int daemon_write(int fd, struct sk_buff *skb, struct uml_net_private *lp) in daemon_write() argument 51 (struct daemon_data *) &lp->user); in daemon_write()
|
D | pcap_kern.c | 34 static int pcap_read(int fd, struct sk_buff *skb, struct uml_net_private *lp) in pcap_read() argument 38 (struct pcap_data *) &lp->user); in pcap_read() 41 static int pcap_write(int fd, struct sk_buff *skb, struct uml_net_private *lp) in pcap_write() argument
|
D | umcast_kern.c | 51 static int umcast_read(int fd, struct sk_buff *skb, struct uml_net_private *lp) in umcast_read() argument 57 static int umcast_write(int fd, struct sk_buff *skb, struct uml_net_private *lp) in umcast_write() argument 60 (struct umcast_data *) &lp->user); in umcast_write()
|
/arch/s390/lib/ |
D | spinlock.c | 117 static inline void arch_spin_lock_queued(arch_spinlock_t *lp) in arch_spin_lock_queued() argument 131 old = READ_ONCE(lp->lock); in arch_spin_lock_queued() 142 if (__atomic_cmpxchg_bool(&lp->lock, old, new)) in arch_spin_lock_queued() 150 if (__atomic_cmpxchg_bool(&lp->lock, old, new)) in arch_spin_lock_queued() 182 old = READ_ONCE(lp->lock); in arch_spin_lock_queued() 187 if (__atomic_cmpxchg_bool(&lp->lock, old, new)) in arch_spin_lock_queued() 211 static inline void arch_spin_lock_classic(arch_spinlock_t *lp) in arch_spin_lock_classic() argument 218 owner = arch_spin_yield_target(READ_ONCE(lp->lock), NULL); in arch_spin_lock_classic() 224 old = arch_load_niai4(&lp->lock); in arch_spin_lock_classic() 229 if (arch_cmpxchg_niai8(&lp->lock, old, new)) { in arch_spin_lock_classic() [all …]
|
/arch/nds32/kernel/ |
D | ex-entry.S | 110 sethi $lp, hi20(ret_from_exception) 111 ori $lp, $lp, lo12(ret_from_exception) 131 sethi $lp, hi20(ret_from_exception) 132 ori $lp, $lp, lo12(ret_from_exception) 144 sethi $lp, hi20(ret_from_intr) 145 ori $lp, $lp, lo12(ret_from_intr)
|
D | perf_event_cpu.c | 1223 frame->lp = ((unsigned long *)frame->fp)[-1]; in unwind_frame_kernel() 1226 if (__kernel_text_address(frame->lp)) in unwind_frame_kernel() 1227 frame->lp = ftrace_graph_ret_addr in unwind_frame_kernel() 1228 (NULL, &graph, frame->lp, NULL); in unwind_frame_kernel() 1249 frame->lp = ((unsigned long *)frame->sp)[1]; in unwind_frame_kernel() 1253 if (__kernel_text_address(frame->lp)) in unwind_frame_kernel() 1254 frame->lp = ftrace_graph_ret_addr in unwind_frame_kernel() 1255 (tsk, &graph, frame->lp, NULL); in unwind_frame_kernel() 1292 perf_callchain_store(entry, fr->lp); in callchain_trace() 1304 unsigned long lp = 0; in user_backtrace() local [all …]
|
D | ex-scall.S | 23 smw.bi $r6, [$p1], $r14, #0xb ! push r6~r14, fp, lp, sp 29 lmw.bi $r6, [$r1], $r14, #0xb ! pop r6~r14, fp, lp, sp 59 la $lp, ret_fast_syscall ! return address 82 la $lp, __sys_trace_return ! return address
|
/arch/nds32/lib/ |
D | memzero.S | 9 push $lp 15 pop $lp
|
/arch/powerpc/kernel/ |
D | paca.c | 123 struct lppaca *lp; in new_lppaca() local 131 lp = alloc_shared_lppaca(LPPACA_SIZE, limit, cpu); in new_lppaca() 133 lp = alloc_paca_data(LPPACA_SIZE, 0x400, limit, cpu); in new_lppaca() 135 init_lppaca(lp); in new_lppaca() 137 return lp; in new_lppaca()
|
/arch/powerpc/include/asm/ |
D | cputime.h | 94 struct lppaca *lp = local_paca->lppaca_ptr; in account_stolen_time() local 96 if (unlikely(local_paca->dtl_ridx != be64_to_cpu(lp->dtl_idx))) in account_stolen_time()
|
/arch/um/os-Linux/drivers/ |
D | tuntap_kern.c | 38 static int tuntap_read(int fd, struct sk_buff *skb, struct uml_net_private *lp) in tuntap_read() argument 44 static int tuntap_write(int fd, struct sk_buff *skb, struct uml_net_private *lp) in tuntap_write() argument
|
D | ethertap_kern.c | 39 static int etap_read(int fd, struct sk_buff *skb, struct uml_net_private *lp) in etap_read() argument 53 static int etap_write(int fd, struct sk_buff *skb, struct uml_net_private *lp) in etap_write() argument
|
/arch/nds32/include/uapi/asm/ |
D | ptrace.h | 16 long lp; member
|
/arch/nds32/include/asm/ |
D | stacktrace.h | 11 unsigned long lp; member
|