Home
last modified time | relevance | path

Searched refs:mr (Results 1 – 25 of 301) sorted by relevance

12345678910>>...13

/kernel/linux/linux-5.10/drivers/infiniband/sw/rdmavt/
Dmr.c117 static void rvt_deinit_mregion(struct rvt_mregion *mr) in rvt_deinit_mregion() argument
119 int i = mr->mapsz; in rvt_deinit_mregion()
121 mr->mapsz = 0; in rvt_deinit_mregion()
123 kfree(mr->map[--i]); in rvt_deinit_mregion()
124 percpu_ref_exit(&mr->refcount); in rvt_deinit_mregion()
129 struct rvt_mregion *mr = container_of(ref, struct rvt_mregion, in __rvt_mregion_complete() local
132 complete(&mr->comp); in __rvt_mregion_complete()
135 static int rvt_init_mregion(struct rvt_mregion *mr, struct ib_pd *pd, in rvt_init_mregion() argument
141 mr->mapsz = 0; in rvt_init_mregion()
144 mr->map[i] = kzalloc_node(sizeof(*mr->map[0]), GFP_KERNEL, in rvt_init_mregion()
[all …]
Dtrace_mr.h63 TP_PROTO(struct rvt_mregion *mr, u16 m, u16 n, void *v, size_t len),
64 TP_ARGS(mr, m, n, v, len),
66 RDI_DEV_ENTRY(ib_to_rvt(mr->pd->device))
79 RDI_DEV_ASSIGN(ib_to_rvt(mr->pd->device));
82 __entry->iova = mr->iova;
83 __entry->user_base = mr->user_base;
84 __entry->lkey = mr->lkey;
88 __entry->length = mr->length;
89 __entry->offset = mr->offset;
109 TP_PROTO(struct rvt_mregion *mr, u16 m, u16 n, void *v, size_t len),
[all …]
/kernel/linux/linux-5.10/drivers/vdpa/mlx5/core/
Dmr.c34 static void populate_mtts(struct mlx5_vdpa_direct_mr *mr, __be64 *mtt) in populate_mtts() argument
37 int nsg = mr->nsg; in populate_mtts()
43 for_each_sg(mr->sg_head.sgl, sg, mr->nent, i) { in populate_mtts()
46 nsg--, dma_addr += BIT(mr->log_size), dma_len -= BIT(mr->log_size)) in populate_mtts()
51 static int create_direct_mr(struct mlx5_vdpa_dev *mvdev, struct mlx5_vdpa_direct_mr *mr) in create_direct_mr() argument
58 inlen = MLX5_ST_SZ_BYTES(create_mkey_in) + roundup(MLX5_ST_SZ_BYTES(mtt) * mr->nsg, 16); in create_direct_mr()
65 MLX5_SET(mkc, mkc, lw, !!(mr->perm & VHOST_MAP_WO)); in create_direct_mr()
66 MLX5_SET(mkc, mkc, lr, !!(mr->perm & VHOST_MAP_RO)); in create_direct_mr()
70 MLX5_SET64(mkc, mkc, start_addr, mr->offset); in create_direct_mr()
71 MLX5_SET64(mkc, mkc, len, mr->end - mr->start); in create_direct_mr()
[all …]
/kernel/linux/linux-5.10/drivers/infiniband/hw/mlx5/
Dmr.c109 static void clean_mr(struct mlx5_ib_dev *dev, struct mlx5_ib_mr *mr);
110 static void dereg_mr(struct mlx5_ib_dev *dev, struct mlx5_ib_mr *mr);
119 static int destroy_mkey(struct mlx5_ib_dev *dev, struct mlx5_ib_mr *mr) in destroy_mkey() argument
121 WARN_ON(xa_load(&dev->odp_mkeys, mlx5_base_mkey(mr->mmkey.key))); in destroy_mkey()
123 return mlx5_core_destroy_mkey(dev->mdev, &mr->mmkey); in destroy_mkey()
126 static inline bool mlx5_ib_pas_fits_in_mr(struct mlx5_ib_mr *mr, u64 start, in mlx5_ib_pas_fits_in_mr() argument
129 return ((u64)1 << mr->order) * MLX5_ADAPTER_PAGE_SIZE >= in mlx5_ib_pas_fits_in_mr()
135 struct mlx5_ib_mr *mr = in create_mkey_callback() local
137 struct mlx5_ib_dev *dev = mr->dev; in create_mkey_callback()
138 struct mlx5_cache_ent *ent = mr->cache_ent; in create_mkey_callback()
[all …]
Dodp.c161 struct mlx5_ib_mr *mr, int flags) in populate_mtt() argument
163 struct ib_umem_odp *odp = to_ib_umem_odp(mr->umem); in populate_mtt()
177 struct mlx5_ib_mr *mr, int flags) in mlx5_odp_populate_xlt() argument
180 populate_klm(xlt, idx, nentries, mr, flags); in mlx5_odp_populate_xlt()
182 populate_mtt(xlt, idx, nentries, mr, flags); in mlx5_odp_populate_xlt()
186 static void dma_fence_odp_mr(struct mlx5_ib_mr *mr) in dma_fence_odp_mr() argument
188 struct ib_umem_odp *odp = to_ib_umem_odp(mr->umem); in dma_fence_odp_mr()
193 mlx5_mr_cache_invalidate(mr); in dma_fence_odp_mr()
201 if (!mr->cache_ent) { in dma_fence_odp_mr()
202 mlx5_core_destroy_mkey(mr->dev->mdev, &mr->mmkey); in dma_fence_odp_mr()
[all …]
/kernel/linux/linux-5.10/drivers/infiniband/hw/vmw_pvrdma/
Dpvrdma_mr.c61 struct pvrdma_user_mr *mr; in pvrdma_get_dma_mr() local
75 mr = kzalloc(sizeof(*mr), GFP_KERNEL); in pvrdma_get_dma_mr()
76 if (!mr) in pvrdma_get_dma_mr()
89 kfree(mr); in pvrdma_get_dma_mr()
93 mr->mmr.mr_handle = resp->mr_handle; in pvrdma_get_dma_mr()
94 mr->ibmr.lkey = resp->lkey; in pvrdma_get_dma_mr()
95 mr->ibmr.rkey = resp->rkey; in pvrdma_get_dma_mr()
97 return &mr->ibmr; in pvrdma_get_dma_mr()
116 struct pvrdma_user_mr *mr = NULL; in pvrdma_reg_user_mr() local
144 mr = kzalloc(sizeof(*mr), GFP_KERNEL); in pvrdma_reg_user_mr()
[all …]
/kernel/linux/linux-5.10/net/sunrpc/xprtrdma/
Dfrwr_ops.c57 void frwr_release_mr(struct rpcrdma_mr *mr) in frwr_release_mr() argument
61 rc = ib_dereg_mr(mr->frwr.fr_mr); in frwr_release_mr()
63 trace_xprtrdma_frwr_dereg(mr, rc); in frwr_release_mr()
64 kfree(mr->mr_sg); in frwr_release_mr()
65 kfree(mr); in frwr_release_mr()
68 static void frwr_mr_recycle(struct rpcrdma_mr *mr) in frwr_mr_recycle() argument
70 struct rpcrdma_xprt *r_xprt = mr->mr_xprt; in frwr_mr_recycle()
72 trace_xprtrdma_mr_recycle(mr); in frwr_mr_recycle()
74 if (mr->mr_dir != DMA_NONE) { in frwr_mr_recycle()
75 trace_xprtrdma_mr_unmap(mr); in frwr_mr_recycle()
[all …]
/kernel/linux/linux-5.10/drivers/gpu/drm/nouveau/nvkm/subdev/fb/
Dgddr5.c75 ram->mr[0] &= ~0xf7f; in nvkm_gddr5_calc()
76 ram->mr[0] |= (WR & 0x0f) << 8; in nvkm_gddr5_calc()
77 ram->mr[0] |= (CL & 0x0f) << 3; in nvkm_gddr5_calc()
78 ram->mr[0] |= (WL & 0x07) << 0; in nvkm_gddr5_calc()
80 ram->mr[1] &= ~0x0bf; in nvkm_gddr5_calc()
81 ram->mr[1] |= (xd & 0x01) << 7; in nvkm_gddr5_calc()
82 ram->mr[1] |= (at[0] & 0x03) << 4; in nvkm_gddr5_calc()
83 ram->mr[1] |= (dt & 0x03) << 2; in nvkm_gddr5_calc()
84 ram->mr[1] |= (ds & 0x03) << 0; in nvkm_gddr5_calc()
89 ram->mr1_nuts = ram->mr[1]; in nvkm_gddr5_calc()
[all …]
Dsddr3.c92 ODT = (ram->mr[1] & 0x004) >> 2 | in nvkm_sddr3_calc()
93 (ram->mr[1] & 0x040) >> 5 | in nvkm_sddr3_calc()
94 (ram->mr[1] & 0x200) >> 7; in nvkm_sddr3_calc()
106 ram->mr[0] &= ~0xf74; in nvkm_sddr3_calc()
107 ram->mr[0] |= (WR & 0x07) << 9; in nvkm_sddr3_calc()
108 ram->mr[0] |= (CL & 0x0e) << 3; in nvkm_sddr3_calc()
109 ram->mr[0] |= (CL & 0x01) << 2; in nvkm_sddr3_calc()
111 ram->mr[1] &= ~0x245; in nvkm_sddr3_calc()
112 ram->mr[1] |= (ODT & 0x1) << 2; in nvkm_sddr3_calc()
113 ram->mr[1] |= (ODT & 0x2) << 5; in nvkm_sddr3_calc()
[all …]
Dgddr3.c89 DLL = !(ram->mr[1] & 0x1); in nvkm_gddr3_calc()
90 RON = !((ram->mr[1] & 0x300) >> 8); in nvkm_gddr3_calc()
98 ODT = (ram->mr[1] & 0xc) >> 2; in nvkm_gddr3_calc()
101 hi = ram->mr[2] & 0x1; in nvkm_gddr3_calc()
107 ram->mr[0] &= ~0xf74; in nvkm_gddr3_calc()
108 ram->mr[0] |= (CWL & 0x07) << 9; in nvkm_gddr3_calc()
109 ram->mr[0] |= (CL & 0x07) << 4; in nvkm_gddr3_calc()
110 ram->mr[0] |= (CL & 0x08) >> 1; in nvkm_gddr3_calc()
112 ram->mr[1] &= ~0x3fc; in nvkm_gddr3_calc()
113 ram->mr[1] |= (ODT & 0x03) << 2; in nvkm_gddr3_calc()
[all …]
/kernel/linux/linux-5.10/drivers/scsi/
Dmesh.c305 volatile struct mesh_regs __iomem *mr = ms->mesh; in mesh_dump_regs() local
311 ms, mr, md); in mesh_dump_regs()
314 (mr->count_hi << 8) + mr->count_lo, mr->sequence, in mesh_dump_regs()
315 (mr->bus_status1 << 8) + mr->bus_status0, mr->fifo_count, in mesh_dump_regs()
316 mr->exception, mr->error, mr->intr_mask, mr->interrupt, in mesh_dump_regs()
317 mr->sync_params); in mesh_dump_regs()
318 while(in_8(&mr->fifo_count)) in mesh_dump_regs()
319 printk(KERN_DEBUG " fifo data=%.2x\n",in_8(&mr->fifo)); in mesh_dump_regs()
339 static inline void mesh_flush_io(volatile struct mesh_regs __iomem *mr) in mesh_flush_io() argument
341 (void)in_8(&mr->mesh_id); in mesh_flush_io()
[all …]
/kernel/linux/linux-5.10/drivers/infiniband/hw/mlx4/
Dmr.c60 struct mlx4_ib_mr *mr; in mlx4_ib_get_dma_mr() local
63 mr = kzalloc(sizeof(*mr), GFP_KERNEL); in mlx4_ib_get_dma_mr()
64 if (!mr) in mlx4_ib_get_dma_mr()
68 ~0ull, convert_access(acc), 0, 0, &mr->mmr); in mlx4_ib_get_dma_mr()
72 err = mlx4_mr_enable(to_mdev(pd->device)->dev, &mr->mmr); in mlx4_ib_get_dma_mr()
76 mr->ibmr.rkey = mr->ibmr.lkey = mr->mmr.key; in mlx4_ib_get_dma_mr()
77 mr->umem = NULL; in mlx4_ib_get_dma_mr()
79 return &mr->ibmr; in mlx4_ib_get_dma_mr()
82 (void) mlx4_mr_free(to_mdev(pd->device)->dev, &mr->mmr); in mlx4_ib_get_dma_mr()
85 kfree(mr); in mlx4_ib_get_dma_mr()
[all …]
/kernel/linux/linux-5.10/drivers/infiniband/hw/hns/
Dhns_roce_mr.c69 static int alloc_mr_key(struct hns_roce_dev *hr_dev, struct hns_roce_mr *mr, in alloc_mr_key() argument
85 mr->iova = iova; /* MR va starting addr */ in alloc_mr_key()
86 mr->size = size; /* MR addr range */ in alloc_mr_key()
87 mr->pd = pd; /* MR num */ in alloc_mr_key()
88 mr->access = access; /* MR access permit */ in alloc_mr_key()
89 mr->enabled = 0; /* MR active status */ in alloc_mr_key()
90 mr->key = hw_index_to_key(obj); /* MR key */ in alloc_mr_key()
104 static void free_mr_key(struct hns_roce_dev *hr_dev, struct hns_roce_mr *mr) in free_mr_key() argument
106 unsigned long obj = key_to_hw_index(mr->key); in free_mr_key()
112 static int alloc_mr_pbl(struct hns_roce_dev *hr_dev, struct hns_roce_mr *mr, in alloc_mr_pbl() argument
[all …]
/kernel/linux/linux-5.10/arch/powerpc/platforms/pseries/
DhvCall.S62 mr r4,r3; \
63 mr r3,r0; \
75 mr r5,BUFREG; \
147 mr r4,r5
148 mr r5,r6
149 mr r6,r7
150 mr r7,r8
151 mr r8,r9
152 mr r9,r10
172 mr r0,r4
[all …]
/kernel/linux/linux-5.10/drivers/infiniband/core/
Dmr_pool.c10 struct ib_mr *mr; in ib_mr_pool_get() local
14 mr = list_first_entry_or_null(list, struct ib_mr, qp_entry); in ib_mr_pool_get()
15 if (mr) { in ib_mr_pool_get()
16 list_del(&mr->qp_entry); in ib_mr_pool_get()
21 return mr; in ib_mr_pool_get()
25 void ib_mr_pool_put(struct ib_qp *qp, struct list_head *list, struct ib_mr *mr) in ib_mr_pool_put() argument
30 list_add(&mr->qp_entry, list); in ib_mr_pool_put()
39 struct ib_mr *mr; in ib_mr_pool_init() local
45 mr = ib_alloc_mr_integrity(qp->pd, max_num_sg, in ib_mr_pool_init()
48 mr = ib_alloc_mr(qp->pd, type, max_num_sg); in ib_mr_pool_init()
[all …]
Duverbs_std_types_mr.c94 struct ib_mr *mr; in UVERBS_HANDLER() local
126 mr = pd->device->ops.reg_dm_mr(pd, dm, &attr, attrs); in UVERBS_HANDLER()
127 if (IS_ERR(mr)) in UVERBS_HANDLER()
128 return PTR_ERR(mr); in UVERBS_HANDLER()
130 mr->device = pd->device; in UVERBS_HANDLER()
131 mr->pd = pd; in UVERBS_HANDLER()
132 mr->type = IB_MR_TYPE_DM; in UVERBS_HANDLER()
133 mr->dm = dm; in UVERBS_HANDLER()
134 mr->uobject = uobj; in UVERBS_HANDLER()
138 rdma_restrack_new(&mr->res, RDMA_RESTRACK_MR); in UVERBS_HANDLER()
[all …]
/kernel/linux/linux-5.10/net/rds/
Drdma.c70 struct rds_mr *mr; in rds_mr_tree_walk() local
74 mr = rb_entry(parent, struct rds_mr, r_rb_node); in rds_mr_tree_walk()
76 if (key < mr->r_key) in rds_mr_tree_walk()
78 else if (key > mr->r_key) in rds_mr_tree_walk()
81 return mr; in rds_mr_tree_walk()
95 static void rds_destroy_mr(struct rds_mr *mr) in rds_destroy_mr() argument
97 struct rds_sock *rs = mr->r_sock; in rds_destroy_mr()
102 mr->r_key, kref_read(&mr->r_kref)); in rds_destroy_mr()
105 if (!RB_EMPTY_NODE(&mr->r_rb_node)) in rds_destroy_mr()
106 rb_erase(&mr->r_rb_node, &rs->rs_rdma_keys); in rds_destroy_mr()
[all …]
/kernel/linux/linux-5.10/drivers/rtc/
Drtc-at91sam9.c133 u32 offset, alarm, mr; in at91_rtc_settime() local
140 mr = rtt_readl(rtc, MR); in at91_rtc_settime()
143 rtt_writel(rtc, MR, mr & ~(AT91_RTT_ALMIEN | AT91_RTT_RTTINCIEN)); in at91_rtc_settime()
164 mr &= ~AT91_RTT_ALMIEN; in at91_rtc_settime()
170 rtt_writel(rtc, MR, mr | AT91_RTT_RTTRST); in at91_rtc_settime()
205 u32 mr; in at91_rtc_setalarm() local
214 mr = rtt_readl(rtc, MR); in at91_rtc_setalarm()
215 rtt_writel(rtc, MR, mr & ~AT91_RTT_ALMIEN); in at91_rtc_setalarm()
226 rtt_writel(rtc, MR, mr | AT91_RTT_ALMIEN); in at91_rtc_setalarm()
236 u32 mr = rtt_readl(rtc, MR); in at91_rtc_alarm_irq_enable() local
[all …]
/kernel/linux/linux-5.10/drivers/net/ethernet/mellanox/mlxsw/
Dspectrum_mr.c254 struct mlxsw_sp_mr *mr = mlxsw_sp->mr; in mlxsw_sp_mr_route_write() local
264 mr_route->route_priv = kzalloc(mr->mr_ops->route_priv_size, in mlxsw_sp_mr_route_write()
274 err = mr->mr_ops->route_create(mlxsw_sp, mr->priv, in mlxsw_sp_mr_route_write()
280 err = mr->mr_ops->route_update(mlxsw_sp, mr_route->route_priv, in mlxsw_sp_mr_route_write()
292 struct mlxsw_sp_mr *mr = mlxsw_sp->mr; in mlxsw_sp_mr_route_erase() local
294 mr->mr_ops->route_destroy(mlxsw_sp, mr->priv, mr_route->route_priv); in mlxsw_sp_mr_route_erase()
485 struct mlxsw_sp_mr *mr = mlxsw_sp->mr; in mlxsw_sp_mr_route_ivif_resolve() local
495 err = mr->mr_ops->route_irif_update(mlxsw_sp, rve->mr_route->route_priv, in mlxsw_sp_mr_route_ivif_resolve()
500 err = mr->mr_ops->route_action_update(mlxsw_sp, in mlxsw_sp_mr_route_ivif_resolve()
519 struct mlxsw_sp_mr *mr = mlxsw_sp->mr; in mlxsw_sp_mr_route_ivif_unresolve() local
[all …]
/kernel/linux/linux-5.10/drivers/sh/intc/
Dhandle.c44 struct intc_mask_reg *mr = desc->hw.mask_regs; in _intc_mask_data() local
48 while (mr && enum_id && *reg_idx < desc->hw.nr_mask_regs) { in _intc_mask_data()
49 mr = desc->hw.mask_regs + *reg_idx; in _intc_mask_data()
51 for (; *fld_idx < ARRAY_SIZE(mr->enum_ids); (*fld_idx)++) { in _intc_mask_data()
52 if (mr->enum_ids[*fld_idx] != enum_id) in _intc_mask_data()
55 if (mr->set_reg && mr->clr_reg) { in _intc_mask_data()
58 reg_e = mr->clr_reg; in _intc_mask_data()
59 reg_d = mr->set_reg; in _intc_mask_data()
62 if (mr->set_reg) { in _intc_mask_data()
64 reg_e = mr->set_reg; in _intc_mask_data()
[all …]
Dbalancing.c44 struct intc_mask_reg *mr = desc->hw.mask_regs; in intc_dist_data() local
48 for (i = 0; mr && enum_id && i < desc->hw.nr_mask_regs; i++) { in intc_dist_data()
49 mr = desc->hw.mask_regs + i; in intc_dist_data()
55 if (!mr->dist_reg) in intc_dist_data()
58 for (j = 0; j < ARRAY_SIZE(mr->enum_ids); j++) { in intc_dist_data()
59 if (mr->enum_ids[j] != enum_id) in intc_dist_data()
64 reg_e = mr->dist_reg; in intc_dist_data()
65 reg_d = mr->dist_reg; in intc_dist_data()
67 fn += (mr->reg_width >> 3) - 1; in intc_dist_data()
72 (mr->reg_width - 1) - j); in intc_dist_data()
/kernel/linux/linux-5.10/arch/x86/mm/
Dinit.c332 static int __meminit save_mr(struct map_range *mr, int nr_range, in save_mr() argument
339 mr[nr_range].start = start_pfn<<PAGE_SHIFT; in save_mr()
340 mr[nr_range].end = end_pfn<<PAGE_SHIFT; in save_mr()
341 mr[nr_range].page_size_mask = page_size_mask; in save_mr()
352 static void __ref adjust_range_page_size_mask(struct map_range *mr, in adjust_range_page_size_mask() argument
359 !(mr[i].page_size_mask & (1<<PG_LEVEL_2M))) { in adjust_range_page_size_mask()
360 unsigned long start = round_down(mr[i].start, PMD_SIZE); in adjust_range_page_size_mask()
361 unsigned long end = round_up(mr[i].end, PMD_SIZE); in adjust_range_page_size_mask()
369 mr[i].page_size_mask |= 1<<PG_LEVEL_2M; in adjust_range_page_size_mask()
372 !(mr[i].page_size_mask & (1<<PG_LEVEL_1G))) { in adjust_range_page_size_mask()
[all …]
/kernel/linux/linux-5.10/net/mac80211/
Drc80211_minstrel.c73 int minstrel_get_tp_avg(struct minstrel_rate *mr, int prob_avg) in minstrel_get_tp_avg() argument
77 usecs = mr->perfect_tx_time; in minstrel_get_tp_avg()
82 if (mr->stats.prob_avg < MINSTREL_FRAC(10, 100)) in minstrel_get_tp_avg()
202 struct minstrel_rate *mr = &mi->r[i]; in minstrel_update_stats() local
213 mr->adjusted_retry_count = mrs->retry_count >> 1; in minstrel_update_stats()
214 if (mr->adjusted_retry_count > 2) in minstrel_update_stats()
215 mr->adjusted_retry_count = 2; in minstrel_update_stats()
216 mr->sample_limit = 4; in minstrel_update_stats()
218 mr->sample_limit = -1; in minstrel_update_stats()
219 mr->adjusted_retry_count = mrs->retry_count; in minstrel_update_stats()
[all …]
/kernel/linux/linux-5.10/drivers/infiniband/ulp/iser/
Diser_memory.c220 struct ib_mr *mr, in iser_inv_rkey() argument
226 inv_wr->ex.invalidate_rkey = mr->rkey; in iser_inv_rkey()
241 struct ib_mr *mr = rsc->sig_mr; in iser_reg_sig_mr() local
242 struct ib_sig_attrs *sig_attrs = mr->sig_attrs; in iser_reg_sig_mr()
254 iser_inv_rkey(&tx_desc->inv_wr, mr, cqe, &wr->wr); in iser_reg_sig_mr()
256 ib_update_fast_reg_key(mr, ib_inc_rkey(mr->rkey)); in iser_reg_sig_mr()
258 ret = ib_map_mr_sg_pi(mr, mem->sg, mem->dma_nents, NULL, in iser_reg_sig_mr()
272 wr->mr = mr; in iser_reg_sig_mr()
273 wr->key = mr->rkey; in iser_reg_sig_mr()
279 sig_reg->sge.lkey = mr->lkey; in iser_reg_sig_mr()
[all …]
/kernel/linux/linux-5.10/drivers/watchdog/
Dsama5d4_wdt.c32 u32 mr; member
52 #define wdt_enabled (!(wdt->mr & AT91_WDT_WDDIS))
87 wdt->mr &= ~AT91_SAM9X60_WDDIS; in sama5d4_wdt_start()
89 wdt->mr &= ~AT91_WDT_WDDIS; in sama5d4_wdt_start()
91 wdt_write(wdt, AT91_WDT_MR, wdt->mr); in sama5d4_wdt_start()
102 wdt->mr |= AT91_SAM9X60_WDDIS; in sama5d4_wdt_stop()
104 wdt->mr |= AT91_WDT_WDDIS; in sama5d4_wdt_stop()
106 wdt_write(wdt, AT91_WDT_MR, wdt->mr); in sama5d4_wdt_stop()
134 wdt->mr &= ~AT91_WDT_WDV; in sama5d4_wdt_set_timeout()
135 wdt->mr |= AT91_WDT_SET_WDV(value); in sama5d4_wdt_set_timeout()
[all …]

12345678910>>...13