Home
last modified time | relevance | path

Searched refs:ns (Results 1 – 25 of 299) sorted by relevance

12345678910>>...12

/drivers/mtd/nand/raw/
Dnandsim.c189 #define NS_IS_INITIALIZED(ns) ((ns)->geom.totsz != 0) argument
192 #define NS_STATUS_OK(ns) (NAND_STATUS_READY | (NAND_STATUS_WP * ((ns)->lines.wp == 0))) argument
195 #define NS_STATUS_FAILED(ns) (NAND_STATUS_FAIL | NS_STATUS_OK(ns)) argument
198 #define NS_RAW_OFFSET(ns) \ argument
199 (((ns)->regs.row * (ns)->geom.pgszoob) + (ns)->regs.column)
202 #define NS_RAW_OFFSET_OOB(ns) (NS_RAW_OFFSET(ns) + ns->geom.pgsz) argument
205 #define NS_PAGE_BYTE_SHIFT(ns) ((ns)->regs.column + (ns)->regs.off) argument
491 DEFINE_SHOW_ATTRIBUTE(ns);
500 static int ns_debugfs_create(struct nandsim *ns) in ns_debugfs_create() argument
515 ns->dent = debugfs_create_file("nandsim_wear_report", 0400, root, ns, in ns_debugfs_create()
[all …]
/drivers/net/netdevsim/
Dethtool.c14 struct netdevsim *ns = netdev_priv(dev); in nsim_get_pause_stats() local
16 if (ns->ethtool.pauseparam.report_stats_rx) in nsim_get_pause_stats()
18 if (ns->ethtool.pauseparam.report_stats_tx) in nsim_get_pause_stats()
25 struct netdevsim *ns = netdev_priv(dev); in nsim_get_pauseparam() local
28 pause->rx_pause = ns->ethtool.pauseparam.rx; in nsim_get_pauseparam()
29 pause->tx_pause = ns->ethtool.pauseparam.tx; in nsim_get_pauseparam()
35 struct netdevsim *ns = netdev_priv(dev); in nsim_set_pauseparam() local
40 ns->ethtool.pauseparam.rx = pause->rx_pause; in nsim_set_pauseparam()
41 ns->ethtool.pauseparam.tx = pause->tx_pause; in nsim_set_pauseparam()
50 struct netdevsim *ns = netdev_priv(dev); in nsim_get_coalesce() local
[all …]
Dnetdev.c31 struct netdevsim *ns = netdev_priv(dev); in nsim_start_xmit() local
33 if (!nsim_ipsec_tx(ns, skb)) in nsim_start_xmit()
36 u64_stats_update_begin(&ns->syncp); in nsim_start_xmit()
37 ns->tx_packets++; in nsim_start_xmit()
38 ns->tx_bytes += skb->len; in nsim_start_xmit()
39 u64_stats_update_end(&ns->syncp); in nsim_start_xmit()
53 struct netdevsim *ns = netdev_priv(dev); in nsim_change_mtu() local
55 if (ns->xdp.prog && new_mtu > NSIM_XDP_MAX_MTU) in nsim_change_mtu()
66 struct netdevsim *ns = netdev_priv(dev); in nsim_get_stats64() local
70 start = u64_stats_fetch_begin(&ns->syncp); in nsim_get_stats64()
[all …]
Dudp_tunnels.c15 struct netdevsim *ns = netdev_priv(dev); in nsim_udp_tunnel_set_port() local
18 ret = -ns->udp_ports.inject_error; in nsim_udp_tunnel_set_port()
19 ns->udp_ports.inject_error = 0; in nsim_udp_tunnel_set_port()
21 if (ns->udp_ports.sleep) in nsim_udp_tunnel_set_port()
22 msleep(ns->udp_ports.sleep); in nsim_udp_tunnel_set_port()
25 if (ns->udp_ports.ports[table][entry]) { in nsim_udp_tunnel_set_port()
29 ns->udp_ports.ports[table][entry] = in nsim_udp_tunnel_set_port()
44 struct netdevsim *ns = netdev_priv(dev); in nsim_udp_tunnel_unset_port() local
47 ret = -ns->udp_ports.inject_error; in nsim_udp_tunnel_unset_port()
48 ns->udp_ports.inject_error = 0; in nsim_udp_tunnel_unset_port()
[all …]
Dbpf.c41 struct netdevsim *ns; member
87 static bool nsim_xdp_offload_active(struct netdevsim *ns) in nsim_xdp_offload_active() argument
89 return ns->xdp_hw.prog; in nsim_xdp_offload_active()
104 nsim_bpf_offload(struct netdevsim *ns, struct bpf_prog *prog, bool oldprog) in nsim_bpf_offload() argument
106 nsim_prog_set_loaded(ns->bpf_offloaded, false); in nsim_bpf_offload()
108 WARN(!!ns->bpf_offloaded != oldprog, in nsim_bpf_offload()
111 ns->bpf_offloaded = prog; in nsim_bpf_offload()
112 ns->bpf_offloaded_id = prog ? prog->aux->id : 0; in nsim_bpf_offload()
123 struct netdevsim *ns = cb_priv; in nsim_bpf_setup_tc_block_cb() local
132 if (!tc_cls_can_offload_and_chain0(ns->netdev, &cls_bpf->common)) in nsim_bpf_setup_tc_block_cb()
[all …]
Dmacsec.c11 static int nsim_macsec_find_secy(struct netdevsim *ns, sci_t sci) in nsim_macsec_find_secy() argument
16 if (ns->macsec.nsim_secy[i].sci == sci) in nsim_macsec_find_secy()
37 struct netdevsim *ns = netdev_priv(ctx->netdev); in nsim_macsec_add_secy() local
40 if (ns->macsec.nsim_secy_count == NSIM_MACSEC_MAX_SECY_COUNT) in nsim_macsec_add_secy()
44 if (!ns->macsec.nsim_secy[idx].used) in nsim_macsec_add_secy()
56 ns->macsec.nsim_secy[idx].used = true; in nsim_macsec_add_secy()
57 ns->macsec.nsim_secy[idx].nsim_rxsc_count = 0; in nsim_macsec_add_secy()
58 ns->macsec.nsim_secy[idx].sci = ctx->secy->sci; in nsim_macsec_add_secy()
59 ns->macsec.nsim_secy_count++; in nsim_macsec_add_secy()
66 struct netdevsim *ns = netdev_priv(ctx->netdev); in nsim_macsec_upd_secy() local
[all …]
Dipsec.c16 struct netdevsim *ns = filp->private_data; in nsim_dbg_netdev_ops_read() local
17 struct nsim_ipsec *ipsec = &ns->ipsec; in nsim_dbg_netdev_ops_read()
133 struct netdevsim *ns; in nsim_ipsec_add_sa() local
139 ns = netdev_priv(dev); in nsim_ipsec_add_sa()
140 ipsec = &ns->ipsec; in nsim_ipsec_add_sa()
202 struct netdevsim *ns = netdev_priv(xs->xso.real_dev); in nsim_ipsec_del_sa() local
203 struct nsim_ipsec *ipsec = &ns->ipsec; in nsim_ipsec_del_sa()
208 netdev_err(ns->netdev, "Invalid SA for delete sa_idx=%d\n", in nsim_ipsec_del_sa()
219 struct netdevsim *ns = netdev_priv(xs->xso.real_dev); in nsim_ipsec_offload_ok() local
220 struct nsim_ipsec *ipsec = &ns->ipsec; in nsim_ipsec_offload_ok()
[all …]
Dnetdevsim.h132 void nsim_destroy(struct netdevsim *ns);
134 void nsim_ethtool_init(struct netdevsim *ns);
144 int nsim_bpf_init(struct netdevsim *ns);
145 void nsim_bpf_uninit(struct netdevsim *ns);
147 int nsim_bpf_disable_tc(struct netdevsim *ns);
160 static inline int nsim_bpf_init(struct netdevsim *ns) in nsim_bpf_init() argument
165 static inline void nsim_bpf_uninit(struct netdevsim *ns) in nsim_bpf_uninit() argument
174 static inline int nsim_bpf_disable_tc(struct netdevsim *ns) in nsim_bpf_disable_tc() argument
262 struct netdevsim *ns; member
374 void nsim_ipsec_init(struct netdevsim *ns);
[all …]
/drivers/block/drbd/
Ddrbd_state.c26 union drbd_state ns; member
281 union drbd_state ns, enum chg_state_flags flags,
285 static enum drbd_state_rv is_valid_transition(union drbd_state os, union drbd_state ns);
287 union drbd_state ns, enum sanitize_state_warnings *warn);
468 union drbd_state os, union drbd_state ns) in cl_wide_st_chg() argument
470 return (os.conn >= C_CONNECTED && ns.conn >= C_CONNECTED && in cl_wide_st_chg()
471 ((os.role != R_PRIMARY && ns.role == R_PRIMARY) || in cl_wide_st_chg()
472 (os.conn != C_STARTING_SYNC_T && ns.conn == C_STARTING_SYNC_T) || in cl_wide_st_chg()
473 (os.conn != C_STARTING_SYNC_S && ns.conn == C_STARTING_SYNC_S) || in cl_wide_st_chg()
474 (os.disk != D_FAILED && ns.disk == D_FAILED))) || in cl_wide_st_chg()
[all …]
/drivers/nvme/host/
Dmultipath.c85 struct nvme_ns *ns = req->q->queuedata; in nvme_failover_req() local
90 nvme_mpath_clear_current_path(ns); in nvme_failover_req()
97 if (nvme_is_ana_error(status) && ns->ctrl->ana_log_buf) { in nvme_failover_req()
98 set_bit(NVME_NS_ANA_PENDING, &ns->flags); in nvme_failover_req()
99 queue_work(nvme_wq, &ns->ctrl->ana_work); in nvme_failover_req()
102 spin_lock_irqsave(&ns->head->requeue_lock, flags); in nvme_failover_req()
104 bio_set_dev(bio, ns->head->disk->part0); in nvme_failover_req()
118 blk_steal_bios(&ns->head->requeue_list, req); in nvme_failover_req()
119 spin_unlock_irqrestore(&ns->head->requeue_lock, flags); in nvme_failover_req()
123 kblockd_schedule_work(&ns->head->requeue_work); in nvme_failover_req()
[all …]
Dzns.c10 int nvme_revalidate_zones(struct nvme_ns *ns) in nvme_revalidate_zones() argument
12 struct request_queue *q = ns->queue; in nvme_revalidate_zones()
14 blk_queue_chunk_sectors(q, ns->zsze); in nvme_revalidate_zones()
15 blk_queue_max_zone_append_sectors(q, ns->ctrl->max_zone_append); in nvme_revalidate_zones()
17 return blk_revalidate_disk_zones(ns->disk, NULL); in nvme_revalidate_zones()
48 int nvme_update_zone_info(struct nvme_ns *ns, unsigned lbaf) in nvme_update_zone_info() argument
50 struct nvme_effects_log *log = ns->head->effects; in nvme_update_zone_info()
51 struct request_queue *q = ns->queue; in nvme_update_zone_info()
59 if (test_and_clear_bit(NVME_NS_FORCE_RO, &ns->flags)) in nvme_update_zone_info()
60 dev_warn(ns->ctrl->device, in nvme_update_zone_info()
[all …]
Dcore.c311 struct nvme_ns *ns = req->q->queuedata; in nvme_log_error() local
314 if (ns) { in nvme_log_error()
316 ns->disk ? ns->disk->disk_name : "?", in nvme_log_error()
319 (unsigned long long)nvme_sect_to_lba(ns, blk_rq_pos(req)), in nvme_log_error()
320 (unsigned long long)blk_rq_bytes(req) >> ns->lba_shift, in nvme_log_error()
627 struct nvme_ns *ns = container_of(kref, struct nvme_ns, kref); in nvme_free_ns() local
629 put_disk(ns->disk); in nvme_free_ns()
630 nvme_put_ns_head(ns->head); in nvme_free_ns()
631 nvme_put_ctrl(ns->ctrl); in nvme_free_ns()
632 kfree(ns); in nvme_free_ns()
[all …]
Dioctl.c16 static bool nvme_cmd_allowed(struct nvme_ns *ns, struct nvme_command *c, in nvme_cmd_allowed() argument
42 if (!ns) { in nvme_cmd_allowed()
61 effects = nvme_command_effects(ns->ctrl, ns, c->common.opcode); in nvme_cmd_allowed()
172 struct nvme_ns *ns = q->queuedata; in nvme_map_user_request() local
173 struct block_device *bdev = ns ? ns->disk->part0 : NULL; in nvme_map_user_request()
226 struct nvme_ns *ns = q->queuedata; in nvme_submit_user_cmd() local
249 effects = nvme_passthru_start(ctrl, ns, cmd->common.opcode); in nvme_submit_user_cmd()
261 nvme_passthru_end(ctrl, ns, effects, cmd, ret); in nvme_submit_user_cmd()
266 static int nvme_submit_io(struct nvme_ns *ns, struct nvme_user_io __user *uio) in nvme_submit_io() argument
287 length = (io.nblocks + 1) << ns->lba_shift; in nvme_submit_io()
[all …]
/drivers/nvme/target/
Dio-cmd-bdev.c51 void nvmet_bdev_ns_disable(struct nvmet_ns *ns) in nvmet_bdev_ns_disable() argument
53 if (ns->bdev) { in nvmet_bdev_ns_disable()
54 blkdev_put(ns->bdev, NULL); in nvmet_bdev_ns_disable()
55 ns->bdev = NULL; in nvmet_bdev_ns_disable()
59 static void nvmet_bdev_ns_enable_integrity(struct nvmet_ns *ns) in nvmet_bdev_ns_enable_integrity() argument
61 struct blk_integrity *bi = bdev_get_integrity(ns->bdev); in nvmet_bdev_ns_enable_integrity()
64 ns->metadata_size = bi->tuple_size; in nvmet_bdev_ns_enable_integrity()
66 ns->pi_type = NVME_NS_DPS_PI_TYPE1; in nvmet_bdev_ns_enable_integrity()
68 ns->pi_type = NVME_NS_DPS_PI_TYPE3; in nvmet_bdev_ns_enable_integrity()
71 ns->metadata_size = 0; in nvmet_bdev_ns_enable_integrity()
[all …]
Dio-cmd-file.c16 void nvmet_file_ns_revalidate(struct nvmet_ns *ns) in nvmet_file_ns_revalidate() argument
18 ns->size = i_size_read(ns->file->f_mapping->host); in nvmet_file_ns_revalidate()
21 void nvmet_file_ns_disable(struct nvmet_ns *ns) in nvmet_file_ns_disable() argument
23 if (ns->file) { in nvmet_file_ns_disable()
24 if (ns->buffered_io) in nvmet_file_ns_disable()
26 mempool_destroy(ns->bvec_pool); in nvmet_file_ns_disable()
27 ns->bvec_pool = NULL; in nvmet_file_ns_disable()
28 fput(ns->file); in nvmet_file_ns_disable()
29 ns->file = NULL; in nvmet_file_ns_disable()
33 int nvmet_file_ns_enable(struct nvmet_ns *ns) in nvmet_file_ns_enable() argument
[all …]
Dcore.c430 req->ns = xa_load(&subsys->namespaces, nsid); in nvmet_req_find_ns()
431 if (unlikely(!req->ns)) { in nvmet_req_find_ns()
438 percpu_ref_get(&req->ns->ref); in nvmet_req_find_ns()
444 struct nvmet_ns *ns = container_of(ref, struct nvmet_ns, ref); in nvmet_destroy_namespace() local
446 complete(&ns->disable_done); in nvmet_destroy_namespace()
449 void nvmet_put_namespace(struct nvmet_ns *ns) in nvmet_put_namespace() argument
451 percpu_ref_put(&ns->ref); in nvmet_put_namespace()
454 static void nvmet_ns_dev_disable(struct nvmet_ns *ns) in nvmet_ns_dev_disable() argument
456 nvmet_bdev_ns_disable(ns); in nvmet_ns_dev_disable()
457 nvmet_file_ns_disable(ns); in nvmet_ns_dev_disable()
[all …]
Dzns.c35 bool nvmet_bdev_zns_enable(struct nvmet_ns *ns) in nvmet_bdev_zns_enable() argument
37 u8 zasl = nvmet_zasl(bdev_max_zone_append_sectors(ns->bdev)); in nvmet_bdev_zns_enable()
38 struct gendisk *bd_disk = ns->bdev->bd_disk; in nvmet_bdev_zns_enable()
41 if (ns->subsys->zasl) { in nvmet_bdev_zns_enable()
42 if (ns->subsys->zasl > zasl) in nvmet_bdev_zns_enable()
45 ns->subsys->zasl = zasl; in nvmet_bdev_zns_enable()
60 if (ns->bdev->bd_disk->conv_zones_bitmap) in nvmet_bdev_zns_enable()
63 ret = blkdev_report_zones(ns->bdev, 0, bdev_nr_zones(ns->bdev), in nvmet_bdev_zns_enable()
68 ns->blksize_shift = blksize_bits(bdev_logical_block_size(ns->bdev)); in nvmet_bdev_zns_enable()
121 if (nvmet_ns_revalidate(req->ns)) { in nvmet_execute_identify_ns_zns()
[all …]
Dadmin-cmd.c85 if (!req->ns->bdev) in nvmet_get_smart_log_nsid()
88 host_reads = part_stat_read(req->ns->bdev, ios[READ]); in nvmet_get_smart_log_nsid()
90 DIV_ROUND_UP(part_stat_read(req->ns->bdev, sectors[READ]), 1000); in nvmet_get_smart_log_nsid()
91 host_writes = part_stat_read(req->ns->bdev, ios[WRITE]); in nvmet_get_smart_log_nsid()
93 DIV_ROUND_UP(part_stat_read(req->ns->bdev, sectors[WRITE]), 1000); in nvmet_get_smart_log_nsid()
108 struct nvmet_ns *ns; in nvmet_get_smart_log_all() local
113 xa_for_each(&ctrl->subsys->namespaces, idx, ns) { in nvmet_get_smart_log_all()
115 if (!ns->bdev) in nvmet_get_smart_log_all()
117 host_reads += part_stat_read(ns->bdev, ios[READ]); in nvmet_get_smart_log_all()
119 part_stat_read(ns->bdev, sectors[READ]), 1000); in nvmet_get_smart_log_all()
[all …]
/drivers/clk/qcom/
Dclk-rcg.c18 static u32 ns_to_src(struct src_sel *s, u32 ns) in ns_to_src() argument
20 ns >>= s->src_sel_shift; in ns_to_src()
21 ns &= SRC_SEL_MASK; in ns_to_src()
22 return ns; in ns_to_src()
25 static u32 src_to_ns(struct src_sel *s, u8 src, u32 ns) in src_to_ns() argument
31 ns &= ~mask; in src_to_ns()
33 ns |= src << s->src_sel_shift; in src_to_ns()
34 return ns; in src_to_ns()
41 u32 ns; in clk_rcg_get_parent() local
44 ret = regmap_read(rcg->clkr.regmap, rcg->ns_reg, &ns); in clk_rcg_get_parent()
[all …]
/drivers/net/ethernet/mellanox/mlx5/core/
Dfs_cmd.h39 int (*create_flow_table)(struct mlx5_flow_root_namespace *ns,
43 int (*destroy_flow_table)(struct mlx5_flow_root_namespace *ns,
46 int (*modify_flow_table)(struct mlx5_flow_root_namespace *ns,
50 int (*create_flow_group)(struct mlx5_flow_root_namespace *ns,
55 int (*destroy_flow_group)(struct mlx5_flow_root_namespace *ns,
59 int (*create_fte)(struct mlx5_flow_root_namespace *ns,
64 int (*update_fte)(struct mlx5_flow_root_namespace *ns,
70 int (*delete_fte)(struct mlx5_flow_root_namespace *ns,
74 int (*update_root_ft)(struct mlx5_flow_root_namespace *ns,
79 int (*packet_reformat_alloc)(struct mlx5_flow_root_namespace *ns,
[all …]
Dfs_cmd.c43 static int mlx5_cmd_stub_update_root_ft(struct mlx5_flow_root_namespace *ns, in mlx5_cmd_stub_update_root_ft() argument
51 static int mlx5_cmd_stub_create_flow_table(struct mlx5_flow_root_namespace *ns, in mlx5_cmd_stub_create_flow_table() argument
63 static int mlx5_cmd_stub_destroy_flow_table(struct mlx5_flow_root_namespace *ns, in mlx5_cmd_stub_destroy_flow_table() argument
69 static int mlx5_cmd_stub_modify_flow_table(struct mlx5_flow_root_namespace *ns, in mlx5_cmd_stub_modify_flow_table() argument
76 static int mlx5_cmd_stub_create_flow_group(struct mlx5_flow_root_namespace *ns, in mlx5_cmd_stub_create_flow_group() argument
84 static int mlx5_cmd_stub_destroy_flow_group(struct mlx5_flow_root_namespace *ns, in mlx5_cmd_stub_destroy_flow_group() argument
91 static int mlx5_cmd_stub_create_fte(struct mlx5_flow_root_namespace *ns, in mlx5_cmd_stub_create_fte() argument
99 static int mlx5_cmd_stub_update_fte(struct mlx5_flow_root_namespace *ns, in mlx5_cmd_stub_update_fte() argument
108 static int mlx5_cmd_stub_delete_fte(struct mlx5_flow_root_namespace *ns, in mlx5_cmd_stub_delete_fte() argument
115 static int mlx5_cmd_stub_packet_reformat_alloc(struct mlx5_flow_root_namespace *ns, in mlx5_cmd_stub_packet_reformat_alloc() argument
[all …]
/drivers/scsi/bfa/
Dbfa_fcs_lport.c3666 static void bfa_fcs_lport_ns_sm_offline(struct bfa_fcs_lport_ns_s *ns,
3668 static void bfa_fcs_lport_ns_sm_plogi_sending(struct bfa_fcs_lport_ns_s *ns,
3670 static void bfa_fcs_lport_ns_sm_plogi(struct bfa_fcs_lport_ns_s *ns,
3672 static void bfa_fcs_lport_ns_sm_plogi_retry(struct bfa_fcs_lport_ns_s *ns,
3675 struct bfa_fcs_lport_ns_s *ns,
3677 static void bfa_fcs_lport_ns_sm_rspn_id(struct bfa_fcs_lport_ns_s *ns,
3679 static void bfa_fcs_lport_ns_sm_rspn_id_retry(struct bfa_fcs_lport_ns_s *ns,
3682 struct bfa_fcs_lport_ns_s *ns,
3684 static void bfa_fcs_lport_ns_sm_rft_id_retry(struct bfa_fcs_lport_ns_s *ns,
3686 static void bfa_fcs_lport_ns_sm_rft_id(struct bfa_fcs_lport_ns_s *ns,
[all …]
/drivers/net/ethernet/mellanox/mlx5/core/steering/
Dfs_dr.c20 static int mlx5_cmd_dr_update_root_ft(struct mlx5_flow_root_namespace *ns, in mlx5_cmd_dr_update_root_ft() argument
25 return mlx5_fs_cmd_get_fw_cmds()->update_root_ft(ns, ft, underlay_qpn, in mlx5_cmd_dr_update_root_ft()
29 static int set_miss_action(struct mlx5_flow_root_namespace *ns, in set_miss_action() argument
49 mlx5_core_err(ns->dev, in set_miss_action()
57 mlx5_core_err(ns->dev, "Failed to destroy action (%d)\n", in set_miss_action()
64 static int mlx5_cmd_dr_create_flow_table(struct mlx5_flow_root_namespace *ns, in mlx5_cmd_dr_create_flow_table() argument
74 return mlx5_fs_cmd_get_fw_cmds()->create_flow_table(ns, ft, in mlx5_cmd_dr_create_flow_table()
79 if (!MLX5_CAP_FLOWTABLE(ns->dev, sw_owner_reformat_supported)) in mlx5_cmd_dr_create_flow_table()
83 tbl = mlx5dr_table_create(ns->fs_dr_domain.dr_domain, ft->level, flags, in mlx5_cmd_dr_create_flow_table()
86 mlx5_core_err(ns->dev, "Failed creating dr flow_table\n"); in mlx5_cmd_dr_create_flow_table()
[all …]
/drivers/ptp/
Dptp_dte.c52 static void dte_write_nco(void __iomem *regs, s64 ns) in dte_write_nco() argument
56 sum2 = (u32)((ns >> DTE_NCO_SUM2_SHIFT) & DTE_NCO_SUM2_MASK); in dte_write_nco()
62 sum3 = (u32)(((ns >> DTE_NCO_SUM3_SHIFT) & DTE_NCO_SUM3_MASK) << in dte_write_nco()
73 s64 ns; in dte_read_nco() local
81 ns = ((s64)sum3 << DTE_NCO_SUM3_SHIFT) | in dte_read_nco()
84 return ns; in dte_read_nco()
89 s64 ns; in dte_write_nco_delta() local
91 ns = dte_read_nco(ptp_dte->regs); in dte_write_nco_delta()
94 if ((delta < 0) && (abs(delta) > ns)) { in dte_write_nco_delta()
96 ns += DTE_NCO_MAX_NS + delta; in dte_write_nco_delta()
[all …]
/drivers/of/
Dfdt_address.c22 #define OF_CHECK_COUNTS(na, ns) ((na) > 0 && (na) <= OF_MAX_ADDR_CELLS && \ argument
23 (ns) > 0)
43 int na, int ns, int pna);
71 int na, int ns, int pna) in fdt_bus_default_map() argument
76 s = of_read_number(range + na + pna, ns); in fdt_bus_default_map()
112 int na, int ns, int pna, const char *rprop) in fdt_translate_one() argument
133 rone = na + pna + ns; in fdt_translate_one()
135 offset = bus->map(addr, ranges, na, ns, pna); in fdt_translate_one()
169 int na, ns, pna, pns; in fdt_translate_address() local
189 bus->count_cells(blob, parent, &na, &ns); in fdt_translate_address()
[all …]

12345678910>>...12