/drivers/mtd/nand/raw/ |
D | nandsim.c | 189 #define NS_IS_INITIALIZED(ns) ((ns)->geom.totsz != 0) argument 192 #define NS_STATUS_OK(ns) (NAND_STATUS_READY | (NAND_STATUS_WP * ((ns)->lines.wp == 0))) argument 195 #define NS_STATUS_FAILED(ns) (NAND_STATUS_FAIL | NS_STATUS_OK(ns)) argument 198 #define NS_RAW_OFFSET(ns) \ argument 199 (((ns)->regs.row * (ns)->geom.pgszoob) + (ns)->regs.column) 202 #define NS_RAW_OFFSET_OOB(ns) (NS_RAW_OFFSET(ns) + ns->geom.pgsz) argument 205 #define NS_PAGE_BYTE_SHIFT(ns) ((ns)->regs.column + (ns)->regs.off) argument 491 DEFINE_SHOW_ATTRIBUTE(ns); 500 static int ns_debugfs_create(struct nandsim *ns) in ns_debugfs_create() argument 515 ns->dent = debugfs_create_file("nandsim_wear_report", 0400, root, ns, in ns_debugfs_create() [all …]
|
/drivers/net/netdevsim/ |
D | ethtool.c | 14 struct netdevsim *ns = netdev_priv(dev); in nsim_get_pause_stats() local 16 if (ns->ethtool.pauseparam.report_stats_rx) in nsim_get_pause_stats() 18 if (ns->ethtool.pauseparam.report_stats_tx) in nsim_get_pause_stats() 25 struct netdevsim *ns = netdev_priv(dev); in nsim_get_pauseparam() local 28 pause->rx_pause = ns->ethtool.pauseparam.rx; in nsim_get_pauseparam() 29 pause->tx_pause = ns->ethtool.pauseparam.tx; in nsim_get_pauseparam() 35 struct netdevsim *ns = netdev_priv(dev); in nsim_set_pauseparam() local 40 ns->ethtool.pauseparam.rx = pause->rx_pause; in nsim_set_pauseparam() 41 ns->ethtool.pauseparam.tx = pause->tx_pause; in nsim_set_pauseparam() 50 struct netdevsim *ns = netdev_priv(dev); in nsim_get_coalesce() local [all …]
|
D | netdev.c | 31 struct netdevsim *ns = netdev_priv(dev); in nsim_start_xmit() local 33 if (!nsim_ipsec_tx(ns, skb)) in nsim_start_xmit() 36 u64_stats_update_begin(&ns->syncp); in nsim_start_xmit() 37 ns->tx_packets++; in nsim_start_xmit() 38 ns->tx_bytes += skb->len; in nsim_start_xmit() 39 u64_stats_update_end(&ns->syncp); in nsim_start_xmit() 53 struct netdevsim *ns = netdev_priv(dev); in nsim_change_mtu() local 55 if (ns->xdp.prog && new_mtu > NSIM_XDP_MAX_MTU) in nsim_change_mtu() 66 struct netdevsim *ns = netdev_priv(dev); in nsim_get_stats64() local 70 start = u64_stats_fetch_begin(&ns->syncp); in nsim_get_stats64() [all …]
|
D | udp_tunnels.c | 15 struct netdevsim *ns = netdev_priv(dev); in nsim_udp_tunnel_set_port() local 18 ret = -ns->udp_ports.inject_error; in nsim_udp_tunnel_set_port() 19 ns->udp_ports.inject_error = 0; in nsim_udp_tunnel_set_port() 21 if (ns->udp_ports.sleep) in nsim_udp_tunnel_set_port() 22 msleep(ns->udp_ports.sleep); in nsim_udp_tunnel_set_port() 25 if (ns->udp_ports.ports[table][entry]) { in nsim_udp_tunnel_set_port() 29 ns->udp_ports.ports[table][entry] = in nsim_udp_tunnel_set_port() 44 struct netdevsim *ns = netdev_priv(dev); in nsim_udp_tunnel_unset_port() local 47 ret = -ns->udp_ports.inject_error; in nsim_udp_tunnel_unset_port() 48 ns->udp_ports.inject_error = 0; in nsim_udp_tunnel_unset_port() [all …]
|
D | bpf.c | 41 struct netdevsim *ns; member 87 static bool nsim_xdp_offload_active(struct netdevsim *ns) in nsim_xdp_offload_active() argument 89 return ns->xdp_hw.prog; in nsim_xdp_offload_active() 104 nsim_bpf_offload(struct netdevsim *ns, struct bpf_prog *prog, bool oldprog) in nsim_bpf_offload() argument 106 nsim_prog_set_loaded(ns->bpf_offloaded, false); in nsim_bpf_offload() 108 WARN(!!ns->bpf_offloaded != oldprog, in nsim_bpf_offload() 111 ns->bpf_offloaded = prog; in nsim_bpf_offload() 112 ns->bpf_offloaded_id = prog ? prog->aux->id : 0; in nsim_bpf_offload() 123 struct netdevsim *ns = cb_priv; in nsim_bpf_setup_tc_block_cb() local 132 if (!tc_cls_can_offload_and_chain0(ns->netdev, &cls_bpf->common)) in nsim_bpf_setup_tc_block_cb() [all …]
|
D | macsec.c | 11 static int nsim_macsec_find_secy(struct netdevsim *ns, sci_t sci) in nsim_macsec_find_secy() argument 16 if (ns->macsec.nsim_secy[i].sci == sci) in nsim_macsec_find_secy() 37 struct netdevsim *ns = netdev_priv(ctx->netdev); in nsim_macsec_add_secy() local 40 if (ns->macsec.nsim_secy_count == NSIM_MACSEC_MAX_SECY_COUNT) in nsim_macsec_add_secy() 44 if (!ns->macsec.nsim_secy[idx].used) in nsim_macsec_add_secy() 56 ns->macsec.nsim_secy[idx].used = true; in nsim_macsec_add_secy() 57 ns->macsec.nsim_secy[idx].nsim_rxsc_count = 0; in nsim_macsec_add_secy() 58 ns->macsec.nsim_secy[idx].sci = ctx->secy->sci; in nsim_macsec_add_secy() 59 ns->macsec.nsim_secy_count++; in nsim_macsec_add_secy() 66 struct netdevsim *ns = netdev_priv(ctx->netdev); in nsim_macsec_upd_secy() local [all …]
|
D | ipsec.c | 16 struct netdevsim *ns = filp->private_data; in nsim_dbg_netdev_ops_read() local 17 struct nsim_ipsec *ipsec = &ns->ipsec; in nsim_dbg_netdev_ops_read() 133 struct netdevsim *ns; in nsim_ipsec_add_sa() local 139 ns = netdev_priv(dev); in nsim_ipsec_add_sa() 140 ipsec = &ns->ipsec; in nsim_ipsec_add_sa() 202 struct netdevsim *ns = netdev_priv(xs->xso.real_dev); in nsim_ipsec_del_sa() local 203 struct nsim_ipsec *ipsec = &ns->ipsec; in nsim_ipsec_del_sa() 208 netdev_err(ns->netdev, "Invalid SA for delete sa_idx=%d\n", in nsim_ipsec_del_sa() 219 struct netdevsim *ns = netdev_priv(xs->xso.real_dev); in nsim_ipsec_offload_ok() local 220 struct nsim_ipsec *ipsec = &ns->ipsec; in nsim_ipsec_offload_ok() [all …]
|
D | netdevsim.h | 132 void nsim_destroy(struct netdevsim *ns); 134 void nsim_ethtool_init(struct netdevsim *ns); 144 int nsim_bpf_init(struct netdevsim *ns); 145 void nsim_bpf_uninit(struct netdevsim *ns); 147 int nsim_bpf_disable_tc(struct netdevsim *ns); 160 static inline int nsim_bpf_init(struct netdevsim *ns) in nsim_bpf_init() argument 165 static inline void nsim_bpf_uninit(struct netdevsim *ns) in nsim_bpf_uninit() argument 174 static inline int nsim_bpf_disable_tc(struct netdevsim *ns) in nsim_bpf_disable_tc() argument 262 struct netdevsim *ns; member 374 void nsim_ipsec_init(struct netdevsim *ns); [all …]
|
/drivers/block/drbd/ |
D | drbd_state.c | 26 union drbd_state ns; member 281 union drbd_state ns, enum chg_state_flags flags, 285 static enum drbd_state_rv is_valid_transition(union drbd_state os, union drbd_state ns); 287 union drbd_state ns, enum sanitize_state_warnings *warn); 468 union drbd_state os, union drbd_state ns) in cl_wide_st_chg() argument 470 return (os.conn >= C_CONNECTED && ns.conn >= C_CONNECTED && in cl_wide_st_chg() 471 ((os.role != R_PRIMARY && ns.role == R_PRIMARY) || in cl_wide_st_chg() 472 (os.conn != C_STARTING_SYNC_T && ns.conn == C_STARTING_SYNC_T) || in cl_wide_st_chg() 473 (os.conn != C_STARTING_SYNC_S && ns.conn == C_STARTING_SYNC_S) || in cl_wide_st_chg() 474 (os.disk != D_FAILED && ns.disk == D_FAILED))) || in cl_wide_st_chg() [all …]
|
/drivers/nvme/host/ |
D | multipath.c | 85 struct nvme_ns *ns = req->q->queuedata; in nvme_failover_req() local 90 nvme_mpath_clear_current_path(ns); in nvme_failover_req() 97 if (nvme_is_ana_error(status) && ns->ctrl->ana_log_buf) { in nvme_failover_req() 98 set_bit(NVME_NS_ANA_PENDING, &ns->flags); in nvme_failover_req() 99 queue_work(nvme_wq, &ns->ctrl->ana_work); in nvme_failover_req() 102 spin_lock_irqsave(&ns->head->requeue_lock, flags); in nvme_failover_req() 104 bio_set_dev(bio, ns->head->disk->part0); in nvme_failover_req() 118 blk_steal_bios(&ns->head->requeue_list, req); in nvme_failover_req() 119 spin_unlock_irqrestore(&ns->head->requeue_lock, flags); in nvme_failover_req() 123 kblockd_schedule_work(&ns->head->requeue_work); in nvme_failover_req() [all …]
|
D | zns.c | 10 int nvme_revalidate_zones(struct nvme_ns *ns) in nvme_revalidate_zones() argument 12 struct request_queue *q = ns->queue; in nvme_revalidate_zones() 14 blk_queue_chunk_sectors(q, ns->zsze); in nvme_revalidate_zones() 15 blk_queue_max_zone_append_sectors(q, ns->ctrl->max_zone_append); in nvme_revalidate_zones() 17 return blk_revalidate_disk_zones(ns->disk, NULL); in nvme_revalidate_zones() 48 int nvme_update_zone_info(struct nvme_ns *ns, unsigned lbaf) in nvme_update_zone_info() argument 50 struct nvme_effects_log *log = ns->head->effects; in nvme_update_zone_info() 51 struct request_queue *q = ns->queue; in nvme_update_zone_info() 59 if (test_and_clear_bit(NVME_NS_FORCE_RO, &ns->flags)) in nvme_update_zone_info() 60 dev_warn(ns->ctrl->device, in nvme_update_zone_info() [all …]
|
D | core.c | 311 struct nvme_ns *ns = req->q->queuedata; in nvme_log_error() local 314 if (ns) { in nvme_log_error() 316 ns->disk ? ns->disk->disk_name : "?", in nvme_log_error() 319 (unsigned long long)nvme_sect_to_lba(ns, blk_rq_pos(req)), in nvme_log_error() 320 (unsigned long long)blk_rq_bytes(req) >> ns->lba_shift, in nvme_log_error() 627 struct nvme_ns *ns = container_of(kref, struct nvme_ns, kref); in nvme_free_ns() local 629 put_disk(ns->disk); in nvme_free_ns() 630 nvme_put_ns_head(ns->head); in nvme_free_ns() 631 nvme_put_ctrl(ns->ctrl); in nvme_free_ns() 632 kfree(ns); in nvme_free_ns() [all …]
|
D | ioctl.c | 16 static bool nvme_cmd_allowed(struct nvme_ns *ns, struct nvme_command *c, in nvme_cmd_allowed() argument 42 if (!ns) { in nvme_cmd_allowed() 61 effects = nvme_command_effects(ns->ctrl, ns, c->common.opcode); in nvme_cmd_allowed() 172 struct nvme_ns *ns = q->queuedata; in nvme_map_user_request() local 173 struct block_device *bdev = ns ? ns->disk->part0 : NULL; in nvme_map_user_request() 226 struct nvme_ns *ns = q->queuedata; in nvme_submit_user_cmd() local 249 effects = nvme_passthru_start(ctrl, ns, cmd->common.opcode); in nvme_submit_user_cmd() 261 nvme_passthru_end(ctrl, ns, effects, cmd, ret); in nvme_submit_user_cmd() 266 static int nvme_submit_io(struct nvme_ns *ns, struct nvme_user_io __user *uio) in nvme_submit_io() argument 287 length = (io.nblocks + 1) << ns->lba_shift; in nvme_submit_io() [all …]
|
/drivers/nvme/target/ |
D | io-cmd-bdev.c | 51 void nvmet_bdev_ns_disable(struct nvmet_ns *ns) in nvmet_bdev_ns_disable() argument 53 if (ns->bdev) { in nvmet_bdev_ns_disable() 54 blkdev_put(ns->bdev, NULL); in nvmet_bdev_ns_disable() 55 ns->bdev = NULL; in nvmet_bdev_ns_disable() 59 static void nvmet_bdev_ns_enable_integrity(struct nvmet_ns *ns) in nvmet_bdev_ns_enable_integrity() argument 61 struct blk_integrity *bi = bdev_get_integrity(ns->bdev); in nvmet_bdev_ns_enable_integrity() 64 ns->metadata_size = bi->tuple_size; in nvmet_bdev_ns_enable_integrity() 66 ns->pi_type = NVME_NS_DPS_PI_TYPE1; in nvmet_bdev_ns_enable_integrity() 68 ns->pi_type = NVME_NS_DPS_PI_TYPE3; in nvmet_bdev_ns_enable_integrity() 71 ns->metadata_size = 0; in nvmet_bdev_ns_enable_integrity() [all …]
|
D | io-cmd-file.c | 16 void nvmet_file_ns_revalidate(struct nvmet_ns *ns) in nvmet_file_ns_revalidate() argument 18 ns->size = i_size_read(ns->file->f_mapping->host); in nvmet_file_ns_revalidate() 21 void nvmet_file_ns_disable(struct nvmet_ns *ns) in nvmet_file_ns_disable() argument 23 if (ns->file) { in nvmet_file_ns_disable() 24 if (ns->buffered_io) in nvmet_file_ns_disable() 26 mempool_destroy(ns->bvec_pool); in nvmet_file_ns_disable() 27 ns->bvec_pool = NULL; in nvmet_file_ns_disable() 28 fput(ns->file); in nvmet_file_ns_disable() 29 ns->file = NULL; in nvmet_file_ns_disable() 33 int nvmet_file_ns_enable(struct nvmet_ns *ns) in nvmet_file_ns_enable() argument [all …]
|
D | core.c | 430 req->ns = xa_load(&subsys->namespaces, nsid); in nvmet_req_find_ns() 431 if (unlikely(!req->ns)) { in nvmet_req_find_ns() 438 percpu_ref_get(&req->ns->ref); in nvmet_req_find_ns() 444 struct nvmet_ns *ns = container_of(ref, struct nvmet_ns, ref); in nvmet_destroy_namespace() local 446 complete(&ns->disable_done); in nvmet_destroy_namespace() 449 void nvmet_put_namespace(struct nvmet_ns *ns) in nvmet_put_namespace() argument 451 percpu_ref_put(&ns->ref); in nvmet_put_namespace() 454 static void nvmet_ns_dev_disable(struct nvmet_ns *ns) in nvmet_ns_dev_disable() argument 456 nvmet_bdev_ns_disable(ns); in nvmet_ns_dev_disable() 457 nvmet_file_ns_disable(ns); in nvmet_ns_dev_disable() [all …]
|
D | zns.c | 35 bool nvmet_bdev_zns_enable(struct nvmet_ns *ns) in nvmet_bdev_zns_enable() argument 37 u8 zasl = nvmet_zasl(bdev_max_zone_append_sectors(ns->bdev)); in nvmet_bdev_zns_enable() 38 struct gendisk *bd_disk = ns->bdev->bd_disk; in nvmet_bdev_zns_enable() 41 if (ns->subsys->zasl) { in nvmet_bdev_zns_enable() 42 if (ns->subsys->zasl > zasl) in nvmet_bdev_zns_enable() 45 ns->subsys->zasl = zasl; in nvmet_bdev_zns_enable() 60 if (ns->bdev->bd_disk->conv_zones_bitmap) in nvmet_bdev_zns_enable() 63 ret = blkdev_report_zones(ns->bdev, 0, bdev_nr_zones(ns->bdev), in nvmet_bdev_zns_enable() 68 ns->blksize_shift = blksize_bits(bdev_logical_block_size(ns->bdev)); in nvmet_bdev_zns_enable() 121 if (nvmet_ns_revalidate(req->ns)) { in nvmet_execute_identify_ns_zns() [all …]
|
D | admin-cmd.c | 85 if (!req->ns->bdev) in nvmet_get_smart_log_nsid() 88 host_reads = part_stat_read(req->ns->bdev, ios[READ]); in nvmet_get_smart_log_nsid() 90 DIV_ROUND_UP(part_stat_read(req->ns->bdev, sectors[READ]), 1000); in nvmet_get_smart_log_nsid() 91 host_writes = part_stat_read(req->ns->bdev, ios[WRITE]); in nvmet_get_smart_log_nsid() 93 DIV_ROUND_UP(part_stat_read(req->ns->bdev, sectors[WRITE]), 1000); in nvmet_get_smart_log_nsid() 108 struct nvmet_ns *ns; in nvmet_get_smart_log_all() local 113 xa_for_each(&ctrl->subsys->namespaces, idx, ns) { in nvmet_get_smart_log_all() 115 if (!ns->bdev) in nvmet_get_smart_log_all() 117 host_reads += part_stat_read(ns->bdev, ios[READ]); in nvmet_get_smart_log_all() 119 part_stat_read(ns->bdev, sectors[READ]), 1000); in nvmet_get_smart_log_all() [all …]
|
/drivers/clk/qcom/ |
D | clk-rcg.c | 18 static u32 ns_to_src(struct src_sel *s, u32 ns) in ns_to_src() argument 20 ns >>= s->src_sel_shift; in ns_to_src() 21 ns &= SRC_SEL_MASK; in ns_to_src() 22 return ns; in ns_to_src() 25 static u32 src_to_ns(struct src_sel *s, u8 src, u32 ns) in src_to_ns() argument 31 ns &= ~mask; in src_to_ns() 33 ns |= src << s->src_sel_shift; in src_to_ns() 34 return ns; in src_to_ns() 41 u32 ns; in clk_rcg_get_parent() local 44 ret = regmap_read(rcg->clkr.regmap, rcg->ns_reg, &ns); in clk_rcg_get_parent() [all …]
|
/drivers/net/ethernet/mellanox/mlx5/core/ |
D | fs_cmd.h | 39 int (*create_flow_table)(struct mlx5_flow_root_namespace *ns, 43 int (*destroy_flow_table)(struct mlx5_flow_root_namespace *ns, 46 int (*modify_flow_table)(struct mlx5_flow_root_namespace *ns, 50 int (*create_flow_group)(struct mlx5_flow_root_namespace *ns, 55 int (*destroy_flow_group)(struct mlx5_flow_root_namespace *ns, 59 int (*create_fte)(struct mlx5_flow_root_namespace *ns, 64 int (*update_fte)(struct mlx5_flow_root_namespace *ns, 70 int (*delete_fte)(struct mlx5_flow_root_namespace *ns, 74 int (*update_root_ft)(struct mlx5_flow_root_namespace *ns, 79 int (*packet_reformat_alloc)(struct mlx5_flow_root_namespace *ns, [all …]
|
D | fs_cmd.c | 43 static int mlx5_cmd_stub_update_root_ft(struct mlx5_flow_root_namespace *ns, in mlx5_cmd_stub_update_root_ft() argument 51 static int mlx5_cmd_stub_create_flow_table(struct mlx5_flow_root_namespace *ns, in mlx5_cmd_stub_create_flow_table() argument 63 static int mlx5_cmd_stub_destroy_flow_table(struct mlx5_flow_root_namespace *ns, in mlx5_cmd_stub_destroy_flow_table() argument 69 static int mlx5_cmd_stub_modify_flow_table(struct mlx5_flow_root_namespace *ns, in mlx5_cmd_stub_modify_flow_table() argument 76 static int mlx5_cmd_stub_create_flow_group(struct mlx5_flow_root_namespace *ns, in mlx5_cmd_stub_create_flow_group() argument 84 static int mlx5_cmd_stub_destroy_flow_group(struct mlx5_flow_root_namespace *ns, in mlx5_cmd_stub_destroy_flow_group() argument 91 static int mlx5_cmd_stub_create_fte(struct mlx5_flow_root_namespace *ns, in mlx5_cmd_stub_create_fte() argument 99 static int mlx5_cmd_stub_update_fte(struct mlx5_flow_root_namespace *ns, in mlx5_cmd_stub_update_fte() argument 108 static int mlx5_cmd_stub_delete_fte(struct mlx5_flow_root_namespace *ns, in mlx5_cmd_stub_delete_fte() argument 115 static int mlx5_cmd_stub_packet_reformat_alloc(struct mlx5_flow_root_namespace *ns, in mlx5_cmd_stub_packet_reformat_alloc() argument [all …]
|
/drivers/scsi/bfa/ |
D | bfa_fcs_lport.c | 3666 static void bfa_fcs_lport_ns_sm_offline(struct bfa_fcs_lport_ns_s *ns, 3668 static void bfa_fcs_lport_ns_sm_plogi_sending(struct bfa_fcs_lport_ns_s *ns, 3670 static void bfa_fcs_lport_ns_sm_plogi(struct bfa_fcs_lport_ns_s *ns, 3672 static void bfa_fcs_lport_ns_sm_plogi_retry(struct bfa_fcs_lport_ns_s *ns, 3675 struct bfa_fcs_lport_ns_s *ns, 3677 static void bfa_fcs_lport_ns_sm_rspn_id(struct bfa_fcs_lport_ns_s *ns, 3679 static void bfa_fcs_lport_ns_sm_rspn_id_retry(struct bfa_fcs_lport_ns_s *ns, 3682 struct bfa_fcs_lport_ns_s *ns, 3684 static void bfa_fcs_lport_ns_sm_rft_id_retry(struct bfa_fcs_lport_ns_s *ns, 3686 static void bfa_fcs_lport_ns_sm_rft_id(struct bfa_fcs_lport_ns_s *ns, [all …]
|
/drivers/net/ethernet/mellanox/mlx5/core/steering/ |
D | fs_dr.c | 20 static int mlx5_cmd_dr_update_root_ft(struct mlx5_flow_root_namespace *ns, in mlx5_cmd_dr_update_root_ft() argument 25 return mlx5_fs_cmd_get_fw_cmds()->update_root_ft(ns, ft, underlay_qpn, in mlx5_cmd_dr_update_root_ft() 29 static int set_miss_action(struct mlx5_flow_root_namespace *ns, in set_miss_action() argument 49 mlx5_core_err(ns->dev, in set_miss_action() 57 mlx5_core_err(ns->dev, "Failed to destroy action (%d)\n", in set_miss_action() 64 static int mlx5_cmd_dr_create_flow_table(struct mlx5_flow_root_namespace *ns, in mlx5_cmd_dr_create_flow_table() argument 74 return mlx5_fs_cmd_get_fw_cmds()->create_flow_table(ns, ft, in mlx5_cmd_dr_create_flow_table() 79 if (!MLX5_CAP_FLOWTABLE(ns->dev, sw_owner_reformat_supported)) in mlx5_cmd_dr_create_flow_table() 83 tbl = mlx5dr_table_create(ns->fs_dr_domain.dr_domain, ft->level, flags, in mlx5_cmd_dr_create_flow_table() 86 mlx5_core_err(ns->dev, "Failed creating dr flow_table\n"); in mlx5_cmd_dr_create_flow_table() [all …]
|
/drivers/ptp/ |
D | ptp_dte.c | 52 static void dte_write_nco(void __iomem *regs, s64 ns) in dte_write_nco() argument 56 sum2 = (u32)((ns >> DTE_NCO_SUM2_SHIFT) & DTE_NCO_SUM2_MASK); in dte_write_nco() 62 sum3 = (u32)(((ns >> DTE_NCO_SUM3_SHIFT) & DTE_NCO_SUM3_MASK) << in dte_write_nco() 73 s64 ns; in dte_read_nco() local 81 ns = ((s64)sum3 << DTE_NCO_SUM3_SHIFT) | in dte_read_nco() 84 return ns; in dte_read_nco() 89 s64 ns; in dte_write_nco_delta() local 91 ns = dte_read_nco(ptp_dte->regs); in dte_write_nco_delta() 94 if ((delta < 0) && (abs(delta) > ns)) { in dte_write_nco_delta() 96 ns += DTE_NCO_MAX_NS + delta; in dte_write_nco_delta() [all …]
|
/drivers/of/ |
D | fdt_address.c | 22 #define OF_CHECK_COUNTS(na, ns) ((na) > 0 && (na) <= OF_MAX_ADDR_CELLS && \ argument 23 (ns) > 0) 43 int na, int ns, int pna); 71 int na, int ns, int pna) in fdt_bus_default_map() argument 76 s = of_read_number(range + na + pna, ns); in fdt_bus_default_map() 112 int na, int ns, int pna, const char *rprop) in fdt_translate_one() argument 133 rone = na + pna + ns; in fdt_translate_one() 135 offset = bus->map(addr, ranges, na, ns, pna); in fdt_translate_one() 169 int na, ns, pna, pns; in fdt_translate_address() local 189 bus->count_cells(blob, parent, &na, &ns); in fdt_translate_address() [all …]
|