/drivers/net/ethernet/amd/xgbe/ |
D | xgbe-ptp.c | 130 u64 nsec; in xgbe_cc_read() local 132 nsec = pdata->hw_if.get_tstamp_time(pdata); in xgbe_cc_read() 134 return nsec; in xgbe_cc_read() 188 u64 nsec; in xgbe_gettime() local 192 nsec = timecounter_read(&pdata->tstamp_tc); in xgbe_gettime() 196 *ts = ns_to_timespec64(nsec); in xgbe_gettime() 208 u64 nsec; in xgbe_settime() local 210 nsec = timespec64_to_ns(ts); in xgbe_settime() 214 timecounter_init(&pdata->tstamp_tc, &pdata->tstamp_cc, nsec); in xgbe_settime()
|
D | xgbe-dev.c | 1541 unsigned int nsec) in xgbe_set_tstamp_time() argument 1547 XGMAC_IOWRITE(pdata, MAC_STNUR, nsec); in xgbe_set_tstamp_time() 1560 u64 nsec; in xgbe_get_tstamp_time() local 1562 nsec = XGMAC_IOREAD(pdata, MAC_STSR); in xgbe_get_tstamp_time() 1563 nsec *= NSEC_PER_SEC; in xgbe_get_tstamp_time() 1564 nsec += XGMAC_IOREAD(pdata, MAC_STNR); in xgbe_get_tstamp_time() 1566 return nsec; in xgbe_get_tstamp_time() 1572 u64 nsec; in xgbe_get_tx_tstamp() local 1585 nsec = tx_ssr; in xgbe_get_tx_tstamp() 1586 nsec *= NSEC_PER_SEC; in xgbe_get_tx_tstamp() [all …]
|
/drivers/net/ethernet/stmicro/stmmac/ |
D | stmmac_hwtstamp.c | 63 static int stmmac_init_systime(void __iomem *ioaddr, u32 sec, u32 nsec) in stmmac_init_systime() argument 69 writel(nsec, ioaddr + PTP_STNSUR); in stmmac_init_systime() 112 static int stmmac_adjust_systime(void __iomem *ioaddr, u32 sec, u32 nsec, in stmmac_adjust_systime() argument 128 nsec = (PTP_DIGITAL_ROLLOVER_MODE - nsec); in stmmac_adjust_systime() 130 nsec = (PTP_BINARY_ROLLOVER_MODE - nsec); in stmmac_adjust_systime() 134 value = (add_sub << PTP_STNSUR_ADDSUB_SHIFT) | nsec; in stmmac_adjust_systime()
|
D | stmmac_ptp.c | 73 u32 sec, nsec; in stmmac_adjust_time() local 84 nsec = reminder; in stmmac_adjust_time() 88 priv->hw->ptp->adjust_systime(priv->ptpaddr, sec, nsec, neg_adj, in stmmac_adjust_time()
|
/drivers/pps/ |
D | kapi.c | 43 ts->nsec += offset->nsec; in pps_add_offset() 44 while (ts->nsec >= NSEC_PER_SEC) { in pps_add_offset() 45 ts->nsec -= NSEC_PER_SEC; in pps_add_offset() 48 while (ts->nsec < 0) { in pps_add_offset() 49 ts->nsec += NSEC_PER_SEC; in pps_add_offset() 177 struct pps_ktime ts_real = { .sec = 0, .nsec = 0, .flags = 0 }; in pps_event()
|
D | sysfs.c | 41 (long long) pps->assert_tu.sec, pps->assert_tu.nsec, in assert_show() 55 (long long) pps->clear_tu.sec, pps->clear_tu.nsec, in clear_show()
|
D | pps.c | 81 fdata->timeout.nsec); in pps_cdev_pps_fetch() 83 ticks += fdata->timeout.nsec / (NSEC_PER_SEC / HZ); in pps_cdev_pps_fetch()
|
/drivers/gpu/drm/nouveau/nvkm/subdev/pmu/ |
D | memx.c | 108 u32 addr, u32 mask, u32 data, u32 nsec) in nvkm_memx_wait() argument 111 addr, mask, data, nsec); in nvkm_memx_wait() 112 memx_cmd(memx, MEMX_WAIT, 4, (u32[]){ addr, mask, data, nsec }); in nvkm_memx_wait() 117 nvkm_memx_nsec(struct nvkm_memx *memx, u32 nsec) in nvkm_memx_nsec() argument 119 nvkm_debug(&memx->pmu->subdev, " DELAY = %d ns\n", nsec); in nvkm_memx_nsec() 120 memx_cmd(memx, MEMX_DELAY, 1, (u32[]){ nsec }); in nvkm_memx_nsec()
|
/drivers/misc/sgi-gru/ |
D | gruhandles.c | 44 unsigned long nsec; in update_mcs_stats() local 46 nsec = CLKS2NSEC(clks); in update_mcs_stats() 48 atomic_long_add(nsec, &mcs_op_statistics[op].total); in update_mcs_stats() 49 if (mcs_op_statistics[op].max < nsec) in update_mcs_stats() 50 mcs_op_statistics[op].max = nsec; in update_mcs_stats()
|
/drivers/ptp/ |
D | ptp_chardev.c | 178 enable = req.perout.period.sec || req.perout.period.nsec; in ptp_ioctl() 202 precise_offset.device.nsec = ts.tv_nsec; in ptp_ioctl() 205 precise_offset.sys_realtime.nsec = ts.tv_nsec; in ptp_ioctl() 208 precise_offset.sys_monoraw.nsec = ts.tv_nsec; in ptp_ioctl() 229 pct->nsec = ts.tv_nsec; in ptp_ioctl() 235 pct->nsec = ts.tv_nsec; in ptp_ioctl() 240 pct->nsec = ts.tv_nsec; in ptp_ioctl()
|
D | ptp_sysfs.c | 102 event.index, event.t.sec, event.t.nsec); in extts_fifo_show() 119 &req.perout.start.sec, &req.perout.start.nsec, in period_store() 120 &req.perout.period.sec, &req.perout.period.nsec); in period_store() 126 enable = req.perout.period.sec || req.perout.period.nsec; in period_store()
|
D | ptp_kvm.c | 79 tspec.tv_nsec = clock_pair.nsec; in ptp_kvm_get_time_fn() 139 tspec.tv_nsec = clock_pair.nsec; in ptp_kvm_gettime()
|
/drivers/gpu/drm/nouveau/nvkm/subdev/fb/ |
D | ramfuc.h | 123 ramfuc_wait(struct ramfuc *ram, u32 addr, u32 mask, u32 data, u32 nsec) in ramfuc_wait() argument 125 nvkm_memx_wait(ram->memx, addr, mask, data, nsec); in ramfuc_wait() 129 ramfuc_nsec(struct ramfuc *ram, u32 nsec) in ramfuc_nsec() argument 131 nvkm_memx_nsec(ram->memx, nsec); in ramfuc_nsec()
|
/drivers/gpu/drm/nouveau/nvkm/subdev/timer/ |
D | base.c | 70 nvkm_timer_alarm(struct nvkm_timer *tmr, u32 nsec, struct nvkm_alarm *alarm) in nvkm_timer_alarm() argument 83 if (nsec) { in nvkm_timer_alarm() 85 alarm->timestamp = nvkm_timer_read(tmr) + nsec; in nvkm_timer_alarm()
|
/drivers/gpu/drm/nouveau/nvkm/subdev/bus/ |
D | hwsq.c | 167 nvkm_hwsq_nsec(struct nvkm_hwsq *hwsq, u32 nsec) in nvkm_hwsq_nsec() argument 169 u8 shift = 0, usec = nsec / 1000; in nvkm_hwsq_nsec() 175 nvkm_debug(hwsq->subdev, " DELAY = %d ns\n", nsec); in nvkm_hwsq_nsec()
|
D | hwsq.h | 144 hwsq_nsec(struct hwsq *ram, u32 nsec) in hwsq_nsec() argument 146 nvkm_hwsq_nsec(ram->hwsq, nsec); in hwsq_nsec()
|
/drivers/gpu/drm/nouveau/include/nvkm/subdev/ |
D | pmu.h | 51 void nvkm_memx_wait(struct nvkm_memx *, u32 addr, u32 mask, u32 data, u32 nsec); 52 void nvkm_memx_nsec(struct nvkm_memx *, u32 nsec);
|
D | bus.h | 19 void nvkm_hwsq_nsec(struct nvkm_hwsq *, u32 nsec);
|
D | timer.h | 29 void nvkm_timer_alarm(struct nvkm_timer *, u32 nsec, struct nvkm_alarm *);
|
/drivers/clocksource/ |
D | jcore-pit.c | 47 u32 seclo, nsec, seclo0; in jcore_sched_clock_read() local 53 nsec = readl(base + REG_NSEC); in jcore_sched_clock_read() 57 return seclo * NSEC_PER_SEC + nsec; in jcore_sched_clock_read()
|
/drivers/net/ethernet/mellanox/mlx4/ |
D | en_clock.c | 66 u64 nsec; in mlx4_en_fill_hwtstamps() local 70 nsec = timecounter_cyc2time(&mdev->clock, timestamp); in mlx4_en_fill_hwtstamps() 74 hwts->hwtstamp = ns_to_ktime(nsec); in mlx4_en_fill_hwtstamps()
|
/drivers/gpu/drm/nouveau/nvkm/subdev/i2c/ |
D | bit.c | 56 nvkm_i2c_delay(struct nvkm_i2c_bus *bus, u32 nsec) in nvkm_i2c_delay() argument 58 udelay((nsec + 500) / 1000); in nvkm_i2c_delay()
|
/drivers/net/ethernet/mellanox/mlx5/core/ |
D | en_clock.c | 68 u64 nsec; in mlx5e_fill_hwstamp() local 71 nsec = timecounter_cyc2time(&tstamp->clock, timestamp); in mlx5e_fill_hwstamp() 74 hwts->hwtstamp = ns_to_ktime(nsec); in mlx5e_fill_hwstamp() 357 ts.tv_nsec = rq->perout.period.nsec; in mlx5e_perout_configure() 364 ts.tv_nsec = rq->perout.start.nsec; in mlx5e_perout_configure()
|
/drivers/net/ethernet/intel/igb/ |
D | igb_ptp.c | 129 u32 sec, nsec; in igb_ptp_read_i210() local 136 nsec = rd32(E1000_SYSTIML); in igb_ptp_read_i210() 140 ts->tv_nsec = nsec; in igb_ptp_read_i210() 531 ts.tv_nsec = rq->perout.period.nsec; in igb_ptp_feature_enable_i210() 578 igb->perout[i].start.tv_nsec = rq->perout.start.nsec; in igb_ptp_feature_enable_i210() 582 wr32(trgttiml, rq->perout.start.nsec); in igb_ptp_feature_enable_i210()
|
/drivers/block/xen-blkback/ |
D | blkback.c | 989 seg[n].nsec = last_sect - first_sect + 1; in xen_blkbk_parse_indirect() 991 preq->nr_sects += seg[n].nsec; in xen_blkbk_parse_indirect() 1295 seg[i].nsec = req->u.rw.seg[i].last_sect - in dispatch_rw_block_io() 1302 preq.nr_sects += seg[i].nsec; in dispatch_rw_block_io() 1325 if (((int)preq.sector_number|(int)seg[i].nsec) & in dispatch_rw_block_io() 1359 seg[i].nsec << 9, in dispatch_rw_block_io() 1375 preq.sector_number += seg[i].nsec; in dispatch_rw_block_io()
|