/arch/parisc/math-emu/ |
D | sfcmp.c | 55 register unsigned int left, right; in sgl_fcmp() local 59 left = *leftptr; in sgl_fcmp() 65 if( (Sgl_exponent(left) == SGL_INFINITY_EXPONENT) in sgl_fcmp() 71 if( ( (Sgl_exponent(left) == SGL_INFINITY_EXPONENT) in sgl_fcmp() 72 && Sgl_isnotzero_mantissa(left) in sgl_fcmp() 73 && (Exception(cond) || Sgl_isone_signaling(left))) in sgl_fcmp() 89 else if( ((Sgl_exponent(left) == SGL_INFINITY_EXPONENT) in sgl_fcmp() 90 && Sgl_isnotzero_mantissa(left)) in sgl_fcmp() 103 Sgl_xortointp1(left,right,xorresult); in sgl_fcmp() 108 if( Sgl_iszero_exponentmantissa(left) in sgl_fcmp() [all …]
|
D | sfsub.c | 55 register unsigned int left, right, result, extent; in sgl_fsub() local 63 left = *leftptr; in sgl_fsub() 68 Sgl_xortointp1(left,right,/*to*/save); in sgl_fsub() 73 if ((result_exponent = Sgl_exponent(left)) == SGL_INFINITY_EXPONENT) in sgl_fsub() 75 if (Sgl_iszero_mantissa(left)) in sgl_fsub() 93 *dstptr = left; in sgl_fsub() 102 if (Sgl_isone_signaling(left)) in sgl_fsub() 108 Sgl_set_quiet(left); in sgl_fsub() 126 *dstptr = left; in sgl_fsub() 163 Sgl_copytoint_exponentmantissa(left,signless_upper_left); in sgl_fsub() [all …]
|
D | sfadd.c | 55 register unsigned int left, right, result, extent; in sgl_fadd() local 65 left = *leftptr; in sgl_fadd() 70 Sgl_xortointp1(left,right,/*to*/save); in sgl_fadd() 75 if ((result_exponent = Sgl_exponent(left)) == SGL_INFINITY_EXPONENT) in sgl_fadd() 77 if (Sgl_iszero_mantissa(left)) in sgl_fadd() 95 *dstptr = left; in sgl_fadd() 104 if (Sgl_isone_signaling(left)) in sgl_fadd() 110 Sgl_set_quiet(left); in sgl_fadd() 128 *dstptr = left; in sgl_fadd() 164 Sgl_copytoint_exponentmantissa(left,signless_upper_left); in sgl_fadd() [all …]
|
D | hppa.h | 28 #define Shiftdouble(left,right,amount,dest) \ argument 30 dest = ((left) << (32-(amount))) | ((unsigned int)(right) >> (amount)) 33 #define Variableshiftdouble(left,right,amount,dest) \ argument 36 else dest = ((((unsigned) left)&0x7fffffff) << (32-(amount))) | \ 40 #define Variable_shift_double(left,right,amount,dest) \ argument 42 dest = (left << (32-(amount))) | ((unsigned) right >> (amount))
|
D | sgl_float.h | 63 #define Sgl_leftshiftby1_withextent(left,right,result) \ argument 64 Shiftdouble(Sall(left),Extall(right),31,Sall(result)) 66 #define Sgl_rightshiftby1_withextent(left,right,dst) \ argument 67 Shiftdouble(Sall(left),Extall(right),1,Extall(right)) 239 #define Sgl_subtract_withextension(left,right,extent,result) \ argument 241 Sgl_subtract(left,right,result); \ 248 #define Sgl_xortointp1(left,right,result) \ argument 249 result = Sall(left) XOR Sall(right); 251 #define Sgl_xorfromintp1(left,right,result) \ argument 252 Sall(result) = left XOR Sall(right) [all …]
|
D | dbl_float.h | 421 #define Dbl_xortointp1(left,right,result) \ argument 422 result = Dallp1(left) XOR Dallp1(right) 424 #define Dbl_xorfromintp1(left,right,result) \ argument 425 Dallp1(result) = left XOR Dallp1(right) 427 #define Dbl_swap_lower(left,right) \ argument 428 Dallp2(left) = Dallp2(left) XOR Dallp2(right); \ 429 Dallp2(right) = Dallp2(left) XOR Dallp2(right); \ 430 Dallp2(left) = Dallp2(left) XOR Dallp2(right) 720 #define Dblext_xortointp1(left,right,result) Dbl_xortointp1(left,right,result) argument 722 #define Dblext_xorfromintp1(left,right,result) \ argument [all …]
|
/arch/x86/crypto/ |
D | des3_ede-asm_64.S | 101 #define initial_permutation(left, right) \ argument 102 do_permutation(left##d, right##d, 4, 0x0f0f0f0f); \ 103 do_permutation(left##d, right##d, 16, 0x0000ffff); \ 104 do_permutation(right##d, left##d, 2, 0x33333333); \ 105 do_permutation(right##d, left##d, 8, 0x00ff00ff); \ 107 movl left##d, RW0d; \ 111 xorl RW0d, left##d; \ 113 roll $1, left##d; \ 115 expand_to_64bits(left, RT3); 117 #define final_permutation(left, right) \ argument [all …]
|
/arch/hexagon/include/asm/ |
D | uaccess.h | 100 long left = raw_copy_from_user(dst, src, n); in hexagon_strncpy_from_user() local 101 if (unlikely(left)) in hexagon_strncpy_from_user() 102 memset(dst + (n - left), 0, left); in hexagon_strncpy_from_user() 105 long left = raw_copy_from_user(dst, src, res); in hexagon_strncpy_from_user() local 106 if (unlikely(left)) in hexagon_strncpy_from_user() 107 memset(dst + (res - left), 0, left); in hexagon_strncpy_from_user()
|
/arch/arm/plat-samsung/ |
D | pm-check.c | 110 unsigned long addr, left; in s3c_pm_makecheck() local 114 left = res->end - addr; in s3c_pm_makecheck() 116 if (left > CHECK_CHUNKSIZE) in s3c_pm_makecheck() 117 left = CHECK_CHUNKSIZE; in s3c_pm_makecheck() 119 *val = crc32_le(~0, phys_to_virt(addr), left); in s3c_pm_makecheck() 168 unsigned long left; in s3c_pm_runcheck() local 177 left = res->end - addr; in s3c_pm_runcheck() 179 if (left > CHECK_CHUNKSIZE) in s3c_pm_runcheck() 180 left = CHECK_CHUNKSIZE; in s3c_pm_runcheck() 184 if (in_region(ptr, left, stkpage, 4096)) { in s3c_pm_runcheck() [all …]
|
/arch/s390/include/asm/ |
D | idals.h | 199 size_t left; in idal_buffer_to_user() local 204 left = copy_to_user(to, ib->data[i], IDA_BLOCK_SIZE); in idal_buffer_to_user() 205 if (left) in idal_buffer_to_user() 206 return left + count - IDA_BLOCK_SIZE; in idal_buffer_to_user() 219 size_t left; in idal_buffer_from_user() local 224 left = copy_from_user(ib->data[i], from, IDA_BLOCK_SIZE); in idal_buffer_from_user() 225 if (left) in idal_buffer_from_user() 226 return left + count - IDA_BLOCK_SIZE; in idal_buffer_from_user()
|
/arch/xtensa/kernel/ |
D | perf_event.c | 161 s64 left; in xtensa_perf_event_set_period() local 164 left = XTENSA_PMU_COUNTER_MAX; in xtensa_perf_event_set_period() 168 left = local64_read(&hwc->period_left); in xtensa_perf_event_set_period() 169 if (left <= -period) { in xtensa_perf_event_set_period() 170 left = period; in xtensa_perf_event_set_period() 171 local64_set(&hwc->period_left, left); in xtensa_perf_event_set_period() 174 } else if (left <= 0) { in xtensa_perf_event_set_period() 175 left += period; in xtensa_perf_event_set_period() 176 local64_set(&hwc->period_left, left); in xtensa_perf_event_set_period() 180 if (left > XTENSA_PMU_COUNTER_MAX) in xtensa_perf_event_set_period() [all …]
|
/arch/powerpc/perf/ |
D | core-fsl-emb.c | 327 s64 left = local64_read(&event->hw.period_left); in fsl_emb_pmu_add() local 328 if (left < 0x80000000L) in fsl_emb_pmu_add() 329 val = 0x80000000L - left; in fsl_emb_pmu_add() 395 s64 left; in fsl_emb_pmu_start() local 410 left = local64_read(&event->hw.period_left); in fsl_emb_pmu_start() 412 if (left < 0x80000000L) in fsl_emb_pmu_start() 413 val = 0x80000000L - left; in fsl_emb_pmu_start() 610 s64 prev, delta, left; in record_and_restart() local 628 left = local64_read(&event->hw.period_left) - delta; in record_and_restart() 630 if (left <= 0) { in record_and_restart() [all …]
|
D | core-book3s.c | 1285 s64 left; in power_pmu_enable() local 1399 left = local64_read(&event->hw.period_left); in power_pmu_enable() 1400 if (left < 0x80000000L) in power_pmu_enable() 1401 val = 0x80000000L - left; in power_pmu_enable() 1608 s64 left; in power_pmu_start() local 1624 left = local64_read(&event->hw.period_left); in power_pmu_start() 1627 if (left < 0x80000000L) in power_pmu_start() 1628 val = 0x80000000L - left; in power_pmu_start() 2032 s64 prev, delta, left; in record_and_restart() local 2050 left = local64_read(&event->hw.period_left) - delta; in record_and_restart() [all …]
|
/arch/sparc/kernel/ |
D | visemul.c | 251 u16 left, right; member 297 u16 left, right; in edge() local 311 left = edge8_tab[rs1 & 0x7].left; in edge() 316 left = edge8_tab_l[rs1 & 0x7].left; in edge() 322 left = edge16_tab[(rs1 >> 1) & 0x3].left; in edge() 328 left = edge16_tab_l[(rs1 >> 1) & 0x3].left; in edge() 334 left = edge32_tab[(rs1 >> 2) & 0x1].left; in edge() 340 left = edge32_tab_l[(rs1 >> 2) & 0x1].left; in edge() 346 rd_val = right & left; in edge() 348 rd_val = left; in edge()
|
/arch/alpha/kernel/ |
D | perf_event.c | 255 long left = local64_read(&hwc->period_left); in alpha_perf_event_set_period() local 259 if (unlikely(left <= -period)) { in alpha_perf_event_set_period() 260 left = period; in alpha_perf_event_set_period() 261 local64_set(&hwc->period_left, left); in alpha_perf_event_set_period() 266 if (unlikely(left <= 0)) { in alpha_perf_event_set_period() 267 left += period; in alpha_perf_event_set_period() 268 local64_set(&hwc->period_left, left); in alpha_perf_event_set_period() 277 if (unlikely(left < alpha_pmu->pmc_left[idx])) in alpha_perf_event_set_period() 278 left = alpha_pmu->pmc_left[idx]; in alpha_perf_event_set_period() 280 if (left > (long)alpha_pmu->pmc_max_period[idx]) in alpha_perf_event_set_period() [all …]
|
/arch/arc/kernel/ |
D | perf_event.c | 219 s64 left = local64_read(&hwc->period_left); in arc_pmu_event_set_period() local 225 if (unlikely(left <= -period)) { in arc_pmu_event_set_period() 227 left = period; in arc_pmu_event_set_period() 228 local64_set(&hwc->period_left, left); in arc_pmu_event_set_period() 231 } else if (unlikely(left <= 0)) { in arc_pmu_event_set_period() 233 left += period; in arc_pmu_event_set_period() 234 local64_set(&hwc->period_left, left); in arc_pmu_event_set_period() 239 if (left > arc_pmu->max_period) in arc_pmu_event_set_period() 240 left = arc_pmu->max_period; in arc_pmu_event_set_period() 242 value = arc_pmu->max_period - left; in arc_pmu_event_set_period()
|
/arch/x86/events/amd/ |
D | ibs.c | 113 s64 left = local64_read(&hwc->period_left); in perf_event_set_period() local 120 if (unlikely(left <= -period)) { in perf_event_set_period() 121 left = period; in perf_event_set_period() 122 local64_set(&hwc->period_left, left); in perf_event_set_period() 127 if (unlikely(left < (s64)min)) { in perf_event_set_period() 128 left += period; in perf_event_set_period() 129 local64_set(&hwc->period_left, left); in perf_event_set_period() 140 if (left > max) { in perf_event_set_period() 141 left -= max; in perf_event_set_period() 142 if (left > max) in perf_event_set_period() [all …]
|
/arch/arm/xen/ |
D | mm.c | 52 size_t left = size; in dma_cache_maint() local 58 size_t len = left; in dma_cache_maint() 83 left -= len; in dma_cache_maint() 84 } while (left); in dma_cache_maint()
|
/arch/metag/kernel/perf/ |
D | perf_event.c | 224 s64 left = local64_read(&hwc->period_left); in metag_pmu_event_set_period() local 230 left += period - hwc->last_period; in metag_pmu_event_set_period() 232 if (unlikely(left <= -period)) { in metag_pmu_event_set_period() 233 left = period; in metag_pmu_event_set_period() 234 local64_set(&hwc->period_left, left); in metag_pmu_event_set_period() 239 if (unlikely(left <= 0)) { in metag_pmu_event_set_period() 240 left += period; in metag_pmu_event_set_period() 241 local64_set(&hwc->period_left, left); in metag_pmu_event_set_period() 246 if (left > (s64)metag_pmu->max_period) in metag_pmu_event_set_period() 247 left = metag_pmu->max_period; in metag_pmu_event_set_period() [all …]
|
/arch/alpha/lib/ |
D | ev6-clear_user.S | 74 addq $0, $4, $0 # .. E .. .. : bytes left -= 8 - misalignment 152 subq $0, 168, $5 # E : U L L U : two trips through the loop left? 171 # zero to 16 quadwords left to store, plus any trailing bytes 172 # $1 is the number of quadwords left to go. 190 # We have an unknown number of bytes left to go. 197 # $0 contains the number of bytes left to copy (0..31)
|
/arch/tile/kernel/ |
D | perf_event.c | 552 s64 left = local64_read(&hwc->period_left); in tile_event_set_period() local 559 if (unlikely(left <= -period)) { in tile_event_set_period() 560 left = period; in tile_event_set_period() 561 local64_set(&hwc->period_left, left); in tile_event_set_period() 566 if (unlikely(left <= 0)) { in tile_event_set_period() 567 left += period; in tile_event_set_period() 568 local64_set(&hwc->period_left, left); in tile_event_set_period() 572 if (left > tile_pmu->max_period) in tile_event_set_period() 573 left = tile_pmu->max_period; in tile_event_set_period() 579 local64_set(&hwc->prev_count, (u64)-left); in tile_event_set_period() [all …]
|
/arch/mips/mm/ |
D | dma-default.c | 242 size_t left = size; in __dma_sync() local 245 size_t len = left; in __dma_sync() 266 left -= len; in __dma_sync() 267 } while (left); in __dma_sync()
|
/arch/x86/events/ |
D | core.c | 1113 s64 left = local64_read(&hwc->period_left); in x86_perf_event_set_period() local 1123 if (unlikely(left <= -period)) { in x86_perf_event_set_period() 1124 left = period; in x86_perf_event_set_period() 1125 local64_set(&hwc->period_left, left); in x86_perf_event_set_period() 1130 if (unlikely(left <= 0)) { in x86_perf_event_set_period() 1131 left += period; in x86_perf_event_set_period() 1132 local64_set(&hwc->period_left, left); in x86_perf_event_set_period() 1139 if (unlikely(left < 2)) in x86_perf_event_set_period() 1140 left = 2; in x86_perf_event_set_period() 1142 if (left > x86_pmu.max_period) in x86_perf_event_set_period() [all …]
|
/arch/x86/xen/ |
D | setup.c | 339 unsigned long left = size; in xen_do_set_identity_and_remap_chunk() local 349 chunk = (left < REMAP_SIZE) ? left : REMAP_SIZE; in xen_do_set_identity_and_remap_chunk() 368 left -= chunk; in xen_do_set_identity_and_remap_chunk() 399 unsigned long left = n - i; in xen_set_identity_and_remap_chunk() local 400 unsigned long size = left; in xen_set_identity_and_remap_chunk() 416 cur_pfn + left, nr_pages); in xen_set_identity_and_remap_chunk()
|
/arch/mips/kernel/ |
D | perf_event_mipsxx.c | 368 u64 left = local64_read(&hwc->period_left); in mipspmu_event_set_period() local 372 if (unlikely((left + period) & (1ULL << 63))) { in mipspmu_event_set_period() 374 left = period; in mipspmu_event_set_period() 375 local64_set(&hwc->period_left, left); in mipspmu_event_set_period() 378 } else if (unlikely((left + period) <= period)) { in mipspmu_event_set_period() 380 left += period; in mipspmu_event_set_period() 381 local64_set(&hwc->period_left, left); in mipspmu_event_set_period() 386 if (left > mipspmu.max_period) { in mipspmu_event_set_period() 387 left = mipspmu.max_period; in mipspmu_event_set_period() 388 local64_set(&hwc->period_left, left); in mipspmu_event_set_period() [all …]
|