Home
last modified time | relevance | path

Searched refs:left (Results 1 – 25 of 106) sorted by relevance

12345

/arch/parisc/math-emu/
Dsfcmp.c55 register unsigned int left, right; in sgl_fcmp() local
59 left = *leftptr; in sgl_fcmp()
65 if( (Sgl_exponent(left) == SGL_INFINITY_EXPONENT) in sgl_fcmp()
71 if( ( (Sgl_exponent(left) == SGL_INFINITY_EXPONENT) in sgl_fcmp()
72 && Sgl_isnotzero_mantissa(left) in sgl_fcmp()
73 && (Exception(cond) || Sgl_isone_signaling(left))) in sgl_fcmp()
89 else if( ((Sgl_exponent(left) == SGL_INFINITY_EXPONENT) in sgl_fcmp()
90 && Sgl_isnotzero_mantissa(left)) in sgl_fcmp()
103 Sgl_xortointp1(left,right,xorresult); in sgl_fcmp()
108 if( Sgl_iszero_exponentmantissa(left) in sgl_fcmp()
[all …]
Dsfsub.c55 register unsigned int left, right, result, extent; in sgl_fsub() local
63 left = *leftptr; in sgl_fsub()
68 Sgl_xortointp1(left,right,/*to*/save); in sgl_fsub()
73 if ((result_exponent = Sgl_exponent(left)) == SGL_INFINITY_EXPONENT) in sgl_fsub()
75 if (Sgl_iszero_mantissa(left)) in sgl_fsub()
93 *dstptr = left; in sgl_fsub()
102 if (Sgl_isone_signaling(left)) in sgl_fsub()
108 Sgl_set_quiet(left); in sgl_fsub()
126 *dstptr = left; in sgl_fsub()
163 Sgl_copytoint_exponentmantissa(left,signless_upper_left); in sgl_fsub()
[all …]
Dsfadd.c55 register unsigned int left, right, result, extent; in sgl_fadd() local
65 left = *leftptr; in sgl_fadd()
70 Sgl_xortointp1(left,right,/*to*/save); in sgl_fadd()
75 if ((result_exponent = Sgl_exponent(left)) == SGL_INFINITY_EXPONENT) in sgl_fadd()
77 if (Sgl_iszero_mantissa(left)) in sgl_fadd()
95 *dstptr = left; in sgl_fadd()
104 if (Sgl_isone_signaling(left)) in sgl_fadd()
110 Sgl_set_quiet(left); in sgl_fadd()
128 *dstptr = left; in sgl_fadd()
164 Sgl_copytoint_exponentmantissa(left,signless_upper_left); in sgl_fadd()
[all …]
Dhppa.h28 #define Shiftdouble(left,right,amount,dest) \ argument
30 dest = ((left) << (32-(amount))) | ((unsigned int)(right) >> (amount))
33 #define Variableshiftdouble(left,right,amount,dest) \ argument
36 else dest = ((((unsigned) left)&0x7fffffff) << (32-(amount))) | \
40 #define Variable_shift_double(left,right,amount,dest) \ argument
42 dest = (left << (32-(amount))) | ((unsigned) right >> (amount))
Dsgl_float.h63 #define Sgl_leftshiftby1_withextent(left,right,result) \ argument
64 Shiftdouble(Sall(left),Extall(right),31,Sall(result))
66 #define Sgl_rightshiftby1_withextent(left,right,dst) \ argument
67 Shiftdouble(Sall(left),Extall(right),1,Extall(right))
239 #define Sgl_subtract_withextension(left,right,extent,result) \ argument
241 Sgl_subtract(left,right,result); \
248 #define Sgl_xortointp1(left,right,result) \ argument
249 result = Sall(left) XOR Sall(right);
251 #define Sgl_xorfromintp1(left,right,result) \ argument
252 Sall(result) = left XOR Sall(right)
[all …]
Ddbl_float.h421 #define Dbl_xortointp1(left,right,result) \ argument
422 result = Dallp1(left) XOR Dallp1(right)
424 #define Dbl_xorfromintp1(left,right,result) \ argument
425 Dallp1(result) = left XOR Dallp1(right)
427 #define Dbl_swap_lower(left,right) \ argument
428 Dallp2(left) = Dallp2(left) XOR Dallp2(right); \
429 Dallp2(right) = Dallp2(left) XOR Dallp2(right); \
430 Dallp2(left) = Dallp2(left) XOR Dallp2(right)
720 #define Dblext_xortointp1(left,right,result) Dbl_xortointp1(left,right,result) argument
722 #define Dblext_xorfromintp1(left,right,result) \ argument
[all …]
/arch/x86/crypto/
Ddes3_ede-asm_64.S101 #define initial_permutation(left, right) \ argument
102 do_permutation(left##d, right##d, 4, 0x0f0f0f0f); \
103 do_permutation(left##d, right##d, 16, 0x0000ffff); \
104 do_permutation(right##d, left##d, 2, 0x33333333); \
105 do_permutation(right##d, left##d, 8, 0x00ff00ff); \
107 movl left##d, RW0d; \
111 xorl RW0d, left##d; \
113 roll $1, left##d; \
115 expand_to_64bits(left, RT3);
117 #define final_permutation(left, right) \ argument
[all …]
/arch/hexagon/include/asm/
Duaccess.h100 long left = raw_copy_from_user(dst, src, n); in hexagon_strncpy_from_user() local
101 if (unlikely(left)) in hexagon_strncpy_from_user()
102 memset(dst + (n - left), 0, left); in hexagon_strncpy_from_user()
105 long left = raw_copy_from_user(dst, src, res); in hexagon_strncpy_from_user() local
106 if (unlikely(left)) in hexagon_strncpy_from_user()
107 memset(dst + (res - left), 0, left); in hexagon_strncpy_from_user()
/arch/arm/plat-samsung/
Dpm-check.c110 unsigned long addr, left; in s3c_pm_makecheck() local
114 left = res->end - addr; in s3c_pm_makecheck()
116 if (left > CHECK_CHUNKSIZE) in s3c_pm_makecheck()
117 left = CHECK_CHUNKSIZE; in s3c_pm_makecheck()
119 *val = crc32_le(~0, phys_to_virt(addr), left); in s3c_pm_makecheck()
168 unsigned long left; in s3c_pm_runcheck() local
177 left = res->end - addr; in s3c_pm_runcheck()
179 if (left > CHECK_CHUNKSIZE) in s3c_pm_runcheck()
180 left = CHECK_CHUNKSIZE; in s3c_pm_runcheck()
184 if (in_region(ptr, left, stkpage, 4096)) { in s3c_pm_runcheck()
[all …]
/arch/s390/include/asm/
Didals.h199 size_t left; in idal_buffer_to_user() local
204 left = copy_to_user(to, ib->data[i], IDA_BLOCK_SIZE); in idal_buffer_to_user()
205 if (left) in idal_buffer_to_user()
206 return left + count - IDA_BLOCK_SIZE; in idal_buffer_to_user()
219 size_t left; in idal_buffer_from_user() local
224 left = copy_from_user(ib->data[i], from, IDA_BLOCK_SIZE); in idal_buffer_from_user()
225 if (left) in idal_buffer_from_user()
226 return left + count - IDA_BLOCK_SIZE; in idal_buffer_from_user()
/arch/xtensa/kernel/
Dperf_event.c161 s64 left; in xtensa_perf_event_set_period() local
164 left = XTENSA_PMU_COUNTER_MAX; in xtensa_perf_event_set_period()
168 left = local64_read(&hwc->period_left); in xtensa_perf_event_set_period()
169 if (left <= -period) { in xtensa_perf_event_set_period()
170 left = period; in xtensa_perf_event_set_period()
171 local64_set(&hwc->period_left, left); in xtensa_perf_event_set_period()
174 } else if (left <= 0) { in xtensa_perf_event_set_period()
175 left += period; in xtensa_perf_event_set_period()
176 local64_set(&hwc->period_left, left); in xtensa_perf_event_set_period()
180 if (left > XTENSA_PMU_COUNTER_MAX) in xtensa_perf_event_set_period()
[all …]
/arch/powerpc/perf/
Dcore-fsl-emb.c327 s64 left = local64_read(&event->hw.period_left); in fsl_emb_pmu_add() local
328 if (left < 0x80000000L) in fsl_emb_pmu_add()
329 val = 0x80000000L - left; in fsl_emb_pmu_add()
395 s64 left; in fsl_emb_pmu_start() local
410 left = local64_read(&event->hw.period_left); in fsl_emb_pmu_start()
412 if (left < 0x80000000L) in fsl_emb_pmu_start()
413 val = 0x80000000L - left; in fsl_emb_pmu_start()
610 s64 prev, delta, left; in record_and_restart() local
628 left = local64_read(&event->hw.period_left) - delta; in record_and_restart()
630 if (left <= 0) { in record_and_restart()
[all …]
Dcore-book3s.c1285 s64 left; in power_pmu_enable() local
1399 left = local64_read(&event->hw.period_left); in power_pmu_enable()
1400 if (left < 0x80000000L) in power_pmu_enable()
1401 val = 0x80000000L - left; in power_pmu_enable()
1608 s64 left; in power_pmu_start() local
1624 left = local64_read(&event->hw.period_left); in power_pmu_start()
1627 if (left < 0x80000000L) in power_pmu_start()
1628 val = 0x80000000L - left; in power_pmu_start()
2032 s64 prev, delta, left; in record_and_restart() local
2050 left = local64_read(&event->hw.period_left) - delta; in record_and_restart()
[all …]
/arch/sparc/kernel/
Dvisemul.c251 u16 left, right; member
297 u16 left, right; in edge() local
311 left = edge8_tab[rs1 & 0x7].left; in edge()
316 left = edge8_tab_l[rs1 & 0x7].left; in edge()
322 left = edge16_tab[(rs1 >> 1) & 0x3].left; in edge()
328 left = edge16_tab_l[(rs1 >> 1) & 0x3].left; in edge()
334 left = edge32_tab[(rs1 >> 2) & 0x1].left; in edge()
340 left = edge32_tab_l[(rs1 >> 2) & 0x1].left; in edge()
346 rd_val = right & left; in edge()
348 rd_val = left; in edge()
/arch/alpha/kernel/
Dperf_event.c255 long left = local64_read(&hwc->period_left); in alpha_perf_event_set_period() local
259 if (unlikely(left <= -period)) { in alpha_perf_event_set_period()
260 left = period; in alpha_perf_event_set_period()
261 local64_set(&hwc->period_left, left); in alpha_perf_event_set_period()
266 if (unlikely(left <= 0)) { in alpha_perf_event_set_period()
267 left += period; in alpha_perf_event_set_period()
268 local64_set(&hwc->period_left, left); in alpha_perf_event_set_period()
277 if (unlikely(left < alpha_pmu->pmc_left[idx])) in alpha_perf_event_set_period()
278 left = alpha_pmu->pmc_left[idx]; in alpha_perf_event_set_period()
280 if (left > (long)alpha_pmu->pmc_max_period[idx]) in alpha_perf_event_set_period()
[all …]
/arch/arc/kernel/
Dperf_event.c219 s64 left = local64_read(&hwc->period_left); in arc_pmu_event_set_period() local
225 if (unlikely(left <= -period)) { in arc_pmu_event_set_period()
227 left = period; in arc_pmu_event_set_period()
228 local64_set(&hwc->period_left, left); in arc_pmu_event_set_period()
231 } else if (unlikely(left <= 0)) { in arc_pmu_event_set_period()
233 left += period; in arc_pmu_event_set_period()
234 local64_set(&hwc->period_left, left); in arc_pmu_event_set_period()
239 if (left > arc_pmu->max_period) in arc_pmu_event_set_period()
240 left = arc_pmu->max_period; in arc_pmu_event_set_period()
242 value = arc_pmu->max_period - left; in arc_pmu_event_set_period()
/arch/x86/events/amd/
Dibs.c113 s64 left = local64_read(&hwc->period_left); in perf_event_set_period() local
120 if (unlikely(left <= -period)) { in perf_event_set_period()
121 left = period; in perf_event_set_period()
122 local64_set(&hwc->period_left, left); in perf_event_set_period()
127 if (unlikely(left < (s64)min)) { in perf_event_set_period()
128 left += period; in perf_event_set_period()
129 local64_set(&hwc->period_left, left); in perf_event_set_period()
140 if (left > max) { in perf_event_set_period()
141 left -= max; in perf_event_set_period()
142 if (left > max) in perf_event_set_period()
[all …]
/arch/arm/xen/
Dmm.c52 size_t left = size; in dma_cache_maint() local
58 size_t len = left; in dma_cache_maint()
83 left -= len; in dma_cache_maint()
84 } while (left); in dma_cache_maint()
/arch/metag/kernel/perf/
Dperf_event.c224 s64 left = local64_read(&hwc->period_left); in metag_pmu_event_set_period() local
230 left += period - hwc->last_period; in metag_pmu_event_set_period()
232 if (unlikely(left <= -period)) { in metag_pmu_event_set_period()
233 left = period; in metag_pmu_event_set_period()
234 local64_set(&hwc->period_left, left); in metag_pmu_event_set_period()
239 if (unlikely(left <= 0)) { in metag_pmu_event_set_period()
240 left += period; in metag_pmu_event_set_period()
241 local64_set(&hwc->period_left, left); in metag_pmu_event_set_period()
246 if (left > (s64)metag_pmu->max_period) in metag_pmu_event_set_period()
247 left = metag_pmu->max_period; in metag_pmu_event_set_period()
[all …]
/arch/alpha/lib/
Dev6-clear_user.S74 addq $0, $4, $0 # .. E .. .. : bytes left -= 8 - misalignment
152 subq $0, 168, $5 # E : U L L U : two trips through the loop left?
171 # zero to 16 quadwords left to store, plus any trailing bytes
172 # $1 is the number of quadwords left to go.
190 # We have an unknown number of bytes left to go.
197 # $0 contains the number of bytes left to copy (0..31)
/arch/tile/kernel/
Dperf_event.c552 s64 left = local64_read(&hwc->period_left); in tile_event_set_period() local
559 if (unlikely(left <= -period)) { in tile_event_set_period()
560 left = period; in tile_event_set_period()
561 local64_set(&hwc->period_left, left); in tile_event_set_period()
566 if (unlikely(left <= 0)) { in tile_event_set_period()
567 left += period; in tile_event_set_period()
568 local64_set(&hwc->period_left, left); in tile_event_set_period()
572 if (left > tile_pmu->max_period) in tile_event_set_period()
573 left = tile_pmu->max_period; in tile_event_set_period()
579 local64_set(&hwc->prev_count, (u64)-left); in tile_event_set_period()
[all …]
/arch/mips/mm/
Ddma-default.c242 size_t left = size; in __dma_sync() local
245 size_t len = left; in __dma_sync()
266 left -= len; in __dma_sync()
267 } while (left); in __dma_sync()
/arch/x86/events/
Dcore.c1113 s64 left = local64_read(&hwc->period_left); in x86_perf_event_set_period() local
1123 if (unlikely(left <= -period)) { in x86_perf_event_set_period()
1124 left = period; in x86_perf_event_set_period()
1125 local64_set(&hwc->period_left, left); in x86_perf_event_set_period()
1130 if (unlikely(left <= 0)) { in x86_perf_event_set_period()
1131 left += period; in x86_perf_event_set_period()
1132 local64_set(&hwc->period_left, left); in x86_perf_event_set_period()
1139 if (unlikely(left < 2)) in x86_perf_event_set_period()
1140 left = 2; in x86_perf_event_set_period()
1142 if (left > x86_pmu.max_period) in x86_perf_event_set_period()
[all …]
/arch/x86/xen/
Dsetup.c339 unsigned long left = size; in xen_do_set_identity_and_remap_chunk() local
349 chunk = (left < REMAP_SIZE) ? left : REMAP_SIZE; in xen_do_set_identity_and_remap_chunk()
368 left -= chunk; in xen_do_set_identity_and_remap_chunk()
399 unsigned long left = n - i; in xen_set_identity_and_remap_chunk() local
400 unsigned long size = left; in xen_set_identity_and_remap_chunk()
416 cur_pfn + left, nr_pages); in xen_set_identity_and_remap_chunk()
/arch/mips/kernel/
Dperf_event_mipsxx.c368 u64 left = local64_read(&hwc->period_left); in mipspmu_event_set_period() local
372 if (unlikely((left + period) & (1ULL << 63))) { in mipspmu_event_set_period()
374 left = period; in mipspmu_event_set_period()
375 local64_set(&hwc->period_left, left); in mipspmu_event_set_period()
378 } else if (unlikely((left + period) <= period)) { in mipspmu_event_set_period()
380 left += period; in mipspmu_event_set_period()
381 local64_set(&hwc->period_left, left); in mipspmu_event_set_period()
386 if (left > mipspmu.max_period) { in mipspmu_event_set_period()
387 left = mipspmu.max_period; in mipspmu_event_set_period()
388 local64_set(&hwc->period_left, left); in mipspmu_event_set_period()
[all …]

12345