/external/python/cpython2/Tools/scripts/ |
D | cleanfuture.py | 166 type, token, (srow, scol), (erow, ecol), line = get() 170 type, token, (srow, scol), (erow, ecol), line = get() 174 type, token, (srow, scol), (erow, ecol), line = get() 180 type, token, (srow, scol), (erow, ecol), line = get() 184 startline = srow - 1 # tokenize is one-based 185 type, token, (srow, scol), (erow, ecol), line = get() 189 type, token, (srow, scol), (erow, ecol), line = get() 193 type, token, (srow, scol), (erow, ecol), line = get() 199 type, token, (srow, scol), (erow, ecol), line = get() 203 type, token, (srow, scol), (erow, ecol), line = get() [all …]
|
/external/python/cpython3/Tools/scripts/ |
D | cleanfuture.py | 165 type, token, (srow, scol), (erow, ecol), line = get() 169 type, token, (srow, scol), (erow, ecol), line = get() 173 type, token, (srow, scol), (erow, ecol), line = get() 179 type, token, (srow, scol), (erow, ecol), line = get() 183 startline = srow - 1 # tokenize is one-based 184 type, token, (srow, scol), (erow, ecol), line = get() 188 type, token, (srow, scol), (erow, ecol), line = get() 192 type, token, (srow, scol), (erow, ecol), line = get() 198 type, token, (srow, scol), (erow, ecol), line = get() 202 type, token, (srow, scol), (erow, ecol), line = get() [all …]
|
D | highlight.py | 21 (srow, scol), (erow, ecol) = start, end 22 if srow == erow: 23 return lines[srow-1][scol:ecol], end 24 rows = [lines[srow-1][scol:]] + lines[srow: erow-1] + [lines[erow-1][:ecol]] 39 tok_type, tok_str, (srow, scol), (erow, ecol), logical_lineno = tok 59 text, written = combine_range(lines, written, (srow, scol))
|
/external/angle/src/tests/gl_tests/ |
D | CubeMapTextureTest.cpp | 158 for (size_t srow = 0; srow < kCubeFaceSectionCountSqrt; ++srow) in runSampleCoordinateTransformTest() local 162 size_t r = row + srow * kTextureSize / kCubeFaceSectionCountSqrt; in runSampleCoordinateTransformTest() 164 size_t s = srow * kCubeFaceSectionCountSqrt + scol; in runSampleCoordinateTransformTest()
|
/external/chromium-trace/catapult/common/py_utils/py_utils/refactor/ |
D | offset_token.py | 75 token_type, string, (srow, scol), _, _ = next_token 80 if erow == srow: 84 token_type, string, (srow - erow, scol)))
|
/external/llvm-project/llvm/test/MC/RISCV/ |
D | rv64zbb-valid.s | 36 # CHECK-ASM-AND-OBJ: srow t0, t1, t2 38 srow t0, t1, t2 label
|
D | rv64zbb-invalid.s | 24 srow t0, t1 # CHECK: :[[@LINE]]:1: error: too few operands for instruction label
|
/external/tensorflow/tensorflow/python/autograph/pyct/ |
D | origin_info.py | 253 srow, _ = loc 255 comments_map[srow] = tok_string.strip()[1:].strip()
|
/external/mesa3d/src/gallium/drivers/etnaviv/ |
D | etnaviv_rs.c | 469 void *smap, *srow, *dmap, *drow; in etna_manual_blit() local 487 srow = smap + src_offset; in etna_manual_blit() 494 memcpy(drow, srow, tile_size * blit_info->src.box.width); in etna_manual_blit() 495 srow += src_lev->stride * 4; in etna_manual_blit()
|
/external/python/cpython2/Lib/lib2to3/pgen2/ |
D | tokenize.py | 155 (srow, scol) = start 158 (srow, scol, erow, ecol, tok_name[type], repr(token))
|
/external/python/cpython2/Lib/ |
D | tokenize.py | 151 srow, scol = srow_scol 154 (srow, scol, erow, ecol, tok_name[type], repr(token))
|
D | inspect.py | 641 srow, scol = srow_scol 652 self.last = srow
|
/external/python/cpython3/Lib/lib2to3/pgen2/ |
D | tokenize.py | 143 (srow, scol) = xxx_todo_changeme 146 (srow, scol, erow, ecol, tok_name[type], repr(token)))
|
/external/python/cpython2/Doc/library/ |
D | tokenize.rst | 37 string; a 2-tuple ``(srow, scol)`` of ints specifying the row and column
|
/external/python/cpython3/Doc/library/ |
D | tokenize.rst | 38 token string; a 2-tuple ``(srow, scol)`` of ints specifying the row and
|
/external/llvm-project/llvm/test/CodeGen/RISCV/ |
D | rv64Zbb.ll | 65 ; RV64IB-NEXT: srow a0, a0, a1 70 ; RV64IBB-NEXT: srow a0, a0, a1
|
/external/llvm-project/llvm/lib/Target/RISCV/ |
D | RISCVInstrInfoB.td | 363 def SROW : ALUW_rr<0b0010000, 0b101, "srow">, Sched<[]>;
|