Searched refs:next_chunk (Results 1 – 13 of 13) sorted by relevance
337 const ChunkRecord& next_chunk = *GetChunkRecordAt(next_chunk_ptr); in DeleteNextChunksFor() local340 next_chunk_ptr - begin() + next_chunk.size, next_chunk.is_valid()); in DeleteNextChunksFor()346 if (PERFETTO_UNLIKELY(!next_chunk.is_valid())) { in DeleteNextChunksFor()356 if (PERFETTO_LIKELY(!next_chunk.is_padding)) { in DeleteNextChunksFor()357 ChunkMeta::Key key(next_chunk); in DeleteNextChunksFor()366 bytes_overwritten += next_chunk.size; in DeleteNextChunksFor()374 next_chunk_ptr - begin(), next_chunk_ptr - begin() + next_chunk.size, in DeleteNextChunksFor()378 padding_bytes_cleared += next_chunk.size; in DeleteNextChunksFor()381 next_chunk_ptr += next_chunk.size; in DeleteNextChunksFor()
972 status, body = request.next_chunk(http=http)984 status, body = request.next_chunk(http=http)990 status, body = request.next_chunk(http=http)996 status, body = request.next_chunk(http=http)1027 status, body = request.next_chunk(http=http)1039 status, body = request.next_chunk(http=http)1045 status, body = request.next_chunk(http=http)1118 status, body = request.next_chunk(http=http)1134 status, body = request.next_chunk(http=http)1214 status, body = request.next_chunk(http=http)[all …]
459 status, done = download.next_chunk()467 status, done = download.next_chunk()490 status, done = download.next_chunk()500 status, done = download.next_chunk()516 status, done = download.next_chunk()529 status, done = download.next_chunk()540 status, done = download.next_chunk()553 status, done = download.next_chunk(num_retries=4)587 status, done = download.next_chunk(num_retries=3)601 status, done = download.next_chunk(num_retries=3)[all …]
31 …ad object and flag it as a resumable upload. You then repeatedly call `next_chunk()` on the [`goog…38 status, response = request.next_chunk()52 … is retryable, the upload can be resumed by continuing to call request.next_chunk(), but subsequen…67 # Call next_chunk() again, but use an exponential backoff for repeated errors.
308 uptr next_chunk = frontier->back(); in FloodFillTag() local310 LsanMetadata m(next_chunk); in FloodFillTag()311 ScanRangeForPointers(next_chunk, next_chunk + m.requested_size(), frontier, in FloodFillTag()
107 WebPChunk* next_chunk = ChunkSearchNextInList(first->next_, tag); in ChunkSearchList() local108 if (next_chunk == NULL) break; in ChunkSearchList()109 first = next_chunk; in ChunkSearchList()
399 #define next_chunk(p) ((mchunkptr)( ((char*)(p)) + ((p)->size & ~PREV_INUSE) )) macro788 mchunkptr next = next_chunk(p);801 assert(next_chunk(prv) == p);1819 if ( (newp = mem2chunk(newmem)) == next_chunk(oldp))2278 for (q = next_chunk(p); in malloc_update_mallinfo()2280 q = next_chunk(q)) in malloc_update_mallinfo()
1318 #define next_chunk(p) ((mchunkptr)( ((char*)(p)) + ((p)->size & ~PREV_INUSE) ))1653 mchunkptr next = next_chunk(p);1666 assert(next_chunk(prv) == p);2656 if ( (newp = mem2chunk(newmem)) == next_chunk(oldp))3057 for (q = next_chunk(p);3059 q = next_chunk(q))
1741 #define next_chunk(p) ((mchunkptr)( ((char*)(p)) + ((p)->head & ~INUSE_BITS))) macro1745 #define next_pinuse(p) ((next_chunk(p)->head) & PINUSE_BIT)2660 assert(is_mmapped(p) || pinuse(p) || next_chunk(prev_chunk(p)) == p); in do_check_inuse_chunk()2781 q = next_chunk(p); in do_check_smallbin()2847 q = next_chunk(q); in traverse_and_check()2907 q = next_chunk(q); in internal_mallinfo()2945 q = next_chunk(q); in internal_malloc_stats()3563 mchunkptr mn = next_chunk(mem2chunk(m)); in sys_alloc()4487 mn = next_chunk(mem2chunk(m)); in init_user_mstate()
1736 #define next_chunk(p) ((mchunkptr)( ((char*)(p)) + ((p)->head & ~INUSE_BITS))) macro1740 #define next_pinuse(p) ((next_chunk(p)->head) & PINUSE_BIT)2655 assert(is_mmapped(p) || pinuse(p) || next_chunk(prev_chunk(p)) == p); in do_check_inuse_chunk()2776 q = next_chunk(p); in do_check_smallbin()2842 q = next_chunk(q); in traverse_and_check()2902 q = next_chunk(q); in internal_mallinfo()2940 q = next_chunk(q); in internal_malloc_stats()3558 mchunkptr mn = next_chunk(mem2chunk(m)); in sys_alloc()4482 mn = next_chunk(mem2chunk(m)); in init_user_mstate()
2273 #define next_chunk(p) ((mchunkptr)( ((char*)(p)) + ((p)->head & ~FLAG_BITS))) macro2277 #define next_pinuse(p) ((next_chunk(p)->head) & PINUSE_BIT)3272 assert(is_mmapped(p) || pinuse(p) || next_chunk(prev_chunk(p)) == p); in do_check_inuse_chunk()3393 q = next_chunk(p); in do_check_smallbin()3459 q = next_chunk(q); in traverse_and_check()3521 q = next_chunk(q); in internal_mallinfo()3563 q = next_chunk(q); in internal_malloc_stats()4216 mchunkptr mn = next_chunk(mem2chunk(m)); in sys_alloc()5146 mchunkptr next = next_chunk(p); in internal_bulk_free()5182 mchunkptr next = next_chunk(q); in internal_inspect_all()[all …]
666 def next_chunk(self, num_retries=0): member in MediaIoBaseDownload827 _, body = self.next_chunk(http=http, num_retries=num_retries)872 def next_chunk(self, http=None, num_retries=0): member in HttpRequest
205 googleapiclient.http.HttpRequest.next_chunk googleapiclient.http.HttpRequest-class.html#next_chunk241 …oogleapiclient.http.MediaIoBaseDownload.next_chunk googleapiclient.http.MediaIoBaseDownload-class.…