/external/ltp/testcases/realtime/lib/ |
D | libstats.c | 75 data->records = calloc(size, sizeof(stats_record_t)); in stats_container_init() 76 if (!data->records) in stats_container_init() 90 data->records[myindex] = rec; in stats_container_append() 97 realloc(data->records, size * sizeof(stats_record_t)); in stats_container_resize() 100 data->records = newrecords; in stats_container_resize() 102 memset(data->records + data->size, 0, size - data->size); in stats_container_resize() 109 free(data->records); in stats_container_free() 116 qsort(data->records, data->index + 1, sizeof(stats_record_t), in stats_sort() 133 sum += data->records[i].y; in stats_stddev() 140 delta = (data->records[i].y - avg); in stats_stddev() [all …]
|
/external/selinux/libsemanage/src/ |
D | database_activedb.c | 39 record_t **records = NULL; in dbase_activedb_cache() local 53 if (ratable->read_list(handle, &records, &rcount) < 0) in dbase_activedb_cache() 58 if (dbase_llist_cache_prepend(handle, &dbase->llist, records[i]) in dbase_activedb_cache() 61 rtable->free(records[i]); in dbase_activedb_cache() 64 free(records); in dbase_activedb_cache() 70 rtable->free(records[i]); in dbase_activedb_cache() 72 free(records); in dbase_activedb_cache() 83 record_t **records = NULL; in dbase_activedb_flush() local 92 if (dbase_llist_list(handle, &dbase->llist, &records, &rcount) < 0) in dbase_activedb_flush() 96 if (ratable->commit_list(handle, records, rcount) < 0) in dbase_activedb_flush() [all …]
|
D | policy_components.c | 16 record_t ** records, in clear_obsolete() argument 31 if (rtable->key_extract(handle, records[i], &key) < 0) in clear_obsolete() 41 rtable->free(records[i]); in clear_obsolete() 42 records[i] = NULL; in clear_obsolete() 61 record_t ** records, unsigned int nrecords, int mode) in load_records() argument 74 if (!records[i]) in load_records() 77 if (rtable->key_extract(handle, records[i], &rkey) < 0) in load_records() 81 dtable->set(handle, dbase, rkey, records[i]) < 0) in load_records() 85 dtable->modify(handle, dbase, rkey, records[i]) < 0) in load_records() 151 record_t **records = NULL; in semanage_base_merge_components() local [all …]
|
D | users_local.c | 29 semanage_seuser_t **records; in lookup_seuser() local 39 &records, in lookup_seuser() 42 sename = semanage_seuser_get_sename(records[i]); in lookup_seuser() 46 sename, semanage_seuser_get_name(records[i])); in lookup_seuser() 51 semanage_seuser_free(records[i]); in lookup_seuser() 52 free(records); in lookup_seuser() 104 semanage_user_t *** records, unsigned int *count) in semanage_user_list_local() argument 108 return dbase_list(handle, dconfig, records, count); in semanage_user_list_local()
|
D | fcontexts_policy.c | 48 semanage_fcontext_t *** records, unsigned int *count) in semanage_fcontext_list() argument 52 return dbase_list(handle, dconfig, records, count); in semanage_fcontext_list() 56 semanage_fcontext_t *** records, unsigned int *count) in semanage_fcontext_list_homedirs() argument 60 return dbase_list(handle, dconfig, records, count); in semanage_fcontext_list_homedirs()
|
D | semanageswig_python.i | 190 semanage_bool_t*** records, 233 semanage_fcontext_t*** records, 277 semanage_iface_t*** records, 321 semanage_seuser_t*** records, 364 semanage_user_t*** records, 407 semanage_port_t*** records, 450 semanage_ibpkey_t*** records, 493 semanage_ibendport_t*** records, 536 semanage_node_t*** records,
|
/external/mp4parser/isoparser/src/main/java/com/googlecode/mp4parser/boxes/piff/ |
D | PlayReadyHeader.java | 41 private List<PlayReadyRecord> records; field in PlayReadyHeader 61 records = PlayReadyRecord.createFor(byteBuffer, recordCount); in parse() 68 for (PlayReadyRecord record : records) { in getData() 75 IsoTypeWriter.writeUInt16BE(byteBuffer, records.size()); in getData() 76 for (PlayReadyRecord record : records) { in getData() 86 public void setRecords(List<PlayReadyRecord> records) { in setRecords() argument 87 this.records = records; in setRecords() 91 return Collections.unmodifiableList(records); in getRecords() 99 sb.append(", recordCount=").append(records.size()); in toString() 100 sb.append(", records=").append(records); in toString() [all …]
|
/external/tensorflow/tensorflow/core/framework/ |
D | tracking_allocator_test.cc | 85 auto records = ta->GetRecordsAndUnRef(); in TEST() local 86 EXPECT_EQ(4, records[0].alloc_bytes); in TEST() 87 EXPECT_EQ(12, records[1].alloc_bytes); in TEST() 109 records = ta->GetRecordsAndUnRef(); in TEST() 110 EXPECT_LE(4, records[0].alloc_bytes); in TEST() 111 EXPECT_GE(-4, records[1].alloc_bytes); in TEST() 112 EXPECT_LE(12, records[2].alloc_bytes); in TEST() 113 EXPECT_GE(-12, records[3].alloc_bytes); in TEST() 135 auto records = ta->GetRecordsAndUnRef(); in TEST() local 136 EXPECT_EQ(12, records[0].alloc_bytes); in TEST() [all …]
|
/external/python/cpython3/Lib/test/ |
D | test_urllibnet.py | 193 records = [] 196 records.append((blocks, block_size, total_size)) 202 records_repr = repr(records) # For use in error messages. 203 self.assertGreater(len(records), 1, msg="There should always be two " 205 self.assertEqual(records[0][0], 0) 206 self.assertGreater(records[0][1], 0, 208 self.assertEqual(records[0][2], expected_size) 209 self.assertEqual(records[-1][2], expected_size) 211 block_sizes = {block_size for _, block_size, _ in records} 212 self.assertEqual({records[0][1]}, block_sizes, [all …]
|
/external/tensorflow/tensorflow/tools/graph_transforms/ |
D | freeze_requantization_ranges_test.cc | 42 std::vector<MinMaxRecord>* records); 179 std::vector<MinMaxRecord> records; in TestExtractMinMaxRecords() local 180 TF_ASSERT_OK(ExtractMinMaxRecords(min_max_log_file_name, &records)); in TestExtractMinMaxRecords() 181 ASSERT_EQ(2, records.size()); in TestExtractMinMaxRecords() 182 EXPECT_EQ("requantization_range_op", records[0].name); in TestExtractMinMaxRecords() 183 EXPECT_NEAR(-2.4313571f, records[0].min, 1e-5f); in TestExtractMinMaxRecords() 184 EXPECT_NEAR(10.584145f, records[0].max, 1e-5f); in TestExtractMinMaxRecords() 185 EXPECT_EQ("other_requantization_range_op", records[1].name); in TestExtractMinMaxRecords() 186 EXPECT_NEAR(-1.0f, records[1].min, 1e-5f); in TestExtractMinMaxRecords() 187 EXPECT_NEAR(2.0f, records[1].max, 1e-5f); in TestExtractMinMaxRecords()
|
D | freeze_requantization_ranges.cc | 33 std::vector<MinMaxRecord>* records) { in ExtractMinMaxRecords() argument 97 records->push_back({name, min, max}); in ExtractMinMaxRecords() 121 std::vector<MinMaxRecord> records; in FreezeRequantizationRanges() local 122 TF_RETURN_IF_ERROR(ExtractMinMaxRecords(min_max_log_file, &records)); in FreezeRequantizationRanges() 123 if (records.empty()) { in FreezeRequantizationRanges() 132 for (const MinMaxRecord& record : records) { in FreezeRequantizationRanges() 148 const std::vector<MinMaxRecord> records = record_info.second; in FreezeRequantizationRanges() local 151 for (const MinMaxRecord& record : records) { in FreezeRequantizationRanges()
|
/external/tensorflow/tensorflow/python/lib/io/ |
D | tf_record_test.py | 85 records = [self._Record(i, j) for j in range(self._num_records)] 86 fn = self._WriteRecordsToFile(records, name, options) 90 def _WriteRecordsToFile(self, records, name="tfrecord", options=None): argument 93 for r in records: 144 def _CompressionSizeDelta(self, records, options_a, options_b): argument 164 fn_a = self._WriteRecordsToFile(records, "tfrecord_a", options=options_a) 166 self.assertEqual(records, test_a, options_a) 168 fn_b = self._WriteRecordsToFile(records, "tfrecord_b", options=options_b) 170 self.assertEqual(records, test_b, options_b) 266 records = [random_record, repeated_record, random_record] [all …]
|
/external/icu/icu4j/tools/misc/src/com/ibm/icu/dev/tool/layout/ |
D | ArabicCharacterData.java | 75 records = new Record[charCount]; in ArabicCharacterData() 80 records[recordIndex++] = new Record(character); in add() 85 if (index < 0 || index >= records.length) { in getRecord() 89 return records[index]; in getRecord() 94 return records.length; in countRecords() 114 private Record[] records; field in ArabicCharacterData
|
/external/python/cpython2/Lib/bsddb/test/ |
D | test_thread.py | 73 records = 1000 variable in ConcurrentDataStoreBase 81 keys=range(self.records) 84 records_per_writer=self.records//self.writers 86 self.assertEqual(self.records,self.writers*records_per_writer) 181 records = 1000 variable in BTreeConcurrentDataStore 188 records = 1000 variable in HashConcurrentDataStore 198 records = 1000 variable in SimpleThreadedBase 209 keys=range(self.records) 212 records_per_writer=self.records//self.writers 214 self.assertEqual(self.records,self.writers*records_per_writer) [all …]
|
/external/boringssl/src/ssl/test/runner/ |
D | dtls.go | 291 var records [][]byte 295 records = append(records, fragment[:n]) 296 records = append(records, fragment[n:]) 298 records = append(records, fragment) 300 …} else if i := len(records) - 1; len(records) > 0 && len(records[i])+len(fragment) <= maxRecordLen… 301 records[i] = append(records[i], fragment...) 304 records = append(records, append([]byte{}, fragment...)) 309 for _, record := range records {
|
/external/ltp/testcases/realtime/func/rt-migrate/ |
D | rt-migrate.c | 268 printf("%6ld ", intervals[t].records[i].y); in print_results() 273 printf("%6ld ", intervals_length[t].records[i].y); in print_results() 278 printf("%6ld ", intervals_loops[t].records[i].y); in print_results() 366 if (i && last < intervals[i].records[l].y && in check_times() 367 ((intervals[i].records[l].y - last) > max_err)) { in check_times() 374 if (intervals_loops[i].records[l].y < last_loops || in check_times() 375 intervals[i].records[l].y > last_length || in check_times() 376 (intervals_length[i].records[l].y > last_length && in check_times() 377 intervals_length[i].records[l].y - last_length > in check_times() 383 last = intervals[i].records[l].y; in check_times() [all …]
|
/external/tcpdump/tests/ |
D | sflow_multiple_counter_30_pdus.out | 3 expanded counter sample (4), length 172, seqnum 87096, type 0, idx 55, records 2 15 expanded counter sample (4), length 172, seqnum 87096, type 0, idx 56, records 2 27 expanded counter sample (4), length 172, seqnum 87099, type 0, idx 57, records 2 39 expanded counter sample (4), length 172, seqnum 87096, type 0, idx 60, records 2 51 expanded counter sample (4), length 172, seqnum 87016, type 0, idx 61, records 2 63 expanded counter sample (4), length 172, seqnum 87096, type 0, idx 62, records 2 75 expanded counter sample (4), length 172, seqnum 87096, type 0, idx 63, records 2 89 expanded counter sample (4), length 172, seqnum 2007195, type 0, idx 1, records 2 101 expanded counter sample (4), length 172, seqnum 2006745, type 0, idx 2, records 2 113 expanded counter sample (4), length 172, seqnum 2007118, type 0, idx 3, records 2 [all …]
|
/external/syzkaller/syz-hub/state/ |
D | state.go | 272 var records []db.Record 287 records = append(records, rec) 293 if len(records) > maxRecords { 294 sort.Sort(recordSeqSorter(records)) 296 maxSeq = records[pos].Seq 297 for pos+1 < len(records) && records[pos+1].Seq == maxSeq { 301 more = len(records) - pos 302 records = records[:pos] 304 progs := make([][]byte, len(records)) 305 for _, rec := range records {
|
/external/grpc-grpc-java/core/src/test/java/io/grpc/internal/ |
D | ManagedChannelOrphanWrapperTest.java | 60 final List<LogRecord> records = new ArrayList<>(1); in orphanedChannelsAreLogged() local 67 synchronized (records) { in orphanedChannelsAreLogged() 68 records.add(record); in orphanedChannelsAreLogged() 92 synchronized (records) { in orphanedChannelsAreLogged() 93 assertEquals(1, records.size()); in orphanedChannelsAreLogged() 94 lr = records.get(0); in orphanedChannelsAreLogged()
|
/external/tensorflow/tensorflow/core/api_def/base_api/ |
D | api_def_DecodeCSV.pbtxt | 4 name: "records" 6 Each string is a record/row in the csv and all records should have 21 Each tensor will have the same shape as records. 44 summary: "Convert CSV records to tensors. Each column maps to one tensor." 46 RFC 4180 format is expected for the CSV records.
|
/external/python/cpython2/Tools/scripts/ |
D | logmerge.py | 63 records = digest_chunk(chunk, branch) 65 del records[-1] 66 database[len(database):] = records 132 records = [] 162 records.append((date, working_file, rev, author, text)) 163 return records
|
/external/toolchain-utils/automation/common/ |
D | events.py | 77 def __init__(self, records=None): argument 78 self._records = records or [] 110 records = [] 116 records.append(_EventRecord( 121 records.append(_EventRecord( 125 return EventHistory(records)
|
/external/tensorflow/tensorflow/python/kernel_tests/ |
D | record_input_test.py | 106 records = data_flow_ops.RecordInput( 116 yield_op = records.get_yield_op() 135 records = data_flow_ops.RecordInput( 142 yield_op = records.get_yield_op() 163 records = data_flow_ops.RecordInput( 173 yield_op = records.get_yield_op()
|
/external/swiftshader/third_party/llvm-7.0/llvm/docs/ |
D | XRayFDRFormat.rst | 96 - ``Metadata Records``: Metadata records serve many purposes. Mostly, they 99 as markers for user-defined Event Data payloads. Metadata records have 16 130 | | | Function records. | 171 | Entry_Args | ``3`` | A function entry that records arguments. | 174 Entry_Args records do not contain the arguments themselves. Instead, metadata 175 records for each of the logged args follow the function record in the stream. 181 Interspersed throughout the buffer are 16 byte Metadata records. For typically 182 instrumented binaries, they will be sparser than Function records, and they 188 The same bit field rules described for function records apply to the first byte 197 | | | Metadata records. | [all …]
|
/external/skia/tools/skqp/ |
D | download_model | 58 records = [] 61 records.append((md5, path)) 62 sys.stderr.write('Downloading %d files.\n' % len(records)) 64 for record in records:
|