Home
last modified time | relevance | path

Searched refs:records (Results 1 – 25 of 683) sorted by relevance

12345678910>>...28

/external/ltp/testcases/realtime/lib/
Dlibstats.c75 data->records = calloc(size, sizeof(stats_record_t)); in stats_container_init()
76 if (!data->records) in stats_container_init()
90 data->records[myindex] = rec; in stats_container_append()
97 realloc(data->records, size * sizeof(stats_record_t)); in stats_container_resize()
100 data->records = newrecords; in stats_container_resize()
102 memset(data->records + data->size, 0, size - data->size); in stats_container_resize()
109 free(data->records); in stats_container_free()
116 qsort(data->records, data->index + 1, sizeof(stats_record_t), in stats_sort()
133 sum += data->records[i].y; in stats_stddev()
140 delta = (data->records[i].y - avg); in stats_stddev()
[all …]
/external/selinux/libsemanage/src/
Ddatabase_activedb.c39 record_t **records = NULL; in dbase_activedb_cache() local
53 if (ratable->read_list(handle, &records, &rcount) < 0) in dbase_activedb_cache()
58 if (dbase_llist_cache_prepend(handle, &dbase->llist, records[i]) in dbase_activedb_cache()
61 rtable->free(records[i]); in dbase_activedb_cache()
64 free(records); in dbase_activedb_cache()
70 rtable->free(records[i]); in dbase_activedb_cache()
72 free(records); in dbase_activedb_cache()
83 record_t **records = NULL; in dbase_activedb_flush() local
92 if (dbase_llist_list(handle, &dbase->llist, &records, &rcount) < 0) in dbase_activedb_flush()
96 if (ratable->commit_list(handle, records, rcount) < 0) in dbase_activedb_flush()
[all …]
Dpolicy_components.c16 record_t ** records, in clear_obsolete() argument
31 if (rtable->key_extract(handle, records[i], &key) < 0) in clear_obsolete()
41 rtable->free(records[i]); in clear_obsolete()
42 records[i] = NULL; in clear_obsolete()
61 record_t ** records, unsigned int nrecords, int mode) in load_records() argument
74 if (!records[i]) in load_records()
77 if (rtable->key_extract(handle, records[i], &rkey) < 0) in load_records()
81 dtable->set(handle, dbase, rkey, records[i]) < 0) in load_records()
85 dtable->modify(handle, dbase, rkey, records[i]) < 0) in load_records()
151 record_t **records = NULL; in semanage_base_merge_components() local
[all …]
Dusers_local.c29 semanage_seuser_t **records; in lookup_seuser() local
39 &records, in lookup_seuser()
42 sename = semanage_seuser_get_sename(records[i]); in lookup_seuser()
46 sename, semanage_seuser_get_name(records[i])); in lookup_seuser()
51 semanage_seuser_free(records[i]); in lookup_seuser()
52 free(records); in lookup_seuser()
104 semanage_user_t *** records, unsigned int *count) in semanage_user_list_local() argument
108 return dbase_list(handle, dconfig, records, count); in semanage_user_list_local()
Dfcontexts_policy.c48 semanage_fcontext_t *** records, unsigned int *count) in semanage_fcontext_list() argument
52 return dbase_list(handle, dconfig, records, count); in semanage_fcontext_list()
56 semanage_fcontext_t *** records, unsigned int *count) in semanage_fcontext_list_homedirs() argument
60 return dbase_list(handle, dconfig, records, count); in semanage_fcontext_list_homedirs()
Dsemanageswig_python.i190 semanage_bool_t*** records,
233 semanage_fcontext_t*** records,
277 semanage_iface_t*** records,
321 semanage_seuser_t*** records,
364 semanage_user_t*** records,
407 semanage_port_t*** records,
450 semanage_ibpkey_t*** records,
493 semanage_ibendport_t*** records,
536 semanage_node_t*** records,
/external/mp4parser/isoparser/src/main/java/com/googlecode/mp4parser/boxes/piff/
DPlayReadyHeader.java41 private List<PlayReadyRecord> records; field in PlayReadyHeader
61 records = PlayReadyRecord.createFor(byteBuffer, recordCount); in parse()
68 for (PlayReadyRecord record : records) { in getData()
75 IsoTypeWriter.writeUInt16BE(byteBuffer, records.size()); in getData()
76 for (PlayReadyRecord record : records) { in getData()
86 public void setRecords(List<PlayReadyRecord> records) { in setRecords() argument
87 this.records = records; in setRecords()
91 return Collections.unmodifiableList(records); in getRecords()
99 sb.append(", recordCount=").append(records.size()); in toString()
100 sb.append(", records=").append(records); in toString()
[all …]
/external/tensorflow/tensorflow/core/framework/
Dtracking_allocator_test.cc85 auto records = ta->GetRecordsAndUnRef(); in TEST() local
86 EXPECT_EQ(4, records[0].alloc_bytes); in TEST()
87 EXPECT_EQ(12, records[1].alloc_bytes); in TEST()
109 records = ta->GetRecordsAndUnRef(); in TEST()
110 EXPECT_LE(4, records[0].alloc_bytes); in TEST()
111 EXPECT_GE(-4, records[1].alloc_bytes); in TEST()
112 EXPECT_LE(12, records[2].alloc_bytes); in TEST()
113 EXPECT_GE(-12, records[3].alloc_bytes); in TEST()
135 auto records = ta->GetRecordsAndUnRef(); in TEST() local
136 EXPECT_EQ(12, records[0].alloc_bytes); in TEST()
[all …]
/external/python/cpython3/Lib/test/
Dtest_urllibnet.py193 records = []
196 records.append((blocks, block_size, total_size))
202 records_repr = repr(records) # For use in error messages.
203 self.assertGreater(len(records), 1, msg="There should always be two "
205 self.assertEqual(records[0][0], 0)
206 self.assertGreater(records[0][1], 0,
208 self.assertEqual(records[0][2], expected_size)
209 self.assertEqual(records[-1][2], expected_size)
211 block_sizes = {block_size for _, block_size, _ in records}
212 self.assertEqual({records[0][1]}, block_sizes,
[all …]
/external/tensorflow/tensorflow/tools/graph_transforms/
Dfreeze_requantization_ranges_test.cc42 std::vector<MinMaxRecord>* records);
179 std::vector<MinMaxRecord> records; in TestExtractMinMaxRecords() local
180 TF_ASSERT_OK(ExtractMinMaxRecords(min_max_log_file_name, &records)); in TestExtractMinMaxRecords()
181 ASSERT_EQ(2, records.size()); in TestExtractMinMaxRecords()
182 EXPECT_EQ("requantization_range_op", records[0].name); in TestExtractMinMaxRecords()
183 EXPECT_NEAR(-2.4313571f, records[0].min, 1e-5f); in TestExtractMinMaxRecords()
184 EXPECT_NEAR(10.584145f, records[0].max, 1e-5f); in TestExtractMinMaxRecords()
185 EXPECT_EQ("other_requantization_range_op", records[1].name); in TestExtractMinMaxRecords()
186 EXPECT_NEAR(-1.0f, records[1].min, 1e-5f); in TestExtractMinMaxRecords()
187 EXPECT_NEAR(2.0f, records[1].max, 1e-5f); in TestExtractMinMaxRecords()
Dfreeze_requantization_ranges.cc33 std::vector<MinMaxRecord>* records) { in ExtractMinMaxRecords() argument
97 records->push_back({name, min, max}); in ExtractMinMaxRecords()
121 std::vector<MinMaxRecord> records; in FreezeRequantizationRanges() local
122 TF_RETURN_IF_ERROR(ExtractMinMaxRecords(min_max_log_file, &records)); in FreezeRequantizationRanges()
123 if (records.empty()) { in FreezeRequantizationRanges()
132 for (const MinMaxRecord& record : records) { in FreezeRequantizationRanges()
148 const std::vector<MinMaxRecord> records = record_info.second; in FreezeRequantizationRanges() local
151 for (const MinMaxRecord& record : records) { in FreezeRequantizationRanges()
/external/tensorflow/tensorflow/python/lib/io/
Dtf_record_test.py85 records = [self._Record(i, j) for j in range(self._num_records)]
86 fn = self._WriteRecordsToFile(records, name, options)
90 def _WriteRecordsToFile(self, records, name="tfrecord", options=None): argument
93 for r in records:
144 def _CompressionSizeDelta(self, records, options_a, options_b): argument
164 fn_a = self._WriteRecordsToFile(records, "tfrecord_a", options=options_a)
166 self.assertEqual(records, test_a, options_a)
168 fn_b = self._WriteRecordsToFile(records, "tfrecord_b", options=options_b)
170 self.assertEqual(records, test_b, options_b)
266 records = [random_record, repeated_record, random_record]
[all …]
/external/icu/icu4j/tools/misc/src/com/ibm/icu/dev/tool/layout/
DArabicCharacterData.java75 records = new Record[charCount]; in ArabicCharacterData()
80 records[recordIndex++] = new Record(character); in add()
85 if (index < 0 || index >= records.length) { in getRecord()
89 return records[index]; in getRecord()
94 return records.length; in countRecords()
114 private Record[] records; field in ArabicCharacterData
/external/python/cpython2/Lib/bsddb/test/
Dtest_thread.py73 records = 1000 variable in ConcurrentDataStoreBase
81 keys=range(self.records)
84 records_per_writer=self.records//self.writers
86 self.assertEqual(self.records,self.writers*records_per_writer)
181 records = 1000 variable in BTreeConcurrentDataStore
188 records = 1000 variable in HashConcurrentDataStore
198 records = 1000 variable in SimpleThreadedBase
209 keys=range(self.records)
212 records_per_writer=self.records//self.writers
214 self.assertEqual(self.records,self.writers*records_per_writer)
[all …]
/external/boringssl/src/ssl/test/runner/
Ddtls.go291 var records [][]byte
295 records = append(records, fragment[:n])
296 records = append(records, fragment[n:])
298 records = append(records, fragment)
300 …} else if i := len(records) - 1; len(records) > 0 && len(records[i])+len(fragment) <= maxRecordLen…
301 records[i] = append(records[i], fragment...)
304 records = append(records, append([]byte{}, fragment...))
309 for _, record := range records {
/external/ltp/testcases/realtime/func/rt-migrate/
Drt-migrate.c268 printf("%6ld ", intervals[t].records[i].y); in print_results()
273 printf("%6ld ", intervals_length[t].records[i].y); in print_results()
278 printf("%6ld ", intervals_loops[t].records[i].y); in print_results()
366 if (i && last < intervals[i].records[l].y && in check_times()
367 ((intervals[i].records[l].y - last) > max_err)) { in check_times()
374 if (intervals_loops[i].records[l].y < last_loops || in check_times()
375 intervals[i].records[l].y > last_length || in check_times()
376 (intervals_length[i].records[l].y > last_length && in check_times()
377 intervals_length[i].records[l].y - last_length > in check_times()
383 last = intervals[i].records[l].y; in check_times()
[all …]
/external/tcpdump/tests/
Dsflow_multiple_counter_30_pdus.out3 expanded counter sample (4), length 172, seqnum 87096, type 0, idx 55, records 2
15 expanded counter sample (4), length 172, seqnum 87096, type 0, idx 56, records 2
27 expanded counter sample (4), length 172, seqnum 87099, type 0, idx 57, records 2
39 expanded counter sample (4), length 172, seqnum 87096, type 0, idx 60, records 2
51 expanded counter sample (4), length 172, seqnum 87016, type 0, idx 61, records 2
63 expanded counter sample (4), length 172, seqnum 87096, type 0, idx 62, records 2
75 expanded counter sample (4), length 172, seqnum 87096, type 0, idx 63, records 2
89 expanded counter sample (4), length 172, seqnum 2007195, type 0, idx 1, records 2
101 expanded counter sample (4), length 172, seqnum 2006745, type 0, idx 2, records 2
113 expanded counter sample (4), length 172, seqnum 2007118, type 0, idx 3, records 2
[all …]
/external/syzkaller/syz-hub/state/
Dstate.go272 var records []db.Record
287 records = append(records, rec)
293 if len(records) > maxRecords {
294 sort.Sort(recordSeqSorter(records))
296 maxSeq = records[pos].Seq
297 for pos+1 < len(records) && records[pos+1].Seq == maxSeq {
301 more = len(records) - pos
302 records = records[:pos]
304 progs := make([][]byte, len(records))
305 for _, rec := range records {
/external/grpc-grpc-java/core/src/test/java/io/grpc/internal/
DManagedChannelOrphanWrapperTest.java60 final List<LogRecord> records = new ArrayList<>(1); in orphanedChannelsAreLogged() local
67 synchronized (records) { in orphanedChannelsAreLogged()
68 records.add(record); in orphanedChannelsAreLogged()
92 synchronized (records) { in orphanedChannelsAreLogged()
93 assertEquals(1, records.size()); in orphanedChannelsAreLogged()
94 lr = records.get(0); in orphanedChannelsAreLogged()
/external/tensorflow/tensorflow/core/api_def/base_api/
Dapi_def_DecodeCSV.pbtxt4 name: "records"
6 Each string is a record/row in the csv and all records should have
21 Each tensor will have the same shape as records.
44 summary: "Convert CSV records to tensors. Each column maps to one tensor."
46 RFC 4180 format is expected for the CSV records.
/external/python/cpython2/Tools/scripts/
Dlogmerge.py63 records = digest_chunk(chunk, branch)
65 del records[-1]
66 database[len(database):] = records
132 records = []
162 records.append((date, working_file, rev, author, text))
163 return records
/external/toolchain-utils/automation/common/
Devents.py77 def __init__(self, records=None): argument
78 self._records = records or []
110 records = []
116 records.append(_EventRecord(
121 records.append(_EventRecord(
125 return EventHistory(records)
/external/tensorflow/tensorflow/python/kernel_tests/
Drecord_input_test.py106 records = data_flow_ops.RecordInput(
116 yield_op = records.get_yield_op()
135 records = data_flow_ops.RecordInput(
142 yield_op = records.get_yield_op()
163 records = data_flow_ops.RecordInput(
173 yield_op = records.get_yield_op()
/external/swiftshader/third_party/llvm-7.0/llvm/docs/
DXRayFDRFormat.rst96 - ``Metadata Records``: Metadata records serve many purposes. Mostly, they
99 as markers for user-defined Event Data payloads. Metadata records have 16
130 | | | Function records. |
171 | Entry_Args | ``3`` | A function entry that records arguments. |
174 Entry_Args records do not contain the arguments themselves. Instead, metadata
175 records for each of the logged args follow the function record in the stream.
181 Interspersed throughout the buffer are 16 byte Metadata records. For typically
182 instrumented binaries, they will be sparser than Function records, and they
188 The same bit field rules described for function records apply to the first byte
197 | | | Metadata records. |
[all …]
/external/skia/tools/skqp/
Ddownload_model58 records = []
61 records.append((md5, path))
62 sys.stderr.write('Downloading %d files.\n' % len(records))
64 for record in records:

12345678910>>...28