Searched refs:date_removed (Results 1 – 9 of 9) sorted by relevance
/external/pigweed/pw_tokenizer/py/pw_tokenizer/ |
D | tokens.py | 111 date_removed: datetime | None = None, 118 self.date_removed = date_removed 139 if self.date_removed is None: 142 if new_date_removed is None or new_date_removed > self.date_removed: 143 self.date_removed = new_date_removed 148 and self.date_removed == other.date_removed 161 if self.date_removed != other.date_removed: 162 return (other.date_removed or datetime.max) < ( 163 self.date_removed or datetime.max 334 entry.date_removed is None or removal_date < entry.date_removed [all …]
|
D | decode.py | 743 def score(self, date_removed: datetime | None = None) -> tuple: 764 date_removed or datetime.max,
|
D | database.py | 238 present = [entry for entry in db.entries() if not entry.date_removed] 367 if not entry.date_removed
|
D | detokenize.py | 150 decode_attempts.append((result.score(entry.date_removed), result))
|
/external/pigweed/pw_tokenizer/py/ |
D | tokens_test.py | 211 date_removed=datetime(1998, 9, 4), 242 date_removed=datetime(1998, 9, 4), 309 self.assertEqual(jello.date_removed, datetime(2019, 6, 11)) 316 self.assertIsNone(llu.date_removed) 417 1, 'one', date_removed=datetime.min 420 2, 'two', 'domain', date_removed=datetime.min 426 self.assertEqual(db.token_to_entries[1][0].date_removed, datetime.min) 427 self.assertEqual(db.token_to_entries[2][0].date_removed, datetime.min) 435 4, 'four', date_removed=datetime.min 443 self.assertIsNone(db.token_to_entries[3][0].date_removed) [all …]
|
D | detokenize_test.py | 118 0xCDAB, '%02d %s %c%%', date_removed=dt.datetime.now() 132 1, 'no args', date_removed=dt.datetime(1, 1, 1) 162 2, '%s', date_removed=dt.datetime(1, 1, 1) 181 2, '%s', date_removed=dt.datetime(1, 1, 1) 201 1, 'no args', date_removed=dt.datetime(100, 1, 1) 387 token, 'REMOVED', date_removed=dt.datetime(9, 1, 1) 391 token, 'A: %d', date_removed=dt.datetime(30, 5, 9) 394 token, 'B: %c', date_removed=dt.datetime(30, 5, 10)
|
/external/pigweed/pw_tokenizer/ |
D | detokenize.cc | 180 for (const auto& [format, date_removed] : entries) { in DetokenizedString() 184 date_removed}); in DetokenizedString() 209 entry.date_removed); in Detokenizer() 277 const std::string& date_removed = row[1]; in FromCsv() local 295 if (!date_removed.empty() && in FromCsv() 296 date_removed.find_first_not_of(' ') != std::string::npos) { in FromCsv() 297 size_t first_dash = date_removed.find('-'); in FromCsv() 303 size_t second_dash = date_removed.find('-', first_dash + 1); in FromCsv() 310 int year = std::stoi(date_removed.substr(0, first_dash), &pos); in FromCsv() 317 date_removed.substr(first_dash + 1, second_dash - first_dash - 1), in FromCsv() [all …]
|
/external/perfetto/src/trace_processor/importers/proto/ |
D | pigweed_detokenizer.cc | 45 uint32_t date_removed; member 92 uint32_t date_removed = in CreateDetokenizer() local 93 ReadUint32(bytes.data + entry_ix + offsetof(Entry, date_removed)); in CreateDetokenizer() 104 if (date_removed == kDateRemovedNever) { in CreateDetokenizer()
|
/external/pigweed/pw_tokenizer/public/pw_tokenizer/ |
D | token_database.h | 82 uint32_t date_removed; member 110 uint32_t date_removed; member 173 entry_.date_removed = ReadUint32(raw_ + sizeof(entry_.token)); in ReadRawEntry()
|