/external/tensorflow/tensorflow/python/kernel_tests/ |
D | decode_compressed_op_test.py | 35 def _compress(self, bytes_in, compression_type): argument 36 if not compression_type: 38 elif compression_type == "ZLIB": 48 for compression_type in ["ZLIB", "GZIP", ""]: 52 in_bytes, compression_type=compression_type) 56 feed_dict={in_bytes: [self._compress(b"AaAA", compression_type), 57 self._compress(b"bBbb", compression_type)]}) 62 for compression_type in ["ZLIB", "GZIP", ""]: 66 in_bytes, compression_type=compression_type) 70 feed_dict={in_bytes: [self._compress(b"AaBC", compression_type)]})
|
D | record_input_test.py | 36 compression_type=tf_record.TFRecordCompressionType.NONE): argument 37 options = tf_record.TFRecordOptions(compression_type) 66 compression_type=tf_record.TFRecordCompressionType.GZIP) 74 compression_type=tf_record.TFRecordCompressionType.GZIP).get_yield_op( 85 compression_type=tf_record.TFRecordCompressionType.ZLIB) 93 compression_type=tf_record.TFRecordCompressionType.ZLIB).get_yield_op(
|
/external/tensorflow/tensorflow/core/lib/io/ |
D | record_writer.cc | 27 return options.compression_type == RecordWriterOptions::ZLIB_COMPRESSION; in IsZlibCompressed() 32 const string& compression_type) { in CreateRecordWriterOptions() argument 34 if (compression_type == "ZLIB") { in CreateRecordWriterOptions() 35 options.compression_type = io::RecordWriterOptions::ZLIB_COMPRESSION; in CreateRecordWriterOptions() 42 } else if (compression_type == compression::kGzip) { in CreateRecordWriterOptions() 43 options.compression_type = io::RecordWriterOptions::ZLIB_COMPRESSION; in CreateRecordWriterOptions() 50 } else if (compression_type != compression::kNone) { in CreateRecordWriterOptions() 51 LOG(ERROR) << "Unsupported compression_type:" << compression_type in CreateRecordWriterOptions() 75 } else if (options.compression_type == RecordWriterOptions::NONE) { in RecordWriter() 78 LOG(FATAL) << "Unspecified compression type :" << options.compression_type; in RecordWriter()
|
D | record_reader.cc | 32 const string& compression_type) { in CreateRecordReaderOptions() argument 34 if (compression_type == "ZLIB") { in CreateRecordReaderOptions() 35 options.compression_type = io::RecordReaderOptions::ZLIB_COMPRESSION; in CreateRecordReaderOptions() 42 } else if (compression_type == compression::kGzip) { in CreateRecordReaderOptions() 43 options.compression_type = io::RecordReaderOptions::ZLIB_COMPRESSION; in CreateRecordReaderOptions() 50 } else if (compression_type != compression::kNone) { in CreateRecordReaderOptions() 51 LOG(ERROR) << "Unsupported compression_type:" << compression_type in CreateRecordReaderOptions() 66 if (options.compression_type == RecordReaderOptions::ZLIB_COMPRESSION) { in RecordReader() 75 } else if (options.compression_type == RecordReaderOptions::NONE) { in RecordReader() 78 LOG(FATAL) << "Unrecognized compression type :" << options.compression_type; in RecordReader()
|
D | record_reader_writer_test.cc | 41 if (options.compression_type == io::RecordWriterOptions::ZLIB_COMPRESSION) { in GetMatchingReaderOptions() 110 options.compression_type = io::RecordWriterOptions::ZLIB_COMPRESSION; in TEST() 173 options.compression_type = io::RecordWriterOptions::ZLIB_COMPRESSION; in TEST() 186 options.compression_type = io::RecordReaderOptions::ZLIB_COMPRESSION; in TEST() 208 options.compression_type = io::RecordWriterOptions::ZLIB_COMPRESSION; in TEST()
|
/external/tensorflow/tensorflow/python/data/kernel_tests/ |
D | text_line_dataset_test.py | 49 compression_type=None): argument 63 if not compression_type: 66 elif compression_type == "GZIP": 69 elif compression_type == "ZLIB": 74 raise ValueError("Unsupported compression_type", compression_type) 78 def _testTextLineDataset(self, compression_type=None): argument 80 2, 5, crlf=True, compression_type=compression_type) 84 filenames, compression_type=compression_type).repeat(num_epochs) 121 self._testTextLineDataset(compression_type="GZIP") 124 self._testTextLineDataset(compression_type="ZLIB")
|
D | fixed_length_record_dataset_test.py | 46 def _createFiles(self, compression_type=None): argument 59 if not compression_type: 62 elif compression_type == "GZIP": 65 elif compression_type == "ZLIB": 70 raise ValueError("Unsupported compression_type", compression_type) 74 def _testFixedLengthRecordDataset(self, compression_type=None): argument 75 test_filenames = self._createFiles(compression_type=compression_type) 83 compression_type=compression_type).repeat(num_epochs) 133 self._testFixedLengthRecordDataset(compression_type="GZIP") 136 self._testFixedLengthRecordDataset(compression_type="ZLIB")
|
D | tf_record_dataset_test.py | 45 compression_type="", argument 50 filenames, compression_type).repeat(num_epochs) 121 dataset = self.dataset_fn(zlib_files, compression_type="ZLIB") 136 dataset = self.dataset_fn(gzip_files, compression_type="GZIP")
|
/external/tensorflow/tensorflow/python/data/ops/ |
D | readers.py | 41 def __init__(self, filenames, compression_type=None, buffer_size=None): argument 56 compression_type, 74 def __init__(self, filenames, compression_type=None, buffer_size=None): argument 75 wrapped = TextLineDatasetV2(filenames, compression_type, buffer_size) 91 def __init__(self, filenames, compression_type=None, buffer_size=None): argument 106 compression_type, 177 def __init__(self, filenames, compression_type=None, buffer_size=None, argument 214 self._compression_type = compression_type 219 return _TFRecordDataset(filename, compression_type, buffer_size) 233 compression_type=None, argument [all …]
|
/external/tensorflow/tensorflow/python/data/experimental/kernel_tests/serialization/ |
D | textline_dataset_serialization_test.py | 30 def _build_iterator_graph(self, test_filenames, compression_type=None): argument 32 test_filenames, compression_type=compression_type, buffer_size=10) 39 for compression_type in compression_types: 44 compression_type=compression_type) 47 lambda: self._build_iterator_graph(test_filenames, compression_type),
|
D | tf_record_dataset_serialization_test.py | 37 compression_type=None, argument 40 if compression_type == "ZLIB": 51 elif compression_type == "GZIP": 62 filenames, compression_type, 91 lambda: self._build_iterator_graph(num_epochs, compression_type="ZLIB"), 94 lambda: self._build_iterator_graph(num_epochs, compression_type="GZIP"),
|
D | csv_dataset_serialization_test.py | 54 compression_type = kwargs.get("compression_type", None) 55 if compression_type == "GZIP": 57 elif compression_type is None: 60 raise ValueError("Invalid compression type:", compression_type)
|
/external/tensorflow/tensorflow/python/lib/io/ |
D | tf_record.py | 53 compression_type=None, argument 91 self.get_compression_type_string(compression_type) 92 self.compression_type = compression_type 118 return cls.get_compression_type_string(options.compression_type) 132 self.get_compression_type_string(self.compression_type))) 171 compression_type = TFRecordOptions.get_compression_type_string(options) 174 compat.as_bytes(path), 0, compat.as_bytes(compression_type), status) 213 options = TFRecordOptions(compression_type=options)
|
/external/tensorflow/tensorflow/core/kernels/data/ |
D | reader_dataset_ops.cc | 46 string compression_type; in MakeDataset() local 48 &compression_type)); in MakeDataset() 59 if (compression_type == "ZLIB") { in MakeDataset() 61 } else if (compression_type == "GZIP") { in MakeDataset() 64 OP_REQUIRES(ctx, compression_type.empty(), in MakeDataset() 79 *output = new Dataset(ctx, std::move(filenames), compression_type, in MakeDataset() 87 const string& compression_type, in Dataset() argument 91 compression_type_(compression_type), in Dataset() 92 use_compression_(!compression_type.empty()), in Dataset() 119 Node* compression_type = nullptr; in AsGraphDefInternal() local [all …]
|
/external/bsdiff/ |
D | patch_reader.cc | 39 std::vector<CompressorType> compression_type; in Init() local 42 compression_type = {CompressorType::kBZ2, CompressorType::kBZ2, in Init() 54 compression_type.push_back(CompressorType::kBZ2); in Init() 57 compression_type.push_back(CompressorType::kBrotli); in Init() 86 ctrl_stream_ = CreateDecompressor(compression_type[0]); in Init() 87 diff_stream_ = CreateDecompressor(compression_type[1]); in Init() 88 extra_stream_ = CreateDecompressor(compression_type[2]); in Init()
|
/external/tensorflow/tensorflow/core/kernels/ |
D | tf_record_reader_op.cc | 30 TFRecordReader(const string& node_name, const string& compression_type, in TFRecordReader() argument 35 compression_type_(compression_type) {} in TFRecordReader() 89 string compression_type; in TFRecordReaderOp() local 91 context->GetAttr("compression_type", &compression_type)); in TFRecordReaderOp() 93 SetReaderFactory([this, compression_type, env]() { in TFRecordReaderOp() 94 return new TFRecordReader(name(), compression_type, env); in TFRecordReaderOp()
|
D | record_input_op.cc | 39 GETATTR(string, compression_type); in RecordInputOp() 42 OP_REQUIRES_OK(ctx, ctx->GetAttr("compression_type", &compression_type)); in RecordInputOp() 50 yopts.compression_type = compression_type; in RecordInputOp()
|
/external/tensorflow/tensorflow/python/data/experimental/kernel_tests/ |
D | csv_dataset_test.py | 41 def _setup_files(self, inputs, linebreak='\n', compression_type=None): argument 46 if compression_type is None: 49 elif compression_type == 'GZIP': 52 elif compression_type == 'ZLIB': 57 raise ValueError('Unsupported compression_type', compression_type) 105 compression_type=None, # Used for both setup and parsing argument 109 filenames = self._setup_files(inputs, linebreak, compression_type) 110 kwargs['compression_type'] = compression_type 423 compression_type=None, argument 431 compression_type=compression_type, [all …]
|
D | tf_record_writer_test.py | 41 def writer_fn(self, filename, compression_type=""): argument 42 input_dataset = readers.TFRecordDataset([filename], compression_type) 44 compression_type).write(input_dataset) 71 self.writer_fn(self._createFile(options), compression_type="ZLIB")) 79 self.writer_fn(self._createFile(options), compression_type="GZIP"))
|
D | make_csv_dataset_test.py | 44 def _setup_files(self, inputs, linebreak="\n", compression_type=None): argument 49 if compression_type is None: 52 elif compression_type == "GZIP": 55 elif compression_type == "ZLIB": 60 raise ValueError("Unsupported compression_type", compression_type) 119 inputs, compression_type=kwargs.get("compression_type", None)) 209 for compression_type in ("GZIP", "ZLIB"): 221 compression_type=compression_type,
|
/external/tensorflow/tensorflow/contrib/data/python/ops/ |
D | readers.py | 52 compression_type=None, argument 142 compression_type) 152 compression_type=None, argument 160 filenames, record_defaults, compression_type, buffer_size, header,
|
D | writers.py | 29 def __init__(self, filename, compression_type=None): argument 30 super(TFRecordWriter, self).__init__(filename, compression_type)
|
/external/tensorflow/tensorflow/core/kernels/data/experimental/ |
D | csv_dataset_op.cc | 42 string compression_type; in MakeDataset() local 44 &compression_type)); in MakeDataset() 100 if (compression_type == "ZLIB") { in MakeDataset() 102 } else if (compression_type == "GZIP") { in MakeDataset() 105 OP_REQUIRES(ctx, compression_type.empty(), in MakeDataset() 107 "Unsupported compression_type: ", compression_type, ".")); in MakeDataset() 129 std::move(compression_type), zlib_compression_options, in MakeDataset() 139 string compression_type, io::ZlibCompressionOptions options, in Dataset() argument 154 use_compression_(!compression_type.empty()), in Dataset() 155 compression_type_(std::move(compression_type)), in Dataset() [all …]
|
D | to_tf_record_op.cc | 53 string compression_type; in ComputeAsync() local 56 &compression_type), in ComputeAsync() 64 compression_type)); in ComputeAsync()
|
/external/tensorflow/tensorflow/python/data/experimental/ops/ |
D | readers.py | 331 compression_type=None, argument 474 compression_type=compression_type 539 compression_type=None, argument 547 compression_type, ignore_errors)) 561 compression_type=None, argument 641 compression_type, 679 compression_type=self._compression_type) 695 compression_type=None, argument 702 wrapped = CsvDatasetV2(filenames, record_defaults, compression_type,
|