Home
last modified time | relevance | path

Searched refs:compression_type (Results 1 – 25 of 102) sorted by relevance

12345

/external/tensorflow/tensorflow/python/kernel_tests/
Ddecode_compressed_op_test.py35 def _compress(self, bytes_in, compression_type): argument
36 if not compression_type:
38 elif compression_type == "ZLIB":
48 for compression_type in ["ZLIB", "GZIP", ""]:
52 in_bytes, compression_type=compression_type)
56 feed_dict={in_bytes: [self._compress(b"AaAA", compression_type),
57 self._compress(b"bBbb", compression_type)]})
62 for compression_type in ["ZLIB", "GZIP", ""]:
66 in_bytes, compression_type=compression_type)
70 feed_dict={in_bytes: [self._compress(b"AaBC", compression_type)]})
Drecord_input_test.py36 compression_type=tf_record.TFRecordCompressionType.NONE): argument
37 options = tf_record.TFRecordOptions(compression_type)
66 compression_type=tf_record.TFRecordCompressionType.GZIP)
74 compression_type=tf_record.TFRecordCompressionType.GZIP).get_yield_op(
85 compression_type=tf_record.TFRecordCompressionType.ZLIB)
93 compression_type=tf_record.TFRecordCompressionType.ZLIB).get_yield_op(
/external/tensorflow/tensorflow/core/lib/io/
Drecord_writer.cc27 return options.compression_type == RecordWriterOptions::ZLIB_COMPRESSION; in IsZlibCompressed()
32 const string& compression_type) { in CreateRecordWriterOptions() argument
34 if (compression_type == "ZLIB") { in CreateRecordWriterOptions()
35 options.compression_type = io::RecordWriterOptions::ZLIB_COMPRESSION; in CreateRecordWriterOptions()
42 } else if (compression_type == compression::kGzip) { in CreateRecordWriterOptions()
43 options.compression_type = io::RecordWriterOptions::ZLIB_COMPRESSION; in CreateRecordWriterOptions()
50 } else if (compression_type != compression::kNone) { in CreateRecordWriterOptions()
51 LOG(ERROR) << "Unsupported compression_type:" << compression_type in CreateRecordWriterOptions()
75 } else if (options.compression_type == RecordWriterOptions::NONE) { in RecordWriter()
78 LOG(FATAL) << "Unspecified compression type :" << options.compression_type; in RecordWriter()
Drecord_reader.cc32 const string& compression_type) { in CreateRecordReaderOptions() argument
34 if (compression_type == "ZLIB") { in CreateRecordReaderOptions()
35 options.compression_type = io::RecordReaderOptions::ZLIB_COMPRESSION; in CreateRecordReaderOptions()
42 } else if (compression_type == compression::kGzip) { in CreateRecordReaderOptions()
43 options.compression_type = io::RecordReaderOptions::ZLIB_COMPRESSION; in CreateRecordReaderOptions()
50 } else if (compression_type != compression::kNone) { in CreateRecordReaderOptions()
51 LOG(ERROR) << "Unsupported compression_type:" << compression_type in CreateRecordReaderOptions()
66 if (options.compression_type == RecordReaderOptions::ZLIB_COMPRESSION) { in RecordReader()
75 } else if (options.compression_type == RecordReaderOptions::NONE) { in RecordReader()
78 LOG(FATAL) << "Unrecognized compression type :" << options.compression_type; in RecordReader()
Drecord_reader_writer_test.cc41 if (options.compression_type == io::RecordWriterOptions::ZLIB_COMPRESSION) { in GetMatchingReaderOptions()
110 options.compression_type = io::RecordWriterOptions::ZLIB_COMPRESSION; in TEST()
173 options.compression_type = io::RecordWriterOptions::ZLIB_COMPRESSION; in TEST()
186 options.compression_type = io::RecordReaderOptions::ZLIB_COMPRESSION; in TEST()
208 options.compression_type = io::RecordWriterOptions::ZLIB_COMPRESSION; in TEST()
/external/tensorflow/tensorflow/python/data/kernel_tests/
Dtext_line_dataset_test.py49 compression_type=None): argument
63 if not compression_type:
66 elif compression_type == "GZIP":
69 elif compression_type == "ZLIB":
74 raise ValueError("Unsupported compression_type", compression_type)
78 def _testTextLineDataset(self, compression_type=None): argument
80 2, 5, crlf=True, compression_type=compression_type)
84 filenames, compression_type=compression_type).repeat(num_epochs)
121 self._testTextLineDataset(compression_type="GZIP")
124 self._testTextLineDataset(compression_type="ZLIB")
Dfixed_length_record_dataset_test.py46 def _createFiles(self, compression_type=None): argument
59 if not compression_type:
62 elif compression_type == "GZIP":
65 elif compression_type == "ZLIB":
70 raise ValueError("Unsupported compression_type", compression_type)
74 def _testFixedLengthRecordDataset(self, compression_type=None): argument
75 test_filenames = self._createFiles(compression_type=compression_type)
83 compression_type=compression_type).repeat(num_epochs)
133 self._testFixedLengthRecordDataset(compression_type="GZIP")
136 self._testFixedLengthRecordDataset(compression_type="ZLIB")
Dtf_record_dataset_test.py45 compression_type="", argument
50 filenames, compression_type).repeat(num_epochs)
121 dataset = self.dataset_fn(zlib_files, compression_type="ZLIB")
136 dataset = self.dataset_fn(gzip_files, compression_type="GZIP")
/external/tensorflow/tensorflow/python/data/ops/
Dreaders.py41 def __init__(self, filenames, compression_type=None, buffer_size=None): argument
56 compression_type,
74 def __init__(self, filenames, compression_type=None, buffer_size=None): argument
75 wrapped = TextLineDatasetV2(filenames, compression_type, buffer_size)
91 def __init__(self, filenames, compression_type=None, buffer_size=None): argument
106 compression_type,
177 def __init__(self, filenames, compression_type=None, buffer_size=None, argument
214 self._compression_type = compression_type
219 return _TFRecordDataset(filename, compression_type, buffer_size)
233 compression_type=None, argument
[all …]
/external/tensorflow/tensorflow/python/data/experimental/kernel_tests/serialization/
Dtextline_dataset_serialization_test.py30 def _build_iterator_graph(self, test_filenames, compression_type=None): argument
32 test_filenames, compression_type=compression_type, buffer_size=10)
39 for compression_type in compression_types:
44 compression_type=compression_type)
47 lambda: self._build_iterator_graph(test_filenames, compression_type),
Dtf_record_dataset_serialization_test.py37 compression_type=None, argument
40 if compression_type == "ZLIB":
51 elif compression_type == "GZIP":
62 filenames, compression_type,
91 lambda: self._build_iterator_graph(num_epochs, compression_type="ZLIB"),
94 lambda: self._build_iterator_graph(num_epochs, compression_type="GZIP"),
Dcsv_dataset_serialization_test.py54 compression_type = kwargs.get("compression_type", None)
55 if compression_type == "GZIP":
57 elif compression_type is None:
60 raise ValueError("Invalid compression type:", compression_type)
/external/tensorflow/tensorflow/python/lib/io/
Dtf_record.py53 compression_type=None, argument
91 self.get_compression_type_string(compression_type)
92 self.compression_type = compression_type
118 return cls.get_compression_type_string(options.compression_type)
132 self.get_compression_type_string(self.compression_type)))
171 compression_type = TFRecordOptions.get_compression_type_string(options)
174 compat.as_bytes(path), 0, compat.as_bytes(compression_type), status)
213 options = TFRecordOptions(compression_type=options)
/external/tensorflow/tensorflow/core/kernels/data/
Dreader_dataset_ops.cc46 string compression_type; in MakeDataset() local
48 &compression_type)); in MakeDataset()
59 if (compression_type == "ZLIB") { in MakeDataset()
61 } else if (compression_type == "GZIP") { in MakeDataset()
64 OP_REQUIRES(ctx, compression_type.empty(), in MakeDataset()
79 *output = new Dataset(ctx, std::move(filenames), compression_type, in MakeDataset()
87 const string& compression_type, in Dataset() argument
91 compression_type_(compression_type), in Dataset()
92 use_compression_(!compression_type.empty()), in Dataset()
119 Node* compression_type = nullptr; in AsGraphDefInternal() local
[all …]
/external/bsdiff/
Dpatch_reader.cc39 std::vector<CompressorType> compression_type; in Init() local
42 compression_type = {CompressorType::kBZ2, CompressorType::kBZ2, in Init()
54 compression_type.push_back(CompressorType::kBZ2); in Init()
57 compression_type.push_back(CompressorType::kBrotli); in Init()
86 ctrl_stream_ = CreateDecompressor(compression_type[0]); in Init()
87 diff_stream_ = CreateDecompressor(compression_type[1]); in Init()
88 extra_stream_ = CreateDecompressor(compression_type[2]); in Init()
/external/tensorflow/tensorflow/core/kernels/
Dtf_record_reader_op.cc30 TFRecordReader(const string& node_name, const string& compression_type, in TFRecordReader() argument
35 compression_type_(compression_type) {} in TFRecordReader()
89 string compression_type; in TFRecordReaderOp() local
91 context->GetAttr("compression_type", &compression_type)); in TFRecordReaderOp()
93 SetReaderFactory([this, compression_type, env]() { in TFRecordReaderOp()
94 return new TFRecordReader(name(), compression_type, env); in TFRecordReaderOp()
Drecord_input_op.cc39 GETATTR(string, compression_type); in RecordInputOp()
42 OP_REQUIRES_OK(ctx, ctx->GetAttr("compression_type", &compression_type)); in RecordInputOp()
50 yopts.compression_type = compression_type; in RecordInputOp()
/external/tensorflow/tensorflow/python/data/experimental/kernel_tests/
Dcsv_dataset_test.py41 def _setup_files(self, inputs, linebreak='\n', compression_type=None): argument
46 if compression_type is None:
49 elif compression_type == 'GZIP':
52 elif compression_type == 'ZLIB':
57 raise ValueError('Unsupported compression_type', compression_type)
105 compression_type=None, # Used for both setup and parsing argument
109 filenames = self._setup_files(inputs, linebreak, compression_type)
110 kwargs['compression_type'] = compression_type
423 compression_type=None, argument
431 compression_type=compression_type,
[all …]
Dtf_record_writer_test.py41 def writer_fn(self, filename, compression_type=""): argument
42 input_dataset = readers.TFRecordDataset([filename], compression_type)
44 compression_type).write(input_dataset)
71 self.writer_fn(self._createFile(options), compression_type="ZLIB"))
79 self.writer_fn(self._createFile(options), compression_type="GZIP"))
Dmake_csv_dataset_test.py44 def _setup_files(self, inputs, linebreak="\n", compression_type=None): argument
49 if compression_type is None:
52 elif compression_type == "GZIP":
55 elif compression_type == "ZLIB":
60 raise ValueError("Unsupported compression_type", compression_type)
119 inputs, compression_type=kwargs.get("compression_type", None))
209 for compression_type in ("GZIP", "ZLIB"):
221 compression_type=compression_type,
/external/tensorflow/tensorflow/contrib/data/python/ops/
Dreaders.py52 compression_type=None, argument
142 compression_type)
152 compression_type=None, argument
160 filenames, record_defaults, compression_type, buffer_size, header,
Dwriters.py29 def __init__(self, filename, compression_type=None): argument
30 super(TFRecordWriter, self).__init__(filename, compression_type)
/external/tensorflow/tensorflow/core/kernels/data/experimental/
Dcsv_dataset_op.cc42 string compression_type; in MakeDataset() local
44 &compression_type)); in MakeDataset()
100 if (compression_type == "ZLIB") { in MakeDataset()
102 } else if (compression_type == "GZIP") { in MakeDataset()
105 OP_REQUIRES(ctx, compression_type.empty(), in MakeDataset()
107 "Unsupported compression_type: ", compression_type, ".")); in MakeDataset()
129 std::move(compression_type), zlib_compression_options, in MakeDataset()
139 string compression_type, io::ZlibCompressionOptions options, in Dataset() argument
154 use_compression_(!compression_type.empty()), in Dataset()
155 compression_type_(std::move(compression_type)), in Dataset()
[all …]
Dto_tf_record_op.cc53 string compression_type; in ComputeAsync() local
56 &compression_type), in ComputeAsync()
64 compression_type)); in ComputeAsync()
/external/tensorflow/tensorflow/python/data/experimental/ops/
Dreaders.py331 compression_type=None, argument
474 compression_type=compression_type
539 compression_type=None, argument
547 compression_type, ignore_errors))
561 compression_type=None, argument
641 compression_type,
679 compression_type=self._compression_type)
695 compression_type=None, argument
702 wrapped = CsvDatasetV2(filenames, record_defaults, compression_type,

12345