• Home
  • Line#
  • Scopes#
  • Navigate#
  • Raw
  • Download
1 /* Copyright 2017 The TensorFlow Authors. All Rights Reserved.
2 
3 Licensed under the Apache License, Version 2.0 (the "License");
4 you may not use this file except in compliance with the License.
5 You may obtain a copy of the License at
6 
7     http://www.apache.org/licenses/LICENSE-2.0
8 
9 Unless required by applicable law or agreed to in writing, software
10 distributed under the License is distributed on an "AS IS" BASIS,
11 WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
12 See the License for the specific language governing permissions and
13 limitations under the License.
14 ==============================================================================*/
15 
16 #ifndef TENSORFLOW_COMPILER_XLA_SERVICE_BACKEND_H_
17 #define TENSORFLOW_COMPILER_XLA_SERVICE_BACKEND_H_
18 
19 #include <map>
20 #include <memory>
21 #include <set>
22 #include <string>
23 #include <vector>
24 
25 #include "absl/container/flat_hash_map.h"
26 #include "absl/strings/str_cat.h"
27 #include "absl/types/span.h"
28 #include "tensorflow/compiler/xla/service/compiler.h"
29 #include "tensorflow/compiler/xla/service/computation_placer.h"
30 #include "tensorflow/compiler/xla/service/device_memory_allocator.h"
31 #include "tensorflow/compiler/xla/service/stream_pool.h"
32 #include "tensorflow/compiler/xla/service/transfer_manager.h"
33 #include "tensorflow/compiler/xla/statusor.h"
34 #include "tensorflow/compiler/xla/types.h"
35 #include "tensorflow/core/platform/mutex.h"
36 #include "tensorflow/core/platform/stream_executor_no_cuda.h"
37 #include "tensorflow/core/platform/thread_annotations.h"
38 
39 namespace Eigen {
40 struct ThreadPoolDevice;
41 }
42 
43 namespace xla {
44 
45 // Options to configure the backend when it is created.
46 class BackendOptions {
47  public:
48   // Set the platform backing the backend, or nullptr for the default platform.
49   BackendOptions& set_platform(se::Platform* platform);
50   se::Platform* platform() const;
51 
52   // Sets the thread pool size for parallel execution of an individual operator.
53   // The default value of -1 will result in initializing the thread pool with
54   // the number of threads equal to the number of cores in the system.
55   BackendOptions& set_intra_op_parallelism_threads(int num_threads);
56   int intra_op_parallelism_threads() const;
57 
58   // Sets the allowed_devices for selectively constructing stream executors
59   // on the platform.
60   BackendOptions& set_allowed_devices(
61       const absl::optional<std::set<int>>& allowed_devices);
62   const absl::optional<std::set<int>>& allowed_devices() const;
63 
64  private:
65   se::Platform* platform_ = nullptr;
66   int intra_op_parallelism_threads_ = -1;
67   absl::optional<std::set<int>> allowed_devices_;
68 };
69 
70 // Class which encapsulates an XLA backend. It includes everything necessary
71 // to compile and execute computations on a particular platform.
72 //
73 // It also offers a pooling API for creation/use of initialized streams:
74 //
75 //    StreamPool::Ptr stream = backend->BorrowStream().ConsumeValueOrDie();
76 class Backend {
77  public:
78   // Creates a new backend.
79   static StatusOr<std::unique_ptr<Backend>> CreateBackend(
80       const BackendOptions& options);
81 
82   // Creates a backend for the default platform. The default platform is defined
83   // in PlatformUtil.
84   static StatusOr<std::unique_ptr<Backend>> CreateDefaultBackend();
85 
86   ~Backend();
87 
88   // Accessors for the various objects.
platform()89   se::Platform* platform() const { return platform_; }
compiler()90   Compiler* compiler() const { return compiler_; }
memory_allocator()91   DeviceMemoryAllocator* memory_allocator() const {
92     return memory_allocator_.get();
93   }
transfer_manager()94   TransferManager* transfer_manager() const { return transfer_manager_; }
computation_placer()95   ComputationPlacer* computation_placer() const { return computation_placer_; }
96 
97   // Returns the number of devices of the platform type which are visible. Not
98   // all of these devices may be usable by XLA.
device_count()99   int device_count() const { return stream_executors_.size(); }
100 
101   // Returns the device ordinal number of the default device.
102   int default_device_ordinal() const;
103 
104   // Returns stream executors of all supported devices for this backend. The
105   // executors are ordered by the device ordinal.
stream_executors()106   const std::vector<se::StreamExecutor*>& stream_executors() const {
107     return stream_executors_;
108   }
109 
110   // Returns the stream executor for the given device ordinal.
111   StatusOr<se::StreamExecutor*> stream_executor(int device_ordinal) const;
112 
113   // Returns the stream executor for the default device ordinal. This stream
114   // executor can only be used when the number of computations is 1 (replication
115   // can be > 1).
default_stream_executor()116   se::StreamExecutor* default_stream_executor() const {
117     CHECK(!stream_executors_.empty());
118     return stream_executors_[0];
119   }
120 
121   // Borrows a stream for use by the caller, either by grabbing it from an
122   // internal pool, or by constructing/initializating it, and returns the result
123   // to the caller.
124   StatusOr<StreamPool::Ptr> BorrowStream(int device_ordinal);
125   StatusOr<StreamPool::Ptr> BorrowStream(se::StreamExecutor* executor);
126 
127   // Returns a function to borrow a stream, as `BorrowStream` above does.
128   // Purely for convenience, the caller could rather make this anonymous
129   // function itself.
StreamBorrower()130   std::function<StatusOr<StreamPool::Ptr>(int)> StreamBorrower() {
131     return [this](int device_ordinal) { return BorrowStream(device_ordinal); };
132   }
133 
134   // Returns whether the given device ordinal of the backend is supported.
device_ordinal_supported(int device_ordinal)135   bool device_ordinal_supported(int device_ordinal) const {
136     return (device_ordinal >= 0 && device_ordinal < device_count() &&
137             stream_executors_[device_ordinal] != nullptr);
138   }
139 
140   // Return a string identifier for the given device, eg: "GPU:3".
device_name(int device_ordinal)141   string device_name(int device_ordinal) const {
142     return absl::StrCat(platform_->Name(), ":", device_ordinal);
143   }
144 
145   // Returns true if the devices with the given ordinals are equivalent from
146   // XLA's perspective. That is, an executable compiled for one device would
147   // be equivalent to an executable compiled for the other.
148   StatusOr<bool> devices_equivalent(int device_ordinal_a, int device_ordinal_b);
149 
150   // For the host platform, returns the configured eigen threadpool device to be
151   // used for scheduling work. For other platforms, returns NULL.
152   const Eigen::ThreadPoolDevice* eigen_intra_op_thread_pool_device() const;
153   tensorflow::thread::ThreadPool* eigen_intra_op_thread_pool() const;
154 
155   // Resets the devices associated with this backend.
156   Status ResetDevices();
157 
158  private:
159   Backend(se::Platform* platform, Compiler* compiler,
160           absl::Span<se::StreamExecutor* const> stream_executors,
161           TransferManager* transfer_manager,
162           ComputationPlacer* computation_placer,
163           int intra_op_parallelism_threads);
164   Backend(const Backend&) = delete;
165   Backend& operator=(const Backend&) = delete;
166 
167   se::Platform* platform_;
168   Compiler* compiler_;
169   TransferManager* transfer_manager_;
170   ComputationPlacer* computation_placer_;
171 
172   // Vector of stream executors. stream_executors_[0] is the default executor.
173   std::vector<se::StreamExecutor*> stream_executors_;
174 
175   tensorflow::mutex mu_;
176 
177   // Mapping from stream executor to stream pools, used by `BorrowStream` above.
178   absl::flat_hash_map<se::StreamExecutor*, std::unique_ptr<StreamPool>>
179       stream_pools_ GUARDED_BY(mu_);
180 
181   // The default memory allocator to use.
182   std::unique_ptr<StreamExecutorMemoryAllocator> memory_allocator_;
183 
184   // For the CPU backend, an Eigen threadpool device for use by Eigen code.
185   struct IntraOpThreadPool;
186   std::unique_ptr<IntraOpThreadPool> intra_op_thread_pool_;
187 };
188 
189 }  // namespace xla
190 
191 #endif  // TENSORFLOW_COMPILER_XLA_SERVICE_BACKEND_H_
192