/home/arjun/llvm-project/llvm/include/llvm/Support/Threading.h
Line | Count | Source (jump to first uncovered line) |
1 | | //===-- llvm/Support/Threading.h - Control multithreading mode --*- C++ -*-===// |
2 | | // |
3 | | // Part of the LLVM Project, under the Apache License v2.0 with LLVM Exceptions. |
4 | | // See https://llvm.org/LICENSE.txt for license information. |
5 | | // SPDX-License-Identifier: Apache-2.0 WITH LLVM-exception |
6 | | // |
7 | | //===----------------------------------------------------------------------===// |
8 | | // |
9 | | // This file declares helper functions for running LLVM in a multi-threaded |
10 | | // environment. |
11 | | // |
12 | | //===----------------------------------------------------------------------===// |
13 | | |
14 | | #ifndef LLVM_SUPPORT_THREADING_H |
15 | | #define LLVM_SUPPORT_THREADING_H |
16 | | |
17 | | #include "llvm/ADT/BitVector.h" |
18 | | #include "llvm/ADT/FunctionExtras.h" |
19 | | #include "llvm/ADT/SmallVector.h" |
20 | | #include "llvm/Config/llvm-config.h" // for LLVM_ON_UNIX |
21 | | #include "llvm/Support/Compiler.h" |
22 | | #include <ciso646> // So we can check the C++ standard lib macros. |
23 | | #include <functional> |
24 | | |
25 | | #if defined(_MSC_VER) |
26 | | // MSVC's call_once implementation worked since VS 2015, which is the minimum |
27 | | // supported version as of this writing. |
28 | | #define LLVM_THREADING_USE_STD_CALL_ONCE 1 |
29 | | #elif defined(LLVM_ON_UNIX) && \ |
30 | | (defined(_LIBCPP_VERSION) || \ |
31 | | !(defined(__NetBSD__) || defined(__OpenBSD__) || \ |
32 | | (defined(__ppc__) || defined(__PPC__)))) |
33 | | // std::call_once from libc++ is used on all Unix platforms. Other |
34 | | // implementations like libstdc++ are known to have problems on NetBSD, |
35 | | // OpenBSD and PowerPC. |
36 | | #define LLVM_THREADING_USE_STD_CALL_ONCE 1 |
37 | | #elif defined(LLVM_ON_UNIX) && \ |
38 | | ((defined(__ppc__) || defined(__PPC__)) && defined(__LITTLE_ENDIAN__)) |
39 | | #define LLVM_THREADING_USE_STD_CALL_ONCE 1 |
40 | | #else |
41 | | #define LLVM_THREADING_USE_STD_CALL_ONCE 0 |
42 | | #endif |
43 | | |
44 | | #if LLVM_THREADING_USE_STD_CALL_ONCE |
45 | | #include <mutex> |
46 | | #else |
47 | | #include "llvm/Support/Atomic.h" |
48 | | #endif |
49 | | |
50 | | namespace llvm { |
51 | | class Twine; |
52 | | |
53 | | /// Returns true if LLVM is compiled with support for multi-threading, and |
54 | | /// false otherwise. |
55 | | bool llvm_is_multithreaded(); |
56 | | |
57 | | /// Execute the given \p UserFn on a separate thread, passing it the provided \p |
58 | | /// UserData and waits for thread completion. |
59 | | /// |
60 | | /// This function does not guarantee that the code will actually be executed |
61 | | /// on a separate thread or honoring the requested stack size, but tries to do |
62 | | /// so where system support is available. |
63 | | /// |
64 | | /// \param UserFn - The callback to execute. |
65 | | /// \param UserData - An argument to pass to the callback function. |
66 | | /// \param StackSizeInBytes - A requested size (in bytes) for the thread stack |
67 | | /// (or None for default) |
68 | | void llvm_execute_on_thread( |
69 | | void (*UserFn)(void *), void *UserData, |
70 | | llvm::Optional<unsigned> StackSizeInBytes = llvm::None); |
71 | | |
72 | | /// Schedule the given \p Func for execution on a separate thread, then return |
73 | | /// to the caller immediately. Roughly equivalent to |
74 | | /// `std::thread(Func).detach()`, except it allows requesting a specific stack |
75 | | /// size, if supported for the platform. |
76 | | /// |
77 | | /// This function would report a fatal error if it can't execute the code |
78 | | /// on a separate thread. |
79 | | /// |
80 | | /// \param Func - The callback to execute. |
81 | | /// \param StackSizeInBytes - A requested size (in bytes) for the thread stack |
82 | | /// (or None for default) |
83 | | void llvm_execute_on_thread_async( |
84 | | llvm::unique_function<void()> Func, |
85 | | llvm::Optional<unsigned> StackSizeInBytes = llvm::None); |
86 | | |
87 | | #if LLVM_THREADING_USE_STD_CALL_ONCE |
88 | | |
89 | | typedef std::once_flag once_flag; |
90 | | |
91 | | #else |
92 | | |
93 | | enum InitStatus { Uninitialized = 0, Wait = 1, Done = 2 }; |
94 | | |
95 | | /// The llvm::once_flag structure |
96 | | /// |
97 | | /// This type is modeled after std::once_flag to use with llvm::call_once. |
98 | | /// This structure must be used as an opaque object. It is a struct to force |
99 | | /// autoinitialization and behave like std::once_flag. |
100 | | struct once_flag { |
101 | | volatile sys::cas_flag status = Uninitialized; |
102 | | }; |
103 | | |
104 | | #endif |
105 | | |
106 | | /// Execute the function specified as a parameter once. |
107 | | /// |
108 | | /// Typical usage: |
109 | | /// \code |
110 | | /// void foo() {...}; |
111 | | /// ... |
112 | | /// static once_flag flag; |
113 | | /// call_once(flag, foo); |
114 | | /// \endcode |
115 | | /// |
116 | | /// \param flag Flag used for tracking whether or not this has run. |
117 | | /// \param F Function to call once. |
118 | | template <typename Function, typename... Args> |
119 | 8 | void call_once(once_flag &flag, Function &&F, Args &&... ArgList) { |
120 | 8 | #if LLVM_THREADING_USE_STD_CALL_ONCE |
121 | 8 | std::call_once(flag, std::forward<Function>(F), |
122 | 8 | std::forward<Args>(ArgList)...); |
123 | | #else |
124 | | // For other platforms we use a generic (if brittle) version based on our |
125 | | // atomics. |
126 | | sys::cas_flag old_val = sys::CompareAndSwap(&flag.status, Wait, Uninitialized); |
127 | | if (old_val == Uninitialized) { |
128 | | std::forward<Function>(F)(std::forward<Args>(ArgList)...); |
129 | | sys::MemoryFence(); |
130 | | TsanIgnoreWritesBegin(); |
131 | | TsanHappensBefore(&flag.status); |
132 | | flag.status = Done; |
133 | | TsanIgnoreWritesEnd(); |
134 | | } else { |
135 | | // Wait until any thread doing the call has finished. |
136 | | sys::cas_flag tmp = flag.status; |
137 | | sys::MemoryFence(); |
138 | | while (tmp != Done) { |
139 | | tmp = flag.status; |
140 | | sys::MemoryFence(); |
141 | | } |
142 | | } |
143 | | TsanHappensAfter(&flag.status); |
144 | | #endif |
145 | | } |
146 | | |
147 | | /// This tells how a thread pool will be used |
148 | | class ThreadPoolStrategy { |
149 | | public: |
150 | | // The default value (0) means all available threads should be used, |
151 | | // taking the affinity mask into account. If set, this value only represents |
152 | | // a suggested high bound, the runtime might choose a lower value (not |
153 | | // higher). |
154 | | unsigned ThreadsRequested = 0; |
155 | | |
156 | | // If SMT is active, use hyper threads. If false, there will be only one |
157 | | // std::thread per core. |
158 | | bool UseHyperThreads = true; |
159 | | |
160 | | // If set, will constrain 'ThreadsRequested' to the number of hardware |
161 | | // threads, or hardware cores. |
162 | | bool Limit = false; |
163 | | |
164 | | /// Retrieves the max available threads for the current strategy. This |
165 | | /// accounts for affinity masks and takes advantage of all CPU sockets. |
166 | | unsigned compute_thread_count() const; |
167 | | |
168 | | /// Assign the current thread to an ideal hardware CPU or NUMA node. In a |
169 | | /// multi-socket system, this ensures threads are assigned to all CPU |
170 | | /// sockets. \p ThreadPoolNum represents a number bounded by [0, |
171 | | /// compute_thread_count()). |
172 | | void apply_thread_strategy(unsigned ThreadPoolNum) const; |
173 | | |
174 | | /// Finds the CPU socket where a thread should go. Returns 'None' if the |
175 | | /// thread shall remain on the actual CPU socket. |
176 | | Optional<unsigned> compute_cpu_socket(unsigned ThreadPoolNum) const; |
177 | | }; |
178 | | |
179 | | /// Build a strategy from a number of threads as a string provided in \p Num. |
180 | | /// When Num is above the max number of threads specified by the \p Default |
181 | | /// strategy, we attempt to equally allocate the threads on all CPU sockets. |
182 | | /// "0" or an empty string will return the \p Default strategy. |
183 | | /// "all" for using all hardware threads. |
184 | | Optional<ThreadPoolStrategy> |
185 | | get_threadpool_strategy(StringRef Num, ThreadPoolStrategy Default = {}); |
186 | | |
187 | | /// Returns a thread strategy for tasks requiring significant memory or other |
188 | | /// resources. To be used for workloads where hardware_concurrency() proves to |
189 | | /// be less efficient. Avoid this strategy if doing lots of I/O. Currently |
190 | | /// based on physical cores, if available for the host system, otherwise falls |
191 | | /// back to hardware_concurrency(). Returns 1 when LLVM is configured with |
192 | | /// LLVM_ENABLE_THREADS = OFF. |
193 | | inline ThreadPoolStrategy |
194 | 0 | heavyweight_hardware_concurrency(unsigned ThreadCount = 0) { |
195 | 0 | ThreadPoolStrategy S; |
196 | 0 | S.UseHyperThreads = false; |
197 | 0 | S.ThreadsRequested = ThreadCount; |
198 | 0 | return S; |
199 | 0 | } |
200 | | |
201 | | /// Like heavyweight_hardware_concurrency() above, but builds a strategy |
202 | | /// based on the rules described for get_threadpool_strategy(). |
203 | | /// If \p Num is invalid, returns a default strategy where one thread per |
204 | | /// hardware core is used. |
205 | 0 | inline ThreadPoolStrategy heavyweight_hardware_concurrency(StringRef Num) { |
206 | 0 | Optional<ThreadPoolStrategy> S = |
207 | 0 | get_threadpool_strategy(Num, heavyweight_hardware_concurrency()); |
208 | 0 | if (S) |
209 | 0 | return *S; |
210 | 0 | return heavyweight_hardware_concurrency(); |
211 | 0 | } |
212 | | |
213 | | /// Returns a default thread strategy where all available hardware ressources |
214 | | /// are to be used, except for those initially excluded by an affinity mask. |
215 | | /// This function takes affinity into consideration. Returns 1 when LLVM is |
216 | | /// configured with LLVM_ENABLE_THREADS=OFF. |
217 | 0 | inline ThreadPoolStrategy hardware_concurrency(unsigned ThreadCount = 0) { |
218 | 0 | ThreadPoolStrategy S; |
219 | 0 | S.ThreadsRequested = ThreadCount; |
220 | 0 | return S; |
221 | 0 | } |
222 | | |
223 | | /// Return the current thread id, as used in various OS system calls. |
224 | | /// Note that not all platforms guarantee that the value returned will be |
225 | | /// unique across the entire system, so portable code should not assume |
226 | | /// this. |
227 | | uint64_t get_threadid(); |
228 | | |
229 | | /// Get the maximum length of a thread name on this platform. |
230 | | /// A value of 0 means there is no limit. |
231 | | uint32_t get_max_thread_name_length(); |
232 | | |
233 | | /// Set the name of the current thread. Setting a thread's name can |
234 | | /// be helpful for enabling useful diagnostics under a debugger or when |
235 | | /// logging. The level of support for setting a thread's name varies |
236 | | /// wildly across operating systems, and we only make a best effort to |
237 | | /// perform the operation on supported platforms. No indication of success |
238 | | /// or failure is returned. |
239 | | void set_thread_name(const Twine &Name); |
240 | | |
241 | | /// Get the name of the current thread. The level of support for |
242 | | /// getting a thread's name varies wildly across operating systems, and it |
243 | | /// is not even guaranteed that if you can successfully set a thread's name |
244 | | /// that you can later get it back. This function is intended for diagnostic |
245 | | /// purposes, and as with setting a thread's name no indication of whether |
246 | | /// the operation succeeded or failed is returned. |
247 | | void get_thread_name(SmallVectorImpl<char> &Name); |
248 | | |
249 | | /// Returns a mask that represents on which hardware thread, core, CPU, NUMA |
250 | | /// group, the calling thread can be executed. On Windows, threads cannot |
251 | | /// cross CPU sockets boundaries. |
252 | | llvm::BitVector get_thread_affinity_mask(); |
253 | | |
254 | | /// Returns how many physical CPUs or NUMA groups the system has. |
255 | | unsigned get_cpus(); |
256 | | |
257 | | enum class ThreadPriority { |
258 | | Background = 0, |
259 | | Default = 1, |
260 | | }; |
261 | | /// If priority is Background tries to lower current threads priority such |
262 | | /// that it does not affect foreground tasks significantly. Can be used for |
263 | | /// long-running, latency-insensitive tasks to make sure cpu is not hogged by |
264 | | /// this task. |
265 | | /// If the priority is default tries to restore current threads priority to |
266 | | /// default scheduling priority. |
267 | | enum class SetThreadPriorityResult { FAILURE, SUCCESS }; |
268 | | SetThreadPriorityResult set_thread_priority(ThreadPriority Priority); |
269 | | } |
270 | | |
271 | | #endif |