// Licensed to the Apache Software Foundation (ASF) under one // or more contributor license agreements. See the NOTICE file // distributed with this work for additional information // regarding copyright ownership. The ASF licenses this file // to you under the Apache License, Version 2.0 (the // "License"); you may not use this file except in compliance // with the License. You may obtain a copy of the License at // // http://www.apache.org/licenses/LICENSE-2.0 // // Unless required by applicable law or agreed to in writing, // software distributed under the License is distributed on an // "AS IS" BASIS, WITHOUT WARRANTIES OR CONDITIONS OF ANY // KIND, either express or implied. See the License for the // specific language governing permissions and limitations // under the License. #pragma once #ifndef _WIN32 #include #endif #ifndef _WIN32 #include #endif #include #include #include #include #include #include "arrow/result.h" #include "arrow/status.h" #include "arrow/util/cancel.h" #include "arrow/util/functional.h" #include "arrow/util/future.h" #include "arrow/util/iterator.h" #include "arrow/util/macros.h" #include "arrow/util/visibility.h" #if defined(_MSC_VER) // Disable harmless warning for decorated name length limit #pragma warning(disable : 4503) #endif namespace arrow { /// \brief Get the capacity of the global thread pool /// /// Return the number of worker threads in the thread pool to which /// Arrow dispatches various CPU-bound tasks. This is an ideal number, /// not necessarily the exact number of threads at a given point in time. /// /// You can change this number using SetCpuThreadPoolCapacity(). ARROW_EXPORT int GetCpuThreadPoolCapacity(); /// \brief Set the capacity of the global thread pool /// /// Set the number of worker threads int the thread pool to which /// Arrow dispatches various CPU-bound tasks. /// /// The current number is returned by GetCpuThreadPoolCapacity(). ARROW_EXPORT Status SetCpuThreadPoolCapacity(int threads); namespace internal { // Hints about a task that may be used by an Executor. // They are ignored by the provided ThreadPool implementation. struct TaskHints { // The lower, the more urgent int32_t priority = 0; // The IO transfer size in bytes int64_t io_size = -1; // The approximate CPU cost in number of instructions int64_t cpu_cost = -1; // An application-specific ID int64_t external_id = -1; }; class ARROW_EXPORT Executor { public: using StopCallback = internal::FnOnce; virtual ~Executor(); // Spawn a fire-and-forget task. template Status Spawn(Function&& func) { return SpawnReal(TaskHints{}, std::forward(func), StopToken::Unstoppable(), StopCallback{}); } template Status Spawn(Function&& func, StopToken stop_token) { return SpawnReal(TaskHints{}, std::forward(func), std::move(stop_token), StopCallback{}); } template Status Spawn(TaskHints hints, Function&& func) { return SpawnReal(hints, std::forward(func), StopToken::Unstoppable(), StopCallback{}); } template Status Spawn(TaskHints hints, Function&& func, StopToken stop_token) { return SpawnReal(hints, std::forward(func), std::move(stop_token), StopCallback{}); } template Status Spawn(TaskHints hints, Function&& func, StopToken stop_token, StopCallback stop_callback) { return SpawnReal(hints, std::forward(func), std::move(stop_token), std::move(stop_callback)); } // Transfers a future to this executor. Any continuations added to the // returned future will run in this executor. Otherwise they would run // on the same thread that called MarkFinished. // // This is necessary when (for example) an I/O task is completing a future. // The continuations of that future should run on the CPU thread pool keeping // CPU heavy work off the I/O thread pool. So the I/O task should transfer // the future to the CPU executor before returning. // // By default this method will only transfer if the future is not already completed. If // the future is already completed then any callback would be run synchronously and so // no transfer is typically necessary. However, in cases where you want to force a // transfer (e.g. to help the scheduler break up units of work across multiple cores) // then you can override this behavior with `always_transfer`. template Future Transfer(Future future) { return DoTransfer(std::move(future), false); } // Overload of Transfer which will always schedule callbacks on new threads even if the // future is finished when the callback is added. // // This can be useful in cases where you want to ensure parallelism template Future TransferAlways(Future future) { return DoTransfer(std::move(future), true); } // Submit a callable and arguments for execution. Return a future that // will return the callable's result value once. // The callable's arguments are copied before execution. template > Result Submit(TaskHints hints, StopToken stop_token, Function&& func, Args&&... args) { using ValueType = typename FutureType::ValueType; auto future = FutureType::Make(); auto task = std::bind(::arrow::detail::ContinueFuture{}, future, std::forward(func), std::forward(args)...); struct { WeakFuture weak_fut; void operator()(const Status& st) { auto fut = weak_fut.get(); if (fut.is_valid()) { fut.MarkFinished(st); } } } stop_callback{WeakFuture(future)}; ARROW_RETURN_NOT_OK(SpawnReal(hints, std::move(task), std::move(stop_token), std::move(stop_callback))); return future; } template > Result Submit(StopToken stop_token, Function&& func, Args&&... args) { return Submit(TaskHints{}, stop_token, std::forward(func), std::forward(args)...); } template > Result Submit(TaskHints hints, Function&& func, Args&&... args) { return Submit(std::move(hints), StopToken::Unstoppable(), std::forward(func), std::forward(args)...); } template > Result Submit(Function&& func, Args&&... args) { return Submit(TaskHints{}, StopToken::Unstoppable(), std::forward(func), std::forward(args)...); } // Return the level of parallelism (the number of tasks that may be executed // concurrently). This may be an approximate number. virtual int GetCapacity() = 0; // Return true if the thread from which this function is called is owned by this // Executor. Returns false if this Executor does not support this property. virtual bool OwnsThisThread() { return false; } /// \brief An interface to represent something with a custom destructor /// /// \see KeepAlive class ARROW_EXPORT Resource { public: virtual ~Resource() = default; }; /// \brief Keep a resource alive until all executor threads have terminated /// /// Executors may have static storage duration. In particular, the CPU and I/O /// executors are currently implemented this way. These threads may access other /// objects with static storage duration such as the OpenTelemetry runtime context /// the default memory pool, or other static executors. /// /// The order in which these objects are destroyed is difficult to control. In order /// to ensure those objects remain alive until all threads have finished those objects /// should be wrapped in a Resource object and passed into this method. The given /// shared_ptr will be kept alive until all threads have finished their worker loops. virtual void KeepAlive(std::shared_ptr resource); protected: ARROW_DISALLOW_COPY_AND_ASSIGN(Executor); Executor() = default; template , typename FTSync = typename FT::SyncType> Future DoTransfer(Future future, bool always_transfer = false) { auto transferred = Future::Make(); if (always_transfer) { CallbackOptions callback_options = CallbackOptions::Defaults(); callback_options.should_schedule = ShouldSchedule::Always; callback_options.executor = this; auto sync_callback = [transferred](const FTSync& result) mutable { transferred.MarkFinished(result); }; future.AddCallback(sync_callback, callback_options); return transferred; } // We could use AddCallback's ShouldSchedule::IfUnfinished but we can save a bit of // work by doing the test here. auto callback = [this, transferred](const FTSync& result) mutable { auto spawn_status = Spawn([transferred, result]() mutable { transferred.MarkFinished(result); }); if (!spawn_status.ok()) { transferred.MarkFinished(spawn_status); } }; auto callback_factory = [&callback]() { return callback; }; if (future.TryAddCallback(callback_factory)) { return transferred; } // If the future is already finished and we aren't going to force spawn a thread // then we don't need to add another layer of callback and can return the original // future return future; } // Subclassing API virtual Status SpawnReal(TaskHints hints, FnOnce task, StopToken, StopCallback&&) = 0; }; /// \brief An executor implementation that runs all tasks on a single thread using an /// event loop. /// /// Note: Any sort of nested parallelism will deadlock this executor. Blocking waits are /// fine but if one task needs to wait for another task it must be expressed as an /// asynchronous continuation. class ARROW_EXPORT SerialExecutor : public Executor { public: template using TopLevelTask = internal::FnOnce(Executor*)>; ~SerialExecutor() override; int GetCapacity() override { return 1; }; Status SpawnReal(TaskHints hints, FnOnce task, StopToken, StopCallback&&) override; /// \brief Runs the TopLevelTask and any scheduled tasks /// /// The TopLevelTask (or one of the tasks it schedules) must either return an invalid /// status or call the finish signal. Failure to do this will result in a deadlock. For /// this reason it is preferable (if possible) to use the helper methods (below) /// RunSynchronously/RunSerially which delegates the responsiblity onto a Future /// producer's existing responsibility to always mark a future finished (which can /// someday be aided by ARROW-12207). template , typename FTSync = typename FT::SyncType> static FTSync RunInSerialExecutor(TopLevelTask initial_task) { Future fut = SerialExecutor().Run(std::move(initial_task)); return FutureToSync(fut); } /// \brief Transform an AsyncGenerator into an Iterator /// /// An event loop will be created and each call to Next will power the event loop with /// the calling thread until the next item is ready to be delivered. /// /// Note: The iterator's destructor will run until the given generator is fully /// exhausted. If you wish to abandon iteration before completion then the correct /// approach is to use a stop token to cause the generator to exhaust early. template static Iterator IterateGenerator( internal::FnOnce()>>(Executor*)> initial_task) { auto serial_executor = std::unique_ptr(new SerialExecutor()); auto maybe_generator = std::move(initial_task)(serial_executor.get()); if (!maybe_generator.ok()) { return MakeErrorIterator(maybe_generator.status()); } auto generator = maybe_generator.MoveValueUnsafe(); struct SerialIterator { SerialIterator(std::unique_ptr executor, std::function()> generator) : executor(std::move(executor)), generator(std::move(generator)) {} ARROW_DISALLOW_COPY_AND_ASSIGN(SerialIterator); ARROW_DEFAULT_MOVE_AND_ASSIGN(SerialIterator); ~SerialIterator() { // A serial iterator must be consumed before it can be destroyed. Allowing it to // do otherwise would lead to resource leakage. There will likely be deadlocks at // this spot in the future but these will be the result of other bugs and not the // fact that we are forcing consumption here. // If a streaming API needs to support early abandonment then it should be done so // with a cancellation token and not simply discarding the iterator and expecting // the underlying work to clean up correctly. if (executor && !executor->IsFinished()) { while (true) { Result maybe_next = Next(); if (!maybe_next.ok() || IsIterationEnd(*maybe_next)) { break; } } } } Result Next() { executor->Unpause(); // This call may lead to tasks being scheduled in the serial executor Future next_fut = generator(); next_fut.AddCallback([this](const Result& res) { // If we're done iterating we should drain the rest of the tasks in the executor if (!res.ok() || IsIterationEnd(*res)) { executor->Finish(); return; } // Otherwise we will break out immediately, leaving the remaining tasks for // the next call. executor->Pause(); }); // Borrow this thread and run tasks until the future is finished executor->RunLoop(); if (!next_fut.is_finished()) { // Not clear this is possible since RunLoop wouldn't generally exit // unless we paused/finished which would imply next_fut has been // finished. return Status::Invalid( "Serial executor terminated before next result computed"); } // At this point we may still have tasks in the executor, that is ok. // We will run those tasks the next time through. return next_fut.result(); } std::unique_ptr executor; std::function()> generator; }; return Iterator(SerialIterator{std::move(serial_executor), std::move(generator)}); } private: SerialExecutor(); // State uses mutex struct State; std::shared_ptr state_; void RunLoop(); // We mark the serial executor "finished" when there should be // no more tasks scheduled on it. It's not strictly needed but // can help catch bugs where we are trying to use the executor // after we are done with it. void Finish(); bool IsFinished(); // We pause the executor when we are running an async generator // and we have received an item that we can deliver. void Pause(); void Unpause(); template ::SyncType> Future Run(TopLevelTask initial_task) { auto final_fut = std::move(initial_task)(this); final_fut.AddCallback([this](const FTSync&) { Finish(); }); RunLoop(); return final_fut; } }; /// An Executor implementation spawning tasks in FIFO manner on a fixed-size /// pool of worker threads. /// /// Note: Any sort of nested parallelism will deadlock this executor. Blocking waits are /// fine but if one task needs to wait for another task it must be expressed as an /// asynchronous continuation. class ARROW_EXPORT ThreadPool : public Executor { public: // Construct a thread pool with the given number of worker threads static Result> Make(int threads); // Like Make(), but takes care that the returned ThreadPool is compatible // with destruction late at process exit. static Result> MakeEternal(int threads); // Destroy thread pool; the pool will first be shut down ~ThreadPool() override; // Return the desired number of worker threads. // The actual number of workers may lag a bit before being adjusted to // match this value. int GetCapacity() override; bool OwnsThisThread() override; // Return the number of tasks either running or in the queue. int GetNumTasks(); // Dynamically change the number of worker threads. // // This function always returns immediately. // If fewer threads are running than this number, new threads are spawned // on-demand when needed for task execution. // If more threads are running than this number, excess threads are reaped // as soon as possible. Status SetCapacity(int threads); // Heuristic for the default capacity of a thread pool for CPU-bound tasks. // This is exposed as a static method to help with testing. static int DefaultCapacity(); // Shutdown the pool. Once the pool starts shutting down, new tasks // cannot be submitted anymore. // If "wait" is true, shutdown waits for all pending tasks to be finished. // If "wait" is false, workers are stopped as soon as currently executing // tasks are finished. Status Shutdown(bool wait = true); // Wait for the thread pool to become idle // // This is useful for sequencing tests void WaitForIdle(); void KeepAlive(std::shared_ptr resource) override; struct State; protected: FRIEND_TEST(TestThreadPool, SetCapacity); FRIEND_TEST(TestGlobalThreadPool, Capacity); friend ARROW_EXPORT ThreadPool* GetCpuThreadPool(); ThreadPool(); Status SpawnReal(TaskHints hints, FnOnce task, StopToken, StopCallback&&) override; // Collect finished worker threads, making sure the OS threads have exited void CollectFinishedWorkersUnlocked(); // Launch a given number of additional workers void LaunchWorkersUnlocked(int threads); // Get the current actual capacity int GetActualCapacity(); // Reinitialize the thread pool if the pid changed void ProtectAgainstFork(); static std::shared_ptr MakeCpuThreadPool(); std::shared_ptr sp_state_; State* state_; bool shutdown_on_destroy_; #ifndef _WIN32 std::atomic pid_; #endif }; // Return the process-global thread pool for CPU-bound tasks. ARROW_EXPORT ThreadPool* GetCpuThreadPool(); /// \brief Potentially run an async operation serially (if use_threads is false) /// \see RunSerially /// /// If `use_threads` is true, the global CPU executor is used. /// If `use_threads` is false, a temporary SerialExecutor is used. /// `get_future` is called (from this thread) with the chosen executor and must /// return a future that will eventually finish. This function returns once the /// future has finished. template typename Fut::SyncType RunSynchronously(FnOnce get_future, bool use_threads) { if (use_threads) { auto fut = std::move(get_future)(GetCpuThreadPool()); return FutureToSync(fut); } else { return SerialExecutor::RunInSerialExecutor(std::move(get_future)); } } } // namespace internal } // namespace arrow