/* * Copyright (C) 2013 The Android Open Source Project * * Licensed under the Apache License, Version 2.0 (the "License"); * you may not use this file except in compliance with the License. * You may obtain a copy of the License at * * http://www.apache.org/licenses/LICENSE-2.0 * * Unless required by applicable law or agreed to in writing, software * distributed under the License is distributed on an "AS IS" BASIS, * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. * See the License for the specific language governing permissions and * limitations under the License. */ #include "RenderThread.h" #include "hwui/Bitmap.h" #include "renderstate/RenderState.h" #include "renderthread/OpenGLPipeline.h" #include "pipeline/skia/SkiaOpenGLReadback.h" #include "pipeline/skia/SkiaOpenGLPipeline.h" #include "pipeline/skia/SkiaVulkanPipeline.h" #include "CanvasContext.h" #include "EglManager.h" #include "OpenGLReadback.h" #include "RenderProxy.h" #include "VulkanManager.h" #include "utils/FatVector.h" #include #include #include #include #include #include #include namespace android { namespace uirenderer { namespace renderthread { // Number of events to read at a time from the DisplayEventReceiver pipe. // The value should be large enough that we can quickly drain the pipe // using just a few large reads. static const size_t EVENT_BUFFER_SIZE = 100; // Slight delay to give the UI time to push us a new frame before we replay static const nsecs_t DISPATCH_FRAME_CALLBACKS_DELAY = milliseconds_to_nanoseconds(4); TaskQueue::TaskQueue() : mHead(nullptr), mTail(nullptr) {} RenderTask* TaskQueue::next() { RenderTask* ret = mHead; if (ret) { mHead = ret->mNext; if (!mHead) { mTail = nullptr; } ret->mNext = nullptr; } return ret; } RenderTask* TaskQueue::peek() { return mHead; } void TaskQueue::queue(RenderTask* task) { // Since the RenderTask itself forms the linked list it is not allowed // to have the same task queued twice LOG_ALWAYS_FATAL_IF(task->mNext || mTail == task, "Task is already in the queue!"); if (mTail) { // Fast path if we can just append if (mTail->mRunAt <= task->mRunAt) { mTail->mNext = task; mTail = task; } else { // Need to find the proper insertion point RenderTask* previous = nullptr; RenderTask* next = mHead; while (next && next->mRunAt <= task->mRunAt) { previous = next; next = next->mNext; } if (!previous) { task->mNext = mHead; mHead = task; } else { previous->mNext = task; if (next) { task->mNext = next; } else { mTail = task; } } } } else { mTail = mHead = task; } } void TaskQueue::queueAtFront(RenderTask* task) { LOG_ALWAYS_FATAL_IF(task->mNext || mHead == task, "Task is already in the queue!"); if (mTail) { task->mNext = mHead; mHead = task; } else { mTail = mHead = task; } } void TaskQueue::remove(RenderTask* task) { // TaskQueue is strict here to enforce that users are keeping track of // their RenderTasks due to how their memory is managed LOG_ALWAYS_FATAL_IF(!task->mNext && mTail != task, "Cannot remove a task that isn't in the queue!"); // If task is the head we can just call next() to pop it off // Otherwise we need to scan through to find the task before it if (peek() == task) { next(); } else { RenderTask* previous = mHead; while (previous->mNext != task) { previous = previous->mNext; } previous->mNext = task->mNext; if (mTail == task) { mTail = previous; } } } class DispatchFrameCallbacks : public RenderTask { private: RenderThread* mRenderThread; public: explicit DispatchFrameCallbacks(RenderThread* rt) : mRenderThread(rt) {} virtual void run() override { mRenderThread->dispatchFrameCallbacks(); } }; static bool gHasRenderThreadInstance = false; bool RenderThread::hasInstance() { return gHasRenderThreadInstance; } RenderThread& RenderThread::getInstance() { // This is a pointer because otherwise __cxa_finalize // will try to delete it like a Good Citizen but that causes us to crash // because we don't want to delete the RenderThread normally. static RenderThread* sInstance = new RenderThread(); gHasRenderThreadInstance = true; return *sInstance; } RenderThread::RenderThread() : Thread(true) , mNextWakeup(LLONG_MAX) , mDisplayEventReceiver(nullptr) , mVsyncRequested(false) , mFrameCallbackTaskPending(false) , mFrameCallbackTask(nullptr) , mRenderState(nullptr) , mEglManager(nullptr) , mVkManager(nullptr) { Properties::load(); mFrameCallbackTask = new DispatchFrameCallbacks(this); mLooper = new Looper(false); run("RenderThread"); } RenderThread::~RenderThread() { LOG_ALWAYS_FATAL("Can't destroy the render thread"); } void RenderThread::initializeDisplayEventReceiver() { LOG_ALWAYS_FATAL_IF(mDisplayEventReceiver, "Initializing a second DisplayEventReceiver?"); mDisplayEventReceiver = new DisplayEventReceiver(); status_t status = mDisplayEventReceiver->initCheck(); LOG_ALWAYS_FATAL_IF(status != NO_ERROR, "Initialization of DisplayEventReceiver " "failed with status: %d", status); // Register the FD mLooper->addFd(mDisplayEventReceiver->getFd(), 0, Looper::EVENT_INPUT, RenderThread::displayEventReceiverCallback, this); } void RenderThread::initThreadLocals() { sp dtoken(SurfaceComposerClient::getBuiltInDisplay( ISurfaceComposer::eDisplayIdMain)); status_t status = SurfaceComposerClient::getDisplayInfo(dtoken, &mDisplayInfo); LOG_ALWAYS_FATAL_IF(status, "Failed to get display info\n"); nsecs_t frameIntervalNanos = static_cast(1000000000 / mDisplayInfo.fps); mTimeLord.setFrameInterval(frameIntervalNanos); initializeDisplayEventReceiver(); mEglManager = new EglManager(*this); mRenderState = new RenderState(*this); mVkManager = new VulkanManager(*this); mCacheManager = new CacheManager(mDisplayInfo); } void RenderThread::dumpGraphicsMemory(int fd) { globalProfileData()->dump(fd); String8 cachesOutput; String8 pipeline; auto renderType = Properties::getRenderPipelineType(); switch (renderType) { case RenderPipelineType::OpenGL: { if (Caches::hasInstance()) { cachesOutput.appendFormat("Caches:\n"); Caches::getInstance().dumpMemoryUsage(cachesOutput); } else { cachesOutput.appendFormat("No caches instance."); } pipeline.appendFormat("FrameBuilder"); break; } case RenderPipelineType::SkiaGL: { mCacheManager->dumpMemoryUsage(cachesOutput, mRenderState); pipeline.appendFormat("Skia (OpenGL)"); break; } case RenderPipelineType::SkiaVulkan: { mCacheManager->dumpMemoryUsage(cachesOutput, mRenderState); pipeline.appendFormat("Skia (Vulkan)"); break; } default: LOG_ALWAYS_FATAL("canvas context type %d not supported", (int32_t) renderType); break; } FILE *file = fdopen(fd, "a"); fprintf(file, "\n%s\n", cachesOutput.string()); fprintf(file, "\nPipeline=%s\n", pipeline.string()); fflush(file); } Readback& RenderThread::readback() { if (!mReadback) { auto renderType = Properties::getRenderPipelineType(); switch (renderType) { case RenderPipelineType::OpenGL: mReadback = new OpenGLReadbackImpl(*this); break; case RenderPipelineType::SkiaGL: case RenderPipelineType::SkiaVulkan: // It works to use the OpenGL pipeline for Vulkan but this is not // ideal as it causes us to create an OpenGL context in addition // to the Vulkan one. mReadback = new skiapipeline::SkiaOpenGLReadback(*this); break; default: LOG_ALWAYS_FATAL("canvas context type %d not supported", (int32_t) renderType); break; } } return *mReadback; } void RenderThread::setGrContext(GrContext* context) { mCacheManager->reset(context); if (mGrContext.get()) { mGrContext->releaseResourcesAndAbandonContext(); } mGrContext.reset(context); } int RenderThread::displayEventReceiverCallback(int fd, int events, void* data) { if (events & (Looper::EVENT_ERROR | Looper::EVENT_HANGUP)) { ALOGE("Display event receiver pipe was closed or an error occurred. " "events=0x%x", events); return 0; // remove the callback } if (!(events & Looper::EVENT_INPUT)) { ALOGW("Received spurious callback for unhandled poll event. " "events=0x%x", events); return 1; // keep the callback } reinterpret_cast(data)->drainDisplayEventQueue(); return 1; // keep the callback } static nsecs_t latestVsyncEvent(DisplayEventReceiver* receiver) { DisplayEventReceiver::Event buf[EVENT_BUFFER_SIZE]; nsecs_t latest = 0; ssize_t n; while ((n = receiver->getEvents(buf, EVENT_BUFFER_SIZE)) > 0) { for (ssize_t i = 0; i < n; i++) { const DisplayEventReceiver::Event& ev = buf[i]; switch (ev.header.type) { case DisplayEventReceiver::DISPLAY_EVENT_VSYNC: latest = ev.header.timestamp; break; } } } if (n < 0) { ALOGW("Failed to get events from display event receiver, status=%d", status_t(n)); } return latest; } void RenderThread::drainDisplayEventQueue() { ATRACE_CALL(); nsecs_t vsyncEvent = latestVsyncEvent(mDisplayEventReceiver); if (vsyncEvent > 0) { mVsyncRequested = false; if (mTimeLord.vsyncReceived(vsyncEvent) && !mFrameCallbackTaskPending) { ATRACE_NAME("queue mFrameCallbackTask"); mFrameCallbackTaskPending = true; nsecs_t runAt = (vsyncEvent + DISPATCH_FRAME_CALLBACKS_DELAY); queueAt(mFrameCallbackTask, runAt); } } } void RenderThread::dispatchFrameCallbacks() { ATRACE_CALL(); mFrameCallbackTaskPending = false; std::set callbacks; mFrameCallbacks.swap(callbacks); if (callbacks.size()) { // Assume one of them will probably animate again so preemptively // request the next vsync in case it occurs mid-frame requestVsync(); for (std::set::iterator it = callbacks.begin(); it != callbacks.end(); it++) { (*it)->doFrame(); } } } void RenderThread::requestVsync() { if (!mVsyncRequested) { mVsyncRequested = true; status_t status = mDisplayEventReceiver->requestNextVsync(); LOG_ALWAYS_FATAL_IF(status != NO_ERROR, "requestNextVsync failed with status: %d", status); } } bool RenderThread::threadLoop() { setpriority(PRIO_PROCESS, 0, PRIORITY_DISPLAY); initThreadLocals(); int timeoutMillis = -1; for (;;) { int result = mLooper->pollOnce(timeoutMillis); LOG_ALWAYS_FATAL_IF(result == Looper::POLL_ERROR, "RenderThread Looper POLL_ERROR!"); nsecs_t nextWakeup; { FatVector workQueue; // Process our queue, if we have anything. By first acquiring // all the pending events then processing them we avoid vsync // starvation if more tasks are queued while we are processing tasks. while (RenderTask* task = nextTask(&nextWakeup)) { workQueue.push_back(task); } for (auto task : workQueue) { task->run(); // task may have deleted itself, do not reference it again } } if (nextWakeup == LLONG_MAX) { timeoutMillis = -1; } else { nsecs_t timeoutNanos = nextWakeup - systemTime(SYSTEM_TIME_MONOTONIC); timeoutMillis = nanoseconds_to_milliseconds(timeoutNanos); if (timeoutMillis < 0) { timeoutMillis = 0; } } if (mPendingRegistrationFrameCallbacks.size() && !mFrameCallbackTaskPending) { drainDisplayEventQueue(); mFrameCallbacks.insert( mPendingRegistrationFrameCallbacks.begin(), mPendingRegistrationFrameCallbacks.end()); mPendingRegistrationFrameCallbacks.clear(); requestVsync(); } if (!mFrameCallbackTaskPending && !mVsyncRequested && mFrameCallbacks.size()) { // TODO: Clean this up. This is working around an issue where a combination // of bad timing and slow drawing can result in dropping a stale vsync // on the floor (correct!) but fails to schedule to listen for the // next vsync (oops), so none of the callbacks are run. requestVsync(); } } return false; } void RenderThread::queue(RenderTask* task) { AutoMutex _lock(mLock); mQueue.queue(task); if (mNextWakeup && task->mRunAt < mNextWakeup) { mNextWakeup = 0; mLooper->wake(); } } void RenderThread::queueAndWait(RenderTask* task) { // These need to be local to the thread to avoid the Condition // signaling the wrong thread. The easiest way to achieve that is to just // make this on the stack, although that has a slight cost to it Mutex mutex; Condition condition; SignalingRenderTask syncTask(task, &mutex, &condition); AutoMutex _lock(mutex); queue(&syncTask); while (!syncTask.hasRun()) { condition.wait(mutex); } } void RenderThread::queueAtFront(RenderTask* task) { AutoMutex _lock(mLock); mQueue.queueAtFront(task); mLooper->wake(); } void RenderThread::queueAt(RenderTask* task, nsecs_t runAtNs) { task->mRunAt = runAtNs; queue(task); } void RenderThread::remove(RenderTask* task) { AutoMutex _lock(mLock); mQueue.remove(task); } void RenderThread::postFrameCallback(IFrameCallback* callback) { mPendingRegistrationFrameCallbacks.insert(callback); } bool RenderThread::removeFrameCallback(IFrameCallback* callback) { size_t erased; erased = mFrameCallbacks.erase(callback); erased |= mPendingRegistrationFrameCallbacks.erase(callback); return erased; } void RenderThread::pushBackFrameCallback(IFrameCallback* callback) { if (mFrameCallbacks.erase(callback)) { mPendingRegistrationFrameCallbacks.insert(callback); } } RenderTask* RenderThread::nextTask(nsecs_t* nextWakeup) { AutoMutex _lock(mLock); RenderTask* next = mQueue.peek(); if (!next) { mNextWakeup = LLONG_MAX; } else { mNextWakeup = next->mRunAt; // Most tasks won't be delayed, so avoid unnecessary systemTime() calls if (next->mRunAt <= 0 || next->mRunAt <= systemTime(SYSTEM_TIME_MONOTONIC)) { next = mQueue.next(); } else { next = nullptr; } } if (nextWakeup) { *nextWakeup = mNextWakeup; } return next; } sk_sp RenderThread::allocateHardwareBitmap(SkBitmap& skBitmap) { auto renderType = Properties::getRenderPipelineType(); switch (renderType) { case RenderPipelineType::OpenGL: return OpenGLPipeline::allocateHardwareBitmap(*this, skBitmap); case RenderPipelineType::SkiaGL: return skiapipeline::SkiaOpenGLPipeline::allocateHardwareBitmap(*this, skBitmap); case RenderPipelineType::SkiaVulkan: return skiapipeline::SkiaVulkanPipeline::allocateHardwareBitmap(*this, skBitmap); default: LOG_ALWAYS_FATAL("canvas context type %d not supported", (int32_t) renderType); break; } return nullptr; } } /* namespace renderthread */ } /* namespace uirenderer */ } /* namespace android */