1 /*
2 * Copyright (C) 2010 The Android Open Source Project
3 *
4 * Licensed under the Apache License, Version 2.0 (the "License");
5 * you may not use this file except in compliance with the License.
6 * You may obtain a copy of the License at
7 *
8 * http://www.apache.org/licenses/LICENSE-2.0
9 *
10 * Unless required by applicable law or agreed to in writing, software
11 * distributed under the License is distributed on an "AS IS" BASIS,
12 * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
13 * See the License for the specific language governing permissions and
14 * limitations under the License.
15 */
16
17 #define LOG_TAG "Surface"
18 #define ATRACE_TAG ATRACE_TAG_GRAPHICS
19 //#define LOG_NDEBUG 0
20
21 #include <gui/Surface.h>
22
23 #include <condition_variable>
24 #include <cstddef>
25 #include <cstdint>
26 #include <deque>
27 #include <mutex>
28 #include <thread>
29
30 #include <inttypes.h>
31
32 #include <android/gui/DisplayStatInfo.h>
33 #include <android/native_window.h>
34
35 #include <gui/FenceMonitor.h>
36 #include <gui/TraceUtils.h>
37 #include <utils/Log.h>
38 #include <utils/NativeHandle.h>
39 #include <utils/Trace.h>
40
41 #include <ui/DynamicDisplayInfo.h>
42 #include <ui/Fence.h>
43 #include <ui/GraphicBuffer.h>
44 #include <ui/Region.h>
45
46 #include <gui/AidlUtil.h>
47 #include <gui/BufferItem.h>
48
49 #include <gui/ISurfaceComposer.h>
50 #include <gui/LayerState.h>
51 #include <private/gui/ComposerService.h>
52 #include <private/gui/ComposerServiceAIDL.h>
53
54 #include <com_android_graphics_libgui_flags.h>
55
56 namespace android {
57
58 using namespace com::android::graphics::libgui;
59 using gui::aidl_utils::statusTFromBinderStatus;
60 using ui::Dataspace;
61
62 namespace {
63
64 enum {
65 // moved from nativewindow/include/system/window.h, to be removed
66 NATIVE_WINDOW_GET_WIDE_COLOR_SUPPORT = 28,
67 NATIVE_WINDOW_GET_HDR_SUPPORT = 29,
68 };
69
isInterceptorRegistrationOp(int op)70 bool isInterceptorRegistrationOp(int op) {
71 return op == NATIVE_WINDOW_SET_CANCEL_INTERCEPTOR ||
72 op == NATIVE_WINDOW_SET_DEQUEUE_INTERCEPTOR ||
73 op == NATIVE_WINDOW_SET_PERFORM_INTERCEPTOR ||
74 op == NATIVE_WINDOW_SET_QUEUE_INTERCEPTOR ||
75 op == NATIVE_WINDOW_SET_QUERY_INTERCEPTOR;
76 }
77
78 } // namespace
79
80 #if COM_ANDROID_GRAPHICS_LIBGUI_FLAGS(WB_PLATFORM_API_IMPROVEMENTS)
ProducerDeathListenerProxy(wp<SurfaceListener> surfaceListener)81 Surface::ProducerDeathListenerProxy::ProducerDeathListenerProxy(wp<SurfaceListener> surfaceListener)
82 : mSurfaceListener(surfaceListener) {}
83
binderDied(const wp<IBinder> &)84 void Surface::ProducerDeathListenerProxy::binderDied(const wp<IBinder>&) {
85 sp<SurfaceListener> surfaceListener = mSurfaceListener.promote();
86 if (!surfaceListener) {
87 return;
88 }
89
90 if (surfaceListener->needsDeathNotify()) {
91 surfaceListener->onRemoteDied();
92 }
93 }
94 #endif // COM_ANDROID_GRAPHICS_LIBGUI_FLAGS(WB_PLATFORM_API_IMPROVEMENTS)
95
Surface(const sp<IGraphicBufferProducer> & bufferProducer,bool controlledByApp,const sp<IBinder> & surfaceControlHandle)96 Surface::Surface(const sp<IGraphicBufferProducer>& bufferProducer, bool controlledByApp,
97 const sp<IBinder>& surfaceControlHandle)
98 : mGraphicBufferProducer(bufferProducer),
99 #if COM_ANDROID_GRAPHICS_LIBGUI_FLAGS(WB_PLATFORM_API_IMPROVEMENTS)
100 mSurfaceDeathListener(nullptr),
101 #endif // COM_ANDROID_GRAPHICS_LIBGUI_FLAGS(WB_PLATFORM_API_IMPROVEMENTS)
102 mCrop(Rect::EMPTY_RECT),
103 mBufferAge(0),
104 mGenerationNumber(0),
105 mSharedBufferMode(false),
106 mAutoRefresh(false),
107 mAutoPrerotation(false),
108 mSharedBufferSlot(BufferItem::INVALID_BUFFER_SLOT),
109 mSharedBufferHasBeenQueued(false),
110 mQueriedSupportedTimestamps(false),
111 mFrameTimestampsSupportsPresent(false),
112 mEnableFrameTimestamps(false),
113 mFrameEventHistory(std::make_unique<ProducerFrameEventHistory>()) {
114 // Initialize the ANativeWindow function pointers.
115 ANativeWindow::setSwapInterval = hook_setSwapInterval;
116 ANativeWindow::dequeueBuffer = hook_dequeueBuffer;
117 ANativeWindow::cancelBuffer = hook_cancelBuffer;
118 ANativeWindow::queueBuffer = hook_queueBuffer;
119 ANativeWindow::query = hook_query;
120 ANativeWindow::perform = hook_perform;
121
122 ANativeWindow::dequeueBuffer_DEPRECATED = hook_dequeueBuffer_DEPRECATED;
123 ANativeWindow::cancelBuffer_DEPRECATED = hook_cancelBuffer_DEPRECATED;
124 ANativeWindow::lockBuffer_DEPRECATED = hook_lockBuffer_DEPRECATED;
125 ANativeWindow::queueBuffer_DEPRECATED = hook_queueBuffer_DEPRECATED;
126
127 const_cast<int&>(ANativeWindow::minSwapInterval) = 0;
128 const_cast<int&>(ANativeWindow::maxSwapInterval) = 1;
129
130 mReqWidth = 0;
131 mReqHeight = 0;
132 mReqFormat = 0;
133 mReqUsage = 0;
134 mTimestamp = NATIVE_WINDOW_TIMESTAMP_AUTO;
135 mDataSpace = Dataspace::UNKNOWN;
136 mScalingMode = NATIVE_WINDOW_SCALING_MODE_FREEZE;
137 mTransform = 0;
138 mStickyTransform = 0;
139 mDefaultWidth = 0;
140 mDefaultHeight = 0;
141 mUserWidth = 0;
142 mUserHeight = 0;
143 mTransformHint = 0;
144 mConsumerRunningBehind = false;
145 mConnectedToCpu = false;
146 mProducerControlledByApp = controlledByApp;
147 mSwapIntervalZero = false;
148 mMaxBufferCount = NUM_BUFFER_SLOTS;
149 mSurfaceControlHandle = surfaceControlHandle;
150 }
151
~Surface()152 Surface::~Surface() {
153 if (mConnectedToCpu) {
154 Surface::disconnect(NATIVE_WINDOW_API_CPU);
155 }
156 #if COM_ANDROID_GRAPHICS_LIBGUI_FLAGS(WB_PLATFORM_API_IMPROVEMENTS)
157 if (mSurfaceDeathListener != nullptr) {
158 IInterface::asBinder(mGraphicBufferProducer)->unlinkToDeath(mSurfaceDeathListener);
159 mSurfaceDeathListener = nullptr;
160 }
161 #endif // COM_ANDROID_GRAPHICS_LIBGUI_FLAGS(WB_PLATFORM_API_IMPROVEMENTS)
162 }
163
composerService() const164 sp<ISurfaceComposer> Surface::composerService() const {
165 return ComposerService::getComposerService();
166 }
167
composerServiceAIDL() const168 sp<gui::ISurfaceComposer> Surface::composerServiceAIDL() const {
169 return ComposerServiceAIDL::getComposerService();
170 }
171
now() const172 nsecs_t Surface::now() const {
173 return systemTime();
174 }
175
getIGraphicBufferProducer() const176 sp<IGraphicBufferProducer> Surface::getIGraphicBufferProducer() const {
177 return mGraphicBufferProducer;
178 }
179
setSidebandStream(const sp<NativeHandle> & stream)180 void Surface::setSidebandStream(const sp<NativeHandle>& stream) {
181 mGraphicBufferProducer->setSidebandStream(stream);
182 }
183
allocateBuffers()184 void Surface::allocateBuffers() {
185 uint32_t reqWidth = mReqWidth ? mReqWidth : mUserWidth;
186 uint32_t reqHeight = mReqHeight ? mReqHeight : mUserHeight;
187 mGraphicBufferProducer->allocateBuffers(reqWidth, reqHeight,
188 mReqFormat, mReqUsage);
189 }
190
191 #if COM_ANDROID_GRAPHICS_LIBGUI_FLAGS(WB_PLATFORM_API_IMPROVEMENTS)
allowAllocation(bool allowAllocation)192 status_t Surface::allowAllocation(bool allowAllocation) {
193 return mGraphicBufferProducer->allowAllocation(allowAllocation);
194 }
195 #endif // COM_ANDROID_GRAPHICS_LIBGUI_FLAGS(WB_PLATFORM_API_IMPROVEMENTS)
196
setGenerationNumber(uint32_t generation)197 status_t Surface::setGenerationNumber(uint32_t generation) {
198 status_t result = mGraphicBufferProducer->setGenerationNumber(generation);
199 if (result == NO_ERROR) {
200 mGenerationNumber = generation;
201 }
202 return result;
203 }
204
getNextFrameNumber() const205 uint64_t Surface::getNextFrameNumber() const {
206 Mutex::Autolock lock(mMutex);
207 return mNextFrameNumber;
208 }
209
getConsumerName() const210 String8 Surface::getConsumerName() const {
211 return mGraphicBufferProducer->getConsumerName();
212 }
213
setDequeueTimeout(nsecs_t timeout)214 status_t Surface::setDequeueTimeout(nsecs_t timeout) {
215 return mGraphicBufferProducer->setDequeueTimeout(timeout);
216 }
217
getLastQueuedBuffer(sp<GraphicBuffer> * outBuffer,sp<Fence> * outFence,float outTransformMatrix[16])218 status_t Surface::getLastQueuedBuffer(sp<GraphicBuffer>* outBuffer,
219 sp<Fence>* outFence, float outTransformMatrix[16]) {
220 return mGraphicBufferProducer->getLastQueuedBuffer(outBuffer, outFence,
221 outTransformMatrix);
222 }
223
getDisplayRefreshCycleDuration(nsecs_t * outRefreshDuration)224 status_t Surface::getDisplayRefreshCycleDuration(nsecs_t* outRefreshDuration) {
225 ATRACE_CALL();
226
227 gui::DisplayStatInfo stats;
228 binder::Status status = composerServiceAIDL()->getDisplayStats(nullptr, &stats);
229 if (!status.isOk()) {
230 return statusTFromBinderStatus(status);
231 }
232
233 *outRefreshDuration = stats.vsyncPeriod;
234
235 return NO_ERROR;
236 }
237
enableFrameTimestamps(bool enable)238 void Surface::enableFrameTimestamps(bool enable) {
239 Mutex::Autolock lock(mMutex);
240 // If going from disabled to enabled, get the initial values for
241 // compositor and display timing.
242 if (!mEnableFrameTimestamps && enable) {
243 FrameEventHistoryDelta delta;
244 mGraphicBufferProducer->getFrameTimestamps(&delta);
245 mFrameEventHistory->applyDelta(delta);
246 }
247 mEnableFrameTimestamps = enable;
248 }
249
getCompositorTiming(nsecs_t * compositeDeadline,nsecs_t * compositeInterval,nsecs_t * compositeToPresentLatency)250 status_t Surface::getCompositorTiming(
251 nsecs_t* compositeDeadline, nsecs_t* compositeInterval,
252 nsecs_t* compositeToPresentLatency) {
253 Mutex::Autolock lock(mMutex);
254 if (!mEnableFrameTimestamps) {
255 return INVALID_OPERATION;
256 }
257
258 if (compositeDeadline != nullptr) {
259 *compositeDeadline =
260 mFrameEventHistory->getNextCompositeDeadline(now());
261 }
262 if (compositeInterval != nullptr) {
263 *compositeInterval = mFrameEventHistory->getCompositeInterval();
264 }
265 if (compositeToPresentLatency != nullptr) {
266 *compositeToPresentLatency =
267 mFrameEventHistory->getCompositeToPresentLatency();
268 }
269 return NO_ERROR;
270 }
271
checkConsumerForUpdates(const FrameEvents * e,const uint64_t lastFrameNumber,const nsecs_t * outLatchTime,const nsecs_t * outFirstRefreshStartTime,const nsecs_t * outLastRefreshStartTime,const nsecs_t * outGpuCompositionDoneTime,const nsecs_t * outDisplayPresentTime,const nsecs_t * outDequeueReadyTime,const nsecs_t * outReleaseTime)272 static bool checkConsumerForUpdates(
273 const FrameEvents* e, const uint64_t lastFrameNumber,
274 const nsecs_t* outLatchTime,
275 const nsecs_t* outFirstRefreshStartTime,
276 const nsecs_t* outLastRefreshStartTime,
277 const nsecs_t* outGpuCompositionDoneTime,
278 const nsecs_t* outDisplayPresentTime,
279 const nsecs_t* outDequeueReadyTime,
280 const nsecs_t* outReleaseTime) {
281 bool checkForLatch = (outLatchTime != nullptr) && !e->hasLatchInfo();
282 bool checkForFirstRefreshStart = (outFirstRefreshStartTime != nullptr) &&
283 !e->hasFirstRefreshStartInfo();
284 bool checkForGpuCompositionDone = (outGpuCompositionDoneTime != nullptr) &&
285 !e->hasGpuCompositionDoneInfo();
286 bool checkForDisplayPresent = (outDisplayPresentTime != nullptr) &&
287 !e->hasDisplayPresentInfo();
288
289 // LastRefreshStart, DequeueReady, and Release are never available for the
290 // last frame.
291 bool checkForLastRefreshStart = (outLastRefreshStartTime != nullptr) &&
292 !e->hasLastRefreshStartInfo() &&
293 (e->frameNumber != lastFrameNumber);
294 bool checkForDequeueReady = (outDequeueReadyTime != nullptr) &&
295 !e->hasDequeueReadyInfo() && (e->frameNumber != lastFrameNumber);
296 bool checkForRelease = (outReleaseTime != nullptr) &&
297 !e->hasReleaseInfo() && (e->frameNumber != lastFrameNumber);
298
299 // RequestedPresent and Acquire info are always available producer-side.
300 return checkForLatch || checkForFirstRefreshStart ||
301 checkForLastRefreshStart || checkForGpuCompositionDone ||
302 checkForDisplayPresent || checkForDequeueReady || checkForRelease;
303 }
304
getFrameTimestamp(nsecs_t * dst,const nsecs_t & src)305 static void getFrameTimestamp(nsecs_t *dst, const nsecs_t& src) {
306 if (dst != nullptr) {
307 // We always get valid timestamps for these eventually.
308 *dst = (src == FrameEvents::TIMESTAMP_PENDING) ?
309 NATIVE_WINDOW_TIMESTAMP_PENDING : src;
310 }
311 }
312
getFrameTimestampFence(nsecs_t * dst,const std::shared_ptr<FenceTime> & src,bool fenceShouldBeKnown)313 static void getFrameTimestampFence(nsecs_t *dst,
314 const std::shared_ptr<FenceTime>& src, bool fenceShouldBeKnown) {
315 if (dst != nullptr) {
316 if (!fenceShouldBeKnown) {
317 *dst = NATIVE_WINDOW_TIMESTAMP_PENDING;
318 return;
319 }
320
321 nsecs_t signalTime = src->getSignalTime();
322 *dst = (signalTime == Fence::SIGNAL_TIME_PENDING) ?
323 NATIVE_WINDOW_TIMESTAMP_PENDING :
324 (signalTime == Fence::SIGNAL_TIME_INVALID) ?
325 NATIVE_WINDOW_TIMESTAMP_INVALID :
326 signalTime;
327 }
328 }
329
getFrameTimestamps(uint64_t frameNumber,nsecs_t * outRequestedPresentTime,nsecs_t * outAcquireTime,nsecs_t * outLatchTime,nsecs_t * outFirstRefreshStartTime,nsecs_t * outLastRefreshStartTime,nsecs_t * outGpuCompositionDoneTime,nsecs_t * outDisplayPresentTime,nsecs_t * outDequeueReadyTime,nsecs_t * outReleaseTime)330 status_t Surface::getFrameTimestamps(uint64_t frameNumber,
331 nsecs_t* outRequestedPresentTime, nsecs_t* outAcquireTime,
332 nsecs_t* outLatchTime, nsecs_t* outFirstRefreshStartTime,
333 nsecs_t* outLastRefreshStartTime, nsecs_t* outGpuCompositionDoneTime,
334 nsecs_t* outDisplayPresentTime, nsecs_t* outDequeueReadyTime,
335 nsecs_t* outReleaseTime) {
336 ATRACE_CALL();
337
338 Mutex::Autolock lock(mMutex);
339
340 if (!mEnableFrameTimestamps) {
341 return INVALID_OPERATION;
342 }
343
344 // Verify the requested timestamps are supported.
345 querySupportedTimestampsLocked();
346 if (outDisplayPresentTime != nullptr && !mFrameTimestampsSupportsPresent) {
347 return BAD_VALUE;
348 }
349
350 FrameEvents* events = mFrameEventHistory->getFrame(frameNumber);
351 if (events == nullptr) {
352 // If the entry isn't available in the producer, it's definitely not
353 // available in the consumer.
354 return NAME_NOT_FOUND;
355 }
356
357 // Update our cache of events if the requested events are not available.
358 if (checkConsumerForUpdates(events, mLastFrameNumber,
359 outLatchTime, outFirstRefreshStartTime, outLastRefreshStartTime,
360 outGpuCompositionDoneTime, outDisplayPresentTime,
361 outDequeueReadyTime, outReleaseTime)) {
362 FrameEventHistoryDelta delta;
363 mGraphicBufferProducer->getFrameTimestamps(&delta);
364 mFrameEventHistory->applyDelta(delta);
365 events = mFrameEventHistory->getFrame(frameNumber);
366 }
367
368 if (events == nullptr) {
369 // The entry was available before the update, but was overwritten
370 // after the update. Make sure not to send the wrong frame's data.
371 return NAME_NOT_FOUND;
372 }
373
374 getFrameTimestamp(outRequestedPresentTime, events->requestedPresentTime);
375 getFrameTimestamp(outLatchTime, events->latchTime);
376
377 nsecs_t firstRefreshStartTime = NATIVE_WINDOW_TIMESTAMP_INVALID;
378 getFrameTimestamp(&firstRefreshStartTime, events->firstRefreshStartTime);
379 if (outFirstRefreshStartTime) {
380 *outFirstRefreshStartTime = firstRefreshStartTime;
381 }
382
383 getFrameTimestamp(outLastRefreshStartTime, events->lastRefreshStartTime);
384 getFrameTimestamp(outDequeueReadyTime, events->dequeueReadyTime);
385
386 nsecs_t acquireTime = NATIVE_WINDOW_TIMESTAMP_INVALID;
387 getFrameTimestampFence(&acquireTime, events->acquireFence,
388 events->hasAcquireInfo());
389 if (outAcquireTime != nullptr) {
390 *outAcquireTime = acquireTime;
391 }
392
393 getFrameTimestampFence(outGpuCompositionDoneTime,
394 events->gpuCompositionDoneFence,
395 events->hasGpuCompositionDoneInfo());
396 getFrameTimestampFence(outDisplayPresentTime, events->displayPresentFence,
397 events->hasDisplayPresentInfo());
398 getFrameTimestampFence(outReleaseTime, events->releaseFence,
399 events->hasReleaseInfo());
400
401 // Fix up the GPU completion fence at this layer -- eglGetFrameTimestampsANDROID() expects
402 // that EGL_FIRST_COMPOSITION_GPU_FINISHED_TIME_ANDROID > EGL_RENDERING_COMPLETE_TIME_ANDROID.
403 // This is typically true, but SurfaceFlinger may opt to cache prior GPU composition results,
404 // which breaks that assumption, so zero out GPU composition time.
405 if (outGpuCompositionDoneTime != nullptr
406 && *outGpuCompositionDoneTime > 0 && (acquireTime > 0 || firstRefreshStartTime > 0)
407 && *outGpuCompositionDoneTime <= std::max(acquireTime, firstRefreshStartTime)) {
408 *outGpuCompositionDoneTime = 0;
409 }
410
411 return NO_ERROR;
412 }
413
414 // Deprecated(b/242763577): to be removed, this method should not be used
415 // The reason this method still exists here is to support compiled vndk
416 // Surface support should not be tied to the display
417 // Return true since most displays should have this support
getWideColorSupport(bool * supported)418 status_t Surface::getWideColorSupport(bool* supported) {
419 ATRACE_CALL();
420
421 *supported = true;
422 return NO_ERROR;
423 }
424
425 // Deprecated(b/242763577): to be removed, this method should not be used
426 // The reason this method still exists here is to support compiled vndk
427 // Surface support should not be tied to the display
428 // Return true since most displays should have this support
getHdrSupport(bool * supported)429 status_t Surface::getHdrSupport(bool* supported) {
430 ATRACE_CALL();
431
432 *supported = true;
433 return NO_ERROR;
434 }
435
hook_setSwapInterval(ANativeWindow * window,int interval)436 int Surface::hook_setSwapInterval(ANativeWindow* window, int interval) {
437 Surface* c = getSelf(window);
438 return c->setSwapInterval(interval);
439 }
440
hook_dequeueBuffer(ANativeWindow * window,ANativeWindowBuffer ** buffer,int * fenceFd)441 int Surface::hook_dequeueBuffer(ANativeWindow* window,
442 ANativeWindowBuffer** buffer, int* fenceFd) {
443 Surface* c = getSelf(window);
444 {
445 std::shared_lock<std::shared_mutex> lock(c->mInterceptorMutex);
446 if (c->mDequeueInterceptor != nullptr) {
447 auto interceptor = c->mDequeueInterceptor;
448 auto data = c->mDequeueInterceptorData;
449 return interceptor(window, Surface::dequeueBufferInternal, data, buffer, fenceFd);
450 }
451 }
452 return c->dequeueBuffer(buffer, fenceFd);
453 }
454
dequeueBufferInternal(ANativeWindow * window,ANativeWindowBuffer ** buffer,int * fenceFd)455 int Surface::dequeueBufferInternal(ANativeWindow* window, ANativeWindowBuffer** buffer,
456 int* fenceFd) {
457 Surface* c = getSelf(window);
458 return c->dequeueBuffer(buffer, fenceFd);
459 }
460
hook_cancelBuffer(ANativeWindow * window,ANativeWindowBuffer * buffer,int fenceFd)461 int Surface::hook_cancelBuffer(ANativeWindow* window,
462 ANativeWindowBuffer* buffer, int fenceFd) {
463 Surface* c = getSelf(window);
464 {
465 std::shared_lock<std::shared_mutex> lock(c->mInterceptorMutex);
466 if (c->mCancelInterceptor != nullptr) {
467 auto interceptor = c->mCancelInterceptor;
468 auto data = c->mCancelInterceptorData;
469 return interceptor(window, Surface::cancelBufferInternal, data, buffer, fenceFd);
470 }
471 }
472 return c->cancelBuffer(buffer, fenceFd);
473 }
474
cancelBufferInternal(ANativeWindow * window,ANativeWindowBuffer * buffer,int fenceFd)475 int Surface::cancelBufferInternal(ANativeWindow* window, ANativeWindowBuffer* buffer, int fenceFd) {
476 Surface* c = getSelf(window);
477 return c->cancelBuffer(buffer, fenceFd);
478 }
479
hook_queueBuffer(ANativeWindow * window,ANativeWindowBuffer * buffer,int fenceFd)480 int Surface::hook_queueBuffer(ANativeWindow* window,
481 ANativeWindowBuffer* buffer, int fenceFd) {
482 Surface* c = getSelf(window);
483 {
484 std::shared_lock<std::shared_mutex> lock(c->mInterceptorMutex);
485 if (c->mQueueInterceptor != nullptr) {
486 auto interceptor = c->mQueueInterceptor;
487 auto data = c->mQueueInterceptorData;
488 return interceptor(window, Surface::queueBufferInternal, data, buffer, fenceFd);
489 }
490 }
491 return c->queueBuffer(buffer, fenceFd);
492 }
493
queueBufferInternal(ANativeWindow * window,ANativeWindowBuffer * buffer,int fenceFd)494 int Surface::queueBufferInternal(ANativeWindow* window, ANativeWindowBuffer* buffer, int fenceFd) {
495 Surface* c = getSelf(window);
496 return c->queueBuffer(buffer, fenceFd);
497 }
498
hook_dequeueBuffer_DEPRECATED(ANativeWindow * window,ANativeWindowBuffer ** buffer)499 int Surface::hook_dequeueBuffer_DEPRECATED(ANativeWindow* window,
500 ANativeWindowBuffer** buffer) {
501 Surface* c = getSelf(window);
502 ANativeWindowBuffer* buf;
503 int fenceFd = -1;
504 int result = c->dequeueBuffer(&buf, &fenceFd);
505 if (result != OK) {
506 return result;
507 }
508 sp<Fence> fence(new Fence(fenceFd));
509 int waitResult = fence->waitForever("dequeueBuffer_DEPRECATED");
510 if (waitResult != OK) {
511 ALOGE("dequeueBuffer_DEPRECATED: Fence::wait returned an error: %d",
512 waitResult);
513 c->cancelBuffer(buf, -1);
514 return waitResult;
515 }
516 *buffer = buf;
517 return result;
518 }
519
hook_cancelBuffer_DEPRECATED(ANativeWindow * window,ANativeWindowBuffer * buffer)520 int Surface::hook_cancelBuffer_DEPRECATED(ANativeWindow* window,
521 ANativeWindowBuffer* buffer) {
522 Surface* c = getSelf(window);
523 return c->cancelBuffer(buffer, -1);
524 }
525
hook_lockBuffer_DEPRECATED(ANativeWindow * window,ANativeWindowBuffer * buffer)526 int Surface::hook_lockBuffer_DEPRECATED(ANativeWindow* window,
527 ANativeWindowBuffer* buffer) {
528 Surface* c = getSelf(window);
529 return c->lockBuffer_DEPRECATED(buffer);
530 }
531
hook_queueBuffer_DEPRECATED(ANativeWindow * window,ANativeWindowBuffer * buffer)532 int Surface::hook_queueBuffer_DEPRECATED(ANativeWindow* window,
533 ANativeWindowBuffer* buffer) {
534 Surface* c = getSelf(window);
535 return c->queueBuffer(buffer, -1);
536 }
537
hook_perform(ANativeWindow * window,int operation,...)538 int Surface::hook_perform(ANativeWindow* window, int operation, ...) {
539 va_list args;
540 va_start(args, operation);
541 Surface* c = getSelf(window);
542 int result;
543 // Don't acquire shared ownership of the interceptor mutex if we're going to
544 // do interceptor registration, as otherwise we'll deadlock on acquiring
545 // exclusive ownership.
546 if (!isInterceptorRegistrationOp(operation)) {
547 std::shared_lock<std::shared_mutex> lock(c->mInterceptorMutex);
548 if (c->mPerformInterceptor != nullptr) {
549 result = c->mPerformInterceptor(window, Surface::performInternal,
550 c->mPerformInterceptorData, operation, args);
551 va_end(args);
552 return result;
553 }
554 }
555 result = c->perform(operation, args);
556 va_end(args);
557 return result;
558 }
559
performInternal(ANativeWindow * window,int operation,va_list args)560 int Surface::performInternal(ANativeWindow* window, int operation, va_list args) {
561 Surface* c = getSelf(window);
562 return c->perform(operation, args);
563 }
564
hook_query(const ANativeWindow * window,int what,int * value)565 int Surface::hook_query(const ANativeWindow* window, int what, int* value) {
566 const Surface* c = getSelf(window);
567 {
568 std::shared_lock<std::shared_mutex> lock(c->mInterceptorMutex);
569 if (c->mQueryInterceptor != nullptr) {
570 auto interceptor = c->mQueryInterceptor;
571 auto data = c->mQueryInterceptorData;
572 return interceptor(window, Surface::queryInternal, data, what, value);
573 }
574 }
575 return c->query(what, value);
576 }
577
queryInternal(const ANativeWindow * window,int what,int * value)578 int Surface::queryInternal(const ANativeWindow* window, int what, int* value) {
579 const Surface* c = getSelf(window);
580 return c->query(what, value);
581 }
582
setSwapInterval(int interval)583 int Surface::setSwapInterval(int interval) {
584 ATRACE_CALL();
585 // EGL specification states:
586 // interval is silently clamped to minimum and maximum implementation
587 // dependent values before being stored.
588
589 if (interval < minSwapInterval)
590 interval = minSwapInterval;
591
592 if (interval > maxSwapInterval)
593 interval = maxSwapInterval;
594
595 const bool wasSwapIntervalZero = mSwapIntervalZero;
596 mSwapIntervalZero = (interval == 0);
597
598 if (mSwapIntervalZero != wasSwapIntervalZero) {
599 mGraphicBufferProducer->setAsyncMode(mSwapIntervalZero);
600 }
601
602 return NO_ERROR;
603 }
604
getDequeueBufferInputLocked(IGraphicBufferProducer::DequeueBufferInput * dequeueInput)605 void Surface::getDequeueBufferInputLocked(
606 IGraphicBufferProducer::DequeueBufferInput* dequeueInput) {
607 LOG_ALWAYS_FATAL_IF(dequeueInput == nullptr, "input is null");
608
609 dequeueInput->width = mReqWidth ? mReqWidth : mUserWidth;
610 dequeueInput->height = mReqHeight ? mReqHeight : mUserHeight;
611
612 dequeueInput->format = mReqFormat;
613 dequeueInput->usage = mReqUsage;
614
615 dequeueInput->getTimestamps = mEnableFrameTimestamps;
616 }
617
dequeueBuffer(android_native_buffer_t ** buffer,int * fenceFd)618 int Surface::dequeueBuffer(android_native_buffer_t** buffer, int* fenceFd) {
619 ATRACE_FORMAT("dequeueBuffer - %s", getDebugName());
620 ALOGV("Surface::dequeueBuffer");
621
622 IGraphicBufferProducer::DequeueBufferInput dqInput;
623 {
624 Mutex::Autolock lock(mMutex);
625 if (mReportRemovedBuffers) {
626 mRemovedBuffers.clear();
627 }
628
629 getDequeueBufferInputLocked(&dqInput);
630
631 if (mSharedBufferMode && mAutoRefresh && mSharedBufferSlot !=
632 BufferItem::INVALID_BUFFER_SLOT) {
633 sp<GraphicBuffer>& gbuf(mSlots[mSharedBufferSlot].buffer);
634 if (gbuf != nullptr) {
635 *buffer = gbuf.get();
636 *fenceFd = -1;
637 return OK;
638 }
639 }
640 } // Drop the lock so that we can still touch the Surface while blocking in IGBP::dequeueBuffer
641
642 int buf = -1;
643 sp<Fence> fence;
644 nsecs_t startTime = systemTime();
645
646 FrameEventHistoryDelta frameTimestamps;
647 status_t result = mGraphicBufferProducer->dequeueBuffer(&buf, &fence, dqInput.width,
648 dqInput.height, dqInput.format,
649 dqInput.usage, &mBufferAge,
650 dqInput.getTimestamps ?
651 &frameTimestamps : nullptr);
652 mLastDequeueDuration = systemTime() - startTime;
653
654 if (result < 0) {
655 ALOGV("dequeueBuffer: IGraphicBufferProducer::dequeueBuffer"
656 "(%d, %d, %d, %#" PRIx64 ") failed: %d",
657 dqInput.width, dqInput.height, dqInput.format, dqInput.usage, result);
658 return result;
659 }
660
661 if (buf < 0 || buf >= NUM_BUFFER_SLOTS) {
662 ALOGE("dequeueBuffer: IGraphicBufferProducer returned invalid slot number %d", buf);
663 android_errorWriteLog(0x534e4554, "36991414"); // SafetyNet logging
664 return FAILED_TRANSACTION;
665 }
666
667 Mutex::Autolock lock(mMutex);
668
669 // Write this while holding the mutex
670 mLastDequeueStartTime = startTime;
671
672 sp<GraphicBuffer>& gbuf(mSlots[buf].buffer);
673
674 // this should never happen
675 ALOGE_IF(fence == nullptr, "Surface::dequeueBuffer: received null Fence! buf=%d", buf);
676
677 if (CC_UNLIKELY(atrace_is_tag_enabled(ATRACE_TAG_GRAPHICS))) {
678 static gui::FenceMonitor hwcReleaseThread("HWC release");
679 hwcReleaseThread.queueFence(fence);
680 }
681
682 if (result & IGraphicBufferProducer::RELEASE_ALL_BUFFERS) {
683 freeAllBuffers();
684 }
685
686 if (dqInput.getTimestamps) {
687 mFrameEventHistory->applyDelta(frameTimestamps);
688 }
689
690 if ((result & IGraphicBufferProducer::BUFFER_NEEDS_REALLOCATION) || gbuf == nullptr) {
691 if (mReportRemovedBuffers && (gbuf != nullptr)) {
692 mRemovedBuffers.push_back(gbuf);
693 }
694 result = mGraphicBufferProducer->requestBuffer(buf, &gbuf);
695 if (result != NO_ERROR) {
696 ALOGE("dequeueBuffer: IGraphicBufferProducer::requestBuffer failed: %d", result);
697 mGraphicBufferProducer->cancelBuffer(buf, fence);
698 return result;
699 }
700 }
701
702 if (fence->isValid()) {
703 *fenceFd = fence->dup();
704 if (*fenceFd == -1) {
705 ALOGE("dequeueBuffer: error duping fence: %d", errno);
706 // dup() should never fail; something is badly wrong. Soldier on
707 // and hope for the best; the worst that should happen is some
708 // visible corruption that lasts until the next frame.
709 }
710 } else {
711 *fenceFd = -1;
712 }
713
714 *buffer = gbuf.get();
715
716 if (mSharedBufferMode && mAutoRefresh) {
717 mSharedBufferSlot = buf;
718 mSharedBufferHasBeenQueued = false;
719 } else if (mSharedBufferSlot == buf) {
720 mSharedBufferSlot = BufferItem::INVALID_BUFFER_SLOT;
721 mSharedBufferHasBeenQueued = false;
722 }
723
724 mDequeuedSlots.insert(buf);
725
726 return OK;
727 }
728
729 #if COM_ANDROID_GRAPHICS_LIBGUI_FLAGS(WB_PLATFORM_API_IMPROVEMENTS)
730
dequeueBuffer(sp<GraphicBuffer> * buffer,sp<Fence> * outFence)731 status_t Surface::dequeueBuffer(sp<GraphicBuffer>* buffer, sp<Fence>* outFence) {
732 if (buffer == nullptr || outFence == nullptr) {
733 return BAD_VALUE;
734 }
735
736 android_native_buffer_t* anb;
737 int fd = -1;
738 status_t res = dequeueBuffer(&anb, &fd);
739 *buffer = GraphicBuffer::from(anb);
740 *outFence = sp<Fence>::make(fd);
741 return res;
742 }
743
queueBuffer(const sp<GraphicBuffer> & buffer,const sp<Fence> & fd,SurfaceQueueBufferOutput * output)744 status_t Surface::queueBuffer(const sp<GraphicBuffer>& buffer, const sp<Fence>& fd,
745 SurfaceQueueBufferOutput* output) {
746 if (buffer == nullptr) {
747 return BAD_VALUE;
748 }
749 return queueBuffer(buffer.get(), fd ? fd->get() : -1, output);
750 }
751
detachBuffer(const sp<GraphicBuffer> & buffer)752 status_t Surface::detachBuffer(const sp<GraphicBuffer>& buffer) {
753 if (nullptr == buffer) {
754 return BAD_VALUE;
755 }
756
757 Mutex::Autolock lock(mMutex);
758
759 uint64_t bufferId = buffer->getId();
760 for (int slot = 0; slot < Surface::NUM_BUFFER_SLOTS; ++slot) {
761 auto& bufferSlot = mSlots[slot];
762 if (bufferSlot.buffer != nullptr && bufferSlot.buffer->getId() == bufferId) {
763 bufferSlot.buffer = nullptr;
764 bufferSlot.dirtyRegion = Region::INVALID_REGION;
765 return mGraphicBufferProducer->detachBuffer(slot);
766 }
767 }
768
769 return BAD_VALUE;
770 }
771
772 #endif // COM_ANDROID_GRAPHICS_LIBGUI_FLAGS(WB_PLATFORM_API_IMPROVEMENTS)
773
dequeueBuffers(std::vector<BatchBuffer> * buffers)774 int Surface::dequeueBuffers(std::vector<BatchBuffer>* buffers) {
775 using DequeueBufferInput = IGraphicBufferProducer::DequeueBufferInput;
776 using DequeueBufferOutput = IGraphicBufferProducer::DequeueBufferOutput;
777 using CancelBufferInput = IGraphicBufferProducer::CancelBufferInput;
778 using RequestBufferOutput = IGraphicBufferProducer::RequestBufferOutput;
779
780 ATRACE_CALL();
781 ALOGV("Surface::dequeueBuffers");
782
783 if (buffers->size() == 0) {
784 ALOGE("%s: must dequeue at least 1 buffer!", __FUNCTION__);
785 return BAD_VALUE;
786 }
787
788 if (mSharedBufferMode) {
789 ALOGE("%s: batch operation is not supported in shared buffer mode!",
790 __FUNCTION__);
791 return INVALID_OPERATION;
792 }
793
794 size_t numBufferRequested = buffers->size();
795 DequeueBufferInput input;
796
797 {
798 Mutex::Autolock lock(mMutex);
799 if (mReportRemovedBuffers) {
800 mRemovedBuffers.clear();
801 }
802
803 getDequeueBufferInputLocked(&input);
804 } // Drop the lock so that we can still touch the Surface while blocking in IGBP::dequeueBuffers
805
806 std::vector<DequeueBufferInput> dequeueInput(numBufferRequested, input);
807 std::vector<DequeueBufferOutput> dequeueOutput;
808
809 nsecs_t startTime = systemTime();
810
811 status_t result = mGraphicBufferProducer->dequeueBuffers(dequeueInput, &dequeueOutput);
812
813 mLastDequeueDuration = systemTime() - startTime;
814
815 if (result < 0) {
816 ALOGV("%s: IGraphicBufferProducer::dequeueBuffers"
817 "(%d, %d, %d, %#" PRIx64 ") failed: %d",
818 __FUNCTION__, input.width, input.height, input.format, input.usage, result);
819 return result;
820 }
821
822 std::vector<CancelBufferInput> cancelBufferInputs;
823 cancelBufferInputs.reserve(numBufferRequested);
824 std::vector<status_t> cancelBufferOutputs;
825 for (size_t i = 0; i < numBufferRequested; i++) {
826 if (dequeueOutput[i].result >= 0) {
827 CancelBufferInput& input = cancelBufferInputs.emplace_back();
828 input.slot = dequeueOutput[i].slot;
829 input.fence = dequeueOutput[i].fence;
830 }
831 }
832
833 for (const auto& output : dequeueOutput) {
834 if (output.result < 0) {
835 mGraphicBufferProducer->cancelBuffers(cancelBufferInputs, &cancelBufferOutputs);
836 ALOGV("%s: IGraphicBufferProducer::dequeueBuffers"
837 "(%d, %d, %d, %#" PRIx64 ") failed: %d",
838 __FUNCTION__, input.width, input.height, input.format, input.usage,
839 output.result);
840 return output.result;
841 }
842
843 if (output.slot < 0 || output.slot >= NUM_BUFFER_SLOTS) {
844 mGraphicBufferProducer->cancelBuffers(cancelBufferInputs, &cancelBufferOutputs);
845 ALOGE("%s: IGraphicBufferProducer returned invalid slot number %d",
846 __FUNCTION__, output.slot);
847 android_errorWriteLog(0x534e4554, "36991414"); // SafetyNet logging
848 return FAILED_TRANSACTION;
849 }
850
851 if (input.getTimestamps && !output.timestamps.has_value()) {
852 mGraphicBufferProducer->cancelBuffers(cancelBufferInputs, &cancelBufferOutputs);
853 ALOGE("%s: no frame timestamp returns!", __FUNCTION__);
854 return FAILED_TRANSACTION;
855 }
856
857 // this should never happen
858 ALOGE_IF(output.fence == nullptr,
859 "%s: received null Fence! slot=%d", __FUNCTION__, output.slot);
860 }
861
862 Mutex::Autolock lock(mMutex);
863
864 // Write this while holding the mutex
865 mLastDequeueStartTime = startTime;
866
867 std::vector<int32_t> requestBufferSlots;
868 requestBufferSlots.reserve(numBufferRequested);
869 // handle release all buffers and request buffers
870 for (const auto& output : dequeueOutput) {
871 if (output.result & IGraphicBufferProducer::RELEASE_ALL_BUFFERS) {
872 ALOGV("%s: RELEASE_ALL_BUFFERS during batch operation", __FUNCTION__);
873 freeAllBuffers();
874 break;
875 }
876 }
877
878 for (const auto& output : dequeueOutput) {
879 // Collect slots that needs requesting buffer
880 sp<GraphicBuffer>& gbuf(mSlots[output.slot].buffer);
881 if ((result & IGraphicBufferProducer::BUFFER_NEEDS_REALLOCATION) || gbuf == nullptr) {
882 if (mReportRemovedBuffers && (gbuf != nullptr)) {
883 mRemovedBuffers.push_back(gbuf);
884 }
885 requestBufferSlots.push_back(output.slot);
886 }
887 }
888
889 // Batch request Buffer
890 std::vector<RequestBufferOutput> reqBufferOutput;
891 if (requestBufferSlots.size() > 0) {
892 result = mGraphicBufferProducer->requestBuffers(requestBufferSlots, &reqBufferOutput);
893 if (result != NO_ERROR) {
894 ALOGE("%s: IGraphicBufferProducer::requestBuffers failed: %d",
895 __FUNCTION__, result);
896 mGraphicBufferProducer->cancelBuffers(cancelBufferInputs, &cancelBufferOutputs);
897 return result;
898 }
899
900 // Check if we have any single failure
901 for (size_t i = 0; i < requestBufferSlots.size(); i++) {
902 if (reqBufferOutput[i].result != OK) {
903 ALOGE("%s: IGraphicBufferProducer::requestBuffers failed at %zu-th buffer, slot %d",
904 __FUNCTION__, i, requestBufferSlots[i]);
905 mGraphicBufferProducer->cancelBuffers(cancelBufferInputs, &cancelBufferOutputs);
906 return reqBufferOutput[i].result;
907 }
908 }
909
910 // Fill request buffer results to mSlots
911 for (size_t i = 0; i < requestBufferSlots.size(); i++) {
912 mSlots[requestBufferSlots[i]].buffer = reqBufferOutput[i].buffer;
913 }
914 }
915
916 for (size_t batchIdx = 0; batchIdx < numBufferRequested; batchIdx++) {
917 const auto& output = dequeueOutput[batchIdx];
918 int slot = output.slot;
919 sp<GraphicBuffer>& gbuf(mSlots[slot].buffer);
920
921 if (CC_UNLIKELY(atrace_is_tag_enabled(ATRACE_TAG_GRAPHICS))) {
922 static gui::FenceMonitor hwcReleaseThread("HWC release");
923 hwcReleaseThread.queueFence(output.fence);
924 }
925
926 if (input.getTimestamps) {
927 mFrameEventHistory->applyDelta(output.timestamps.value());
928 }
929
930 if (output.fence->isValid()) {
931 buffers->at(batchIdx).fenceFd = output.fence->dup();
932 if (buffers->at(batchIdx).fenceFd == -1) {
933 ALOGE("%s: error duping fence: %d", __FUNCTION__, errno);
934 // dup() should never fail; something is badly wrong. Soldier on
935 // and hope for the best; the worst that should happen is some
936 // visible corruption that lasts until the next frame.
937 }
938 } else {
939 buffers->at(batchIdx).fenceFd = -1;
940 }
941
942 buffers->at(batchIdx).buffer = gbuf.get();
943 mDequeuedSlots.insert(slot);
944 }
945 return OK;
946 }
947
cancelBuffer(android_native_buffer_t * buffer,int fenceFd)948 int Surface::cancelBuffer(android_native_buffer_t* buffer,
949 int fenceFd) {
950 ATRACE_CALL();
951 ALOGV("Surface::cancelBuffer");
952 Mutex::Autolock lock(mMutex);
953 int i = getSlotFromBufferLocked(buffer);
954 if (i < 0) {
955 if (fenceFd >= 0) {
956 close(fenceFd);
957 }
958 return i;
959 }
960 if (mSharedBufferSlot == i && mSharedBufferHasBeenQueued) {
961 if (fenceFd >= 0) {
962 close(fenceFd);
963 }
964 return OK;
965 }
966 sp<Fence> fence(fenceFd >= 0 ? new Fence(fenceFd) : Fence::NO_FENCE);
967 mGraphicBufferProducer->cancelBuffer(i, fence);
968
969 if (mSharedBufferMode && mAutoRefresh && mSharedBufferSlot == i) {
970 mSharedBufferHasBeenQueued = true;
971 }
972
973 mDequeuedSlots.erase(i);
974
975 return OK;
976 }
977
cancelBuffers(const std::vector<BatchBuffer> & buffers)978 int Surface::cancelBuffers(const std::vector<BatchBuffer>& buffers) {
979 using CancelBufferInput = IGraphicBufferProducer::CancelBufferInput;
980 ATRACE_CALL();
981 ALOGV("Surface::cancelBuffers");
982
983 if (mSharedBufferMode) {
984 ALOGE("%s: batch operation is not supported in shared buffer mode!",
985 __FUNCTION__);
986 return INVALID_OPERATION;
987 }
988
989 size_t numBuffers = buffers.size();
990 std::vector<CancelBufferInput> cancelBufferInputs(numBuffers);
991 std::vector<status_t> cancelBufferOutputs;
992 size_t numBuffersCancelled = 0;
993 int badSlotResult = 0;
994 for (size_t i = 0; i < numBuffers; i++) {
995 int slot = getSlotFromBufferLocked(buffers[i].buffer);
996 int fenceFd = buffers[i].fenceFd;
997 if (slot < 0) {
998 if (fenceFd >= 0) {
999 close(fenceFd);
1000 }
1001 ALOGE("%s: cannot find slot number for cancelled buffer", __FUNCTION__);
1002 badSlotResult = slot;
1003 } else {
1004 sp<Fence> fence(fenceFd >= 0 ? new Fence(fenceFd) : Fence::NO_FENCE);
1005 cancelBufferInputs[numBuffersCancelled].slot = slot;
1006 cancelBufferInputs[numBuffersCancelled++].fence = fence;
1007 }
1008 }
1009 cancelBufferInputs.resize(numBuffersCancelled);
1010 mGraphicBufferProducer->cancelBuffers(cancelBufferInputs, &cancelBufferOutputs);
1011
1012
1013 for (size_t i = 0; i < numBuffersCancelled; i++) {
1014 mDequeuedSlots.erase(cancelBufferInputs[i].slot);
1015 }
1016
1017 if (badSlotResult != 0) {
1018 return badSlotResult;
1019 }
1020 return OK;
1021 }
1022
getSlotFromBufferLocked(android_native_buffer_t * buffer) const1023 int Surface::getSlotFromBufferLocked(
1024 android_native_buffer_t* buffer) const {
1025 if (buffer == nullptr) {
1026 ALOGE("%s: input buffer is null!", __FUNCTION__);
1027 return BAD_VALUE;
1028 }
1029
1030 for (int i = 0; i < NUM_BUFFER_SLOTS; i++) {
1031 if (mSlots[i].buffer != nullptr &&
1032 mSlots[i].buffer->handle == buffer->handle) {
1033 return i;
1034 }
1035 }
1036 ALOGE("%s: unknown buffer: %p", __FUNCTION__, buffer->handle);
1037 return BAD_VALUE;
1038 }
1039
lockBuffer_DEPRECATED(android_native_buffer_t * buffer)1040 int Surface::lockBuffer_DEPRECATED(android_native_buffer_t* buffer __attribute__((unused))) {
1041 ALOGV("Surface::lockBuffer");
1042 Mutex::Autolock lock(mMutex);
1043 return OK;
1044 }
1045
getQueueBufferInputLocked(android_native_buffer_t * buffer,int fenceFd,nsecs_t timestamp,IGraphicBufferProducer::QueueBufferInput * out)1046 void Surface::getQueueBufferInputLocked(android_native_buffer_t* buffer, int fenceFd,
1047 nsecs_t timestamp, IGraphicBufferProducer::QueueBufferInput* out) {
1048 bool isAutoTimestamp = false;
1049
1050 if (timestamp == NATIVE_WINDOW_TIMESTAMP_AUTO) {
1051 timestamp = systemTime(SYSTEM_TIME_MONOTONIC);
1052 isAutoTimestamp = true;
1053 ALOGV("Surface::queueBuffer making up timestamp: %.2f ms",
1054 timestamp / 1000000.0);
1055 }
1056
1057 // Make sure the crop rectangle is entirely inside the buffer.
1058 Rect crop(Rect::EMPTY_RECT);
1059 mCrop.intersect(Rect(buffer->width, buffer->height), &crop);
1060
1061 sp<Fence> fence(fenceFd >= 0 ? new Fence(fenceFd) : Fence::NO_FENCE);
1062 IGraphicBufferProducer::QueueBufferInput input(timestamp, isAutoTimestamp,
1063 static_cast<android_dataspace>(mDataSpace), crop, mScalingMode,
1064 mTransform ^ mStickyTransform, fence, mStickyTransform,
1065 mEnableFrameTimestamps);
1066
1067 // we should send HDR metadata as needed if this becomes a bottleneck
1068 input.setHdrMetadata(mHdrMetadata);
1069
1070 if (mConnectedToCpu || mDirtyRegion.bounds() == Rect::INVALID_RECT) {
1071 input.setSurfaceDamage(Region::INVALID_REGION);
1072 } else {
1073 // Here we do two things:
1074 // 1) The surface damage was specified using the OpenGL ES convention of
1075 // the origin being in the bottom-left corner. Here we flip to the
1076 // convention that the rest of the system uses (top-left corner) by
1077 // subtracting all top/bottom coordinates from the buffer height.
1078 // 2) If the buffer is coming in rotated (for example, because the EGL
1079 // implementation is reacting to the transform hint coming back from
1080 // SurfaceFlinger), the surface damage needs to be rotated the
1081 // opposite direction, since it was generated assuming an unrotated
1082 // buffer (the app doesn't know that the EGL implementation is
1083 // reacting to the transform hint behind its back). The
1084 // transformations in the switch statement below apply those
1085 // complementary rotations (e.g., if 90 degrees, rotate 270 degrees).
1086
1087 int width = buffer->width;
1088 int height = buffer->height;
1089 bool rotated90 = (mTransform ^ mStickyTransform) &
1090 NATIVE_WINDOW_TRANSFORM_ROT_90;
1091 if (rotated90) {
1092 std::swap(width, height);
1093 }
1094
1095 Region flippedRegion;
1096 for (auto rect : mDirtyRegion) {
1097 int left = rect.left;
1098 int right = rect.right;
1099 int top = height - rect.bottom; // Flip from OpenGL convention
1100 int bottom = height - rect.top; // Flip from OpenGL convention
1101 switch (mTransform ^ mStickyTransform) {
1102 case NATIVE_WINDOW_TRANSFORM_ROT_90: {
1103 // Rotate 270 degrees
1104 Rect flippedRect{top, width - right, bottom, width - left};
1105 flippedRegion.orSelf(flippedRect);
1106 break;
1107 }
1108 case NATIVE_WINDOW_TRANSFORM_ROT_180: {
1109 // Rotate 180 degrees
1110 Rect flippedRect{width - right, height - bottom,
1111 width - left, height - top};
1112 flippedRegion.orSelf(flippedRect);
1113 break;
1114 }
1115 case NATIVE_WINDOW_TRANSFORM_ROT_270: {
1116 // Rotate 90 degrees
1117 Rect flippedRect{height - bottom, left,
1118 height - top, right};
1119 flippedRegion.orSelf(flippedRect);
1120 break;
1121 }
1122 default: {
1123 Rect flippedRect{left, top, right, bottom};
1124 flippedRegion.orSelf(flippedRect);
1125 break;
1126 }
1127 }
1128 }
1129
1130 input.setSurfaceDamage(flippedRegion);
1131 }
1132 *out = input;
1133 }
1134
applyGrallocMetadataLocked(android_native_buffer_t * buffer,const IGraphicBufferProducer::QueueBufferInput & queueBufferInput)1135 void Surface::applyGrallocMetadataLocked(
1136 android_native_buffer_t* buffer,
1137 const IGraphicBufferProducer::QueueBufferInput& queueBufferInput) {
1138 ATRACE_CALL();
1139 auto& mapper = GraphicBufferMapper::get();
1140 mapper.setDataspace(buffer->handle, static_cast<ui::Dataspace>(queueBufferInput.dataSpace));
1141 if (mHdrMetadataIsSet & HdrMetadata::SMPTE2086)
1142 mapper.setSmpte2086(buffer->handle, queueBufferInput.getHdrMetadata().getSmpte2086());
1143 if (mHdrMetadataIsSet & HdrMetadata::CTA861_3)
1144 mapper.setCta861_3(buffer->handle, queueBufferInput.getHdrMetadata().getCta8613());
1145 if (mHdrMetadataIsSet & HdrMetadata::HDR10PLUS)
1146 mapper.setSmpte2094_40(buffer->handle, queueBufferInput.getHdrMetadata().getHdr10Plus());
1147 }
1148
onBufferQueuedLocked(int slot,sp<Fence> fence,const IGraphicBufferProducer::QueueBufferOutput & output)1149 void Surface::onBufferQueuedLocked(int slot, sp<Fence> fence,
1150 const IGraphicBufferProducer::QueueBufferOutput& output) {
1151 mDequeuedSlots.erase(slot);
1152
1153 if (mEnableFrameTimestamps) {
1154 mFrameEventHistory->applyDelta(output.frameTimestamps);
1155 // Update timestamps with the local acquire fence.
1156 // The consumer doesn't send it back to prevent us from having two
1157 // file descriptors of the same fence.
1158 mFrameEventHistory->updateAcquireFence(mNextFrameNumber,
1159 std::make_shared<FenceTime>(fence));
1160
1161 // Cache timestamps of signaled fences so we can close their file
1162 // descriptors.
1163 mFrameEventHistory->updateSignalTimes();
1164 }
1165
1166 mLastFrameNumber = mNextFrameNumber;
1167
1168 mDefaultWidth = output.width;
1169 mDefaultHeight = output.height;
1170 mNextFrameNumber = output.nextFrameNumber;
1171
1172 // Ignore transform hint if sticky transform is set or transform to display inverse flag is
1173 // set.
1174 if (mStickyTransform == 0 && !transformToDisplayInverse()) {
1175 mTransformHint = output.transformHint;
1176 }
1177
1178 mConsumerRunningBehind = (output.numPendingBuffers >= 2);
1179
1180 if (!mConnectedToCpu) {
1181 // Clear surface damage back to full-buffer
1182 mDirtyRegion = Region::INVALID_REGION;
1183 }
1184
1185 if (mSharedBufferMode && mAutoRefresh && mSharedBufferSlot == slot) {
1186 mSharedBufferHasBeenQueued = true;
1187 }
1188
1189 mQueueBufferCondition.broadcast();
1190
1191 if (CC_UNLIKELY(atrace_is_tag_enabled(ATRACE_TAG_GRAPHICS))) {
1192 static gui::FenceMonitor gpuCompletionThread("GPU completion");
1193 gpuCompletionThread.queueFence(fence);
1194 }
1195 }
1196
1197 #if COM_ANDROID_GRAPHICS_LIBGUI_FLAGS(WB_PLATFORM_API_IMPROVEMENTS)
1198
queueBuffer(android_native_buffer_t * buffer,int fenceFd,SurfaceQueueBufferOutput * surfaceOutput)1199 int Surface::queueBuffer(android_native_buffer_t* buffer, int fenceFd,
1200 SurfaceQueueBufferOutput* surfaceOutput) {
1201 ATRACE_CALL();
1202 ALOGV("Surface::queueBuffer");
1203
1204 IGraphicBufferProducer::QueueBufferOutput output;
1205 IGraphicBufferProducer::QueueBufferInput input;
1206 int slot;
1207 sp<Fence> fence;
1208 {
1209 Mutex::Autolock lock(mMutex);
1210
1211 slot = getSlotFromBufferLocked(buffer);
1212 if (slot < 0) {
1213 if (fenceFd >= 0) {
1214 close(fenceFd);
1215 }
1216 return slot;
1217 }
1218 if (mSharedBufferSlot == slot && mSharedBufferHasBeenQueued) {
1219 if (fenceFd >= 0) {
1220 close(fenceFd);
1221 }
1222 return OK;
1223 }
1224
1225 getQueueBufferInputLocked(buffer, fenceFd, mTimestamp, &input);
1226 applyGrallocMetadataLocked(buffer, input);
1227 fence = input.fence;
1228 }
1229 nsecs_t now = systemTime();
1230 // Drop the lock temporarily while we touch the underlying producer. In the case of a local
1231 // BufferQueue, the following should be allowable:
1232 //
1233 // Surface::queueBuffer
1234 // -> IConsumerListener::onFrameAvailable callback triggers automatically
1235 // -> implementation calls IGraphicBufferConsumer::acquire/release immediately
1236 // -> SurfaceListener::onBufferRelesed callback triggers automatically
1237 // -> implementation calls Surface::dequeueBuffer
1238 status_t err = mGraphicBufferProducer->queueBuffer(slot, input, &output);
1239 {
1240 Mutex::Autolock lock(mMutex);
1241
1242 mLastQueueDuration = systemTime() - now;
1243 if (err != OK) {
1244 ALOGE("queueBuffer: error queuing buffer, %d", err);
1245 }
1246
1247 onBufferQueuedLocked(slot, fence, output);
1248 }
1249
1250 if (surfaceOutput != nullptr) {
1251 *surfaceOutput = {.bufferReplaced = output.bufferReplaced};
1252 }
1253
1254 return err;
1255 }
1256
1257 #if COM_ANDROID_GRAPHICS_LIBGUI_FLAGS(WB_PLATFORM_API_IMPROVEMENTS)
queueBuffers(const std::vector<BatchQueuedBuffer> & buffers,std::vector<SurfaceQueueBufferOutput> * queueBufferOutputs)1258 int Surface::queueBuffers(const std::vector<BatchQueuedBuffer>& buffers,
1259 std::vector<SurfaceQueueBufferOutput>* queueBufferOutputs)
1260 #else
1261 int Surface::queueBuffers(const std::vector<BatchQueuedBuffer>& buffers)
1262 #endif // COM_ANDROID_GRAPHICS_LIBGUI_FLAGS(WB_PLATFORM_API_IMPROVEMENTS)
1263 {
1264 ATRACE_CALL();
1265 ALOGV("Surface::queueBuffers");
1266
1267 size_t numBuffers = buffers.size();
1268 std::vector<IGraphicBufferProducer::QueueBufferInput> igbpQueueBufferInputs(numBuffers);
1269 std::vector<IGraphicBufferProducer::QueueBufferOutput> igbpQueueBufferOutputs;
1270 std::vector<int> bufferSlots(numBuffers, -1);
1271 std::vector<sp<Fence>> bufferFences(numBuffers);
1272
1273 int err;
1274 {
1275 Mutex::Autolock lock(mMutex);
1276
1277 if (mSharedBufferMode) {
1278 ALOGE("%s: batched operation is not supported in shared buffer mode", __FUNCTION__);
1279 return INVALID_OPERATION;
1280 }
1281
1282 for (size_t batchIdx = 0; batchIdx < numBuffers; batchIdx++) {
1283 int i = getSlotFromBufferLocked(buffers[batchIdx].buffer);
1284 if (i < 0) {
1285 if (buffers[batchIdx].fenceFd >= 0) {
1286 close(buffers[batchIdx].fenceFd);
1287 }
1288 return i;
1289 }
1290 bufferSlots[batchIdx] = i;
1291
1292 IGraphicBufferProducer::QueueBufferInput input;
1293 getQueueBufferInputLocked(buffers[batchIdx].buffer, buffers[batchIdx].fenceFd,
1294 buffers[batchIdx].timestamp, &input);
1295 input.slot = i;
1296 bufferFences[batchIdx] = input.fence;
1297 igbpQueueBufferInputs[batchIdx] = input;
1298 }
1299 }
1300 nsecs_t now = systemTime();
1301 err = mGraphicBufferProducer->queueBuffers(igbpQueueBufferInputs, &igbpQueueBufferOutputs);
1302 {
1303 Mutex::Autolock lock(mMutex);
1304 mLastQueueDuration = systemTime() - now;
1305 if (err != OK) {
1306 ALOGE("%s: error queuing buffer, %d", __FUNCTION__, err);
1307 }
1308
1309 for (size_t batchIdx = 0; batchIdx < numBuffers; batchIdx++) {
1310 onBufferQueuedLocked(bufferSlots[batchIdx], bufferFences[batchIdx],
1311 igbpQueueBufferOutputs[batchIdx]);
1312 }
1313 }
1314
1315 #if COM_ANDROID_GRAPHICS_LIBGUI_FLAGS(WB_PLATFORM_API_IMPROVEMENTS)
1316 if (queueBufferOutputs != nullptr) {
1317 queueBufferOutputs->clear();
1318 queueBufferOutputs->resize(numBuffers);
1319 for (size_t batchIdx = 0; batchIdx < numBuffers; batchIdx++) {
1320 (*queueBufferOutputs)[batchIdx].bufferReplaced =
1321 igbpQueueBufferOutputs[batchIdx].bufferReplaced;
1322 }
1323 }
1324 #endif // COM_ANDROID_GRAPHICS_LIBGUI_FLAGS(WB_PLATFORM_API_IMPROVEMENTS)
1325
1326 return err;
1327 }
1328
1329 #else
1330
queueBuffer(android_native_buffer_t * buffer,int fenceFd)1331 int Surface::queueBuffer(android_native_buffer_t* buffer, int fenceFd) {
1332 ATRACE_CALL();
1333 ALOGV("Surface::queueBuffer");
1334 Mutex::Autolock lock(mMutex);
1335
1336 int i = getSlotFromBufferLocked(buffer);
1337 if (i < 0) {
1338 if (fenceFd >= 0) {
1339 close(fenceFd);
1340 }
1341 return i;
1342 }
1343 if (mSharedBufferSlot == i && mSharedBufferHasBeenQueued) {
1344 if (fenceFd >= 0) {
1345 close(fenceFd);
1346 }
1347 return OK;
1348 }
1349
1350 IGraphicBufferProducer::QueueBufferOutput output;
1351 IGraphicBufferProducer::QueueBufferInput input;
1352 getQueueBufferInputLocked(buffer, fenceFd, mTimestamp, &input);
1353 applyGrallocMetadataLocked(buffer, input);
1354 sp<Fence> fence = input.fence;
1355
1356 nsecs_t now = systemTime();
1357
1358 status_t err = mGraphicBufferProducer->queueBuffer(i, input, &output);
1359 mLastQueueDuration = systemTime() - now;
1360 if (err != OK) {
1361 ALOGE("queueBuffer: error queuing buffer, %d", err);
1362 }
1363
1364 onBufferQueuedLocked(i, fence, output);
1365 return err;
1366 }
1367
queueBuffers(const std::vector<BatchQueuedBuffer> & buffers)1368 int Surface::queueBuffers(const std::vector<BatchQueuedBuffer>& buffers) {
1369 ATRACE_CALL();
1370 ALOGV("Surface::queueBuffers");
1371 Mutex::Autolock lock(mMutex);
1372
1373 if (mSharedBufferMode) {
1374 ALOGE("%s: batched operation is not supported in shared buffer mode", __FUNCTION__);
1375 return INVALID_OPERATION;
1376 }
1377
1378 size_t numBuffers = buffers.size();
1379 std::vector<IGraphicBufferProducer::QueueBufferInput> queueBufferInputs(numBuffers);
1380 std::vector<IGraphicBufferProducer::QueueBufferOutput> queueBufferOutputs;
1381 std::vector<int> bufferSlots(numBuffers, -1);
1382 std::vector<sp<Fence>> bufferFences(numBuffers);
1383
1384 for (size_t batchIdx = 0; batchIdx < numBuffers; batchIdx++) {
1385 int i = getSlotFromBufferLocked(buffers[batchIdx].buffer);
1386 if (i < 0) {
1387 if (buffers[batchIdx].fenceFd >= 0) {
1388 close(buffers[batchIdx].fenceFd);
1389 }
1390 return i;
1391 }
1392 bufferSlots[batchIdx] = i;
1393
1394 IGraphicBufferProducer::QueueBufferInput input;
1395 getQueueBufferInputLocked(
1396 buffers[batchIdx].buffer, buffers[batchIdx].fenceFd, buffers[batchIdx].timestamp,
1397 &input);
1398 bufferFences[batchIdx] = input.fence;
1399 queueBufferInputs[batchIdx] = input;
1400 }
1401
1402 nsecs_t now = systemTime();
1403 status_t err = mGraphicBufferProducer->queueBuffers(queueBufferInputs, &queueBufferOutputs);
1404 mLastQueueDuration = systemTime() - now;
1405 if (err != OK) {
1406 ALOGE("%s: error queuing buffer, %d", __FUNCTION__, err);
1407 }
1408
1409
1410 for (size_t batchIdx = 0; batchIdx < numBuffers; batchIdx++) {
1411 onBufferQueuedLocked(bufferSlots[batchIdx], bufferFences[batchIdx],
1412 queueBufferOutputs[batchIdx]);
1413 }
1414
1415 return err;
1416 }
1417
1418 #endif // COM_ANDROID_GRAPHICS_LIBGUI_FLAGS(WB_PLATFORM_API_IMPROVEMENTS)
1419
querySupportedTimestampsLocked() const1420 void Surface::querySupportedTimestampsLocked() const {
1421 // mMutex must be locked when calling this method.
1422
1423 if (mQueriedSupportedTimestamps) {
1424 return;
1425 }
1426 mQueriedSupportedTimestamps = true;
1427
1428 std::vector<FrameEvent> supportedFrameTimestamps;
1429 binder::Status status =
1430 composerServiceAIDL()->getSupportedFrameTimestamps(&supportedFrameTimestamps);
1431
1432 if (!status.isOk()) {
1433 return;
1434 }
1435
1436 for (auto sft : supportedFrameTimestamps) {
1437 if (sft == FrameEvent::DISPLAY_PRESENT) {
1438 mFrameTimestampsSupportsPresent = true;
1439 }
1440 }
1441 }
1442
query(int what,int * value) const1443 int Surface::query(int what, int* value) const {
1444 ATRACE_CALL();
1445 ALOGV("Surface::query");
1446 { // scope for the lock
1447 Mutex::Autolock lock(mMutex);
1448 switch (what) {
1449 case NATIVE_WINDOW_FORMAT:
1450 if (mReqFormat) {
1451 *value = static_cast<int>(mReqFormat);
1452 return NO_ERROR;
1453 }
1454 break;
1455 case NATIVE_WINDOW_QUEUES_TO_WINDOW_COMPOSER: {
1456 status_t err = mGraphicBufferProducer->query(what, value);
1457 if (err == NO_ERROR) {
1458 return NO_ERROR;
1459 }
1460 sp<gui::ISurfaceComposer> surfaceComposer = composerServiceAIDL();
1461 if (surfaceComposer == nullptr) {
1462 return -EPERM; // likely permissions error
1463 }
1464 // ISurfaceComposer no longer supports authenticateSurfaceTexture
1465 *value = 0;
1466 return NO_ERROR;
1467 }
1468 case NATIVE_WINDOW_CONCRETE_TYPE:
1469 *value = NATIVE_WINDOW_SURFACE;
1470 return NO_ERROR;
1471 case NATIVE_WINDOW_DEFAULT_WIDTH:
1472 *value = static_cast<int>(
1473 mUserWidth ? mUserWidth : mDefaultWidth);
1474 return NO_ERROR;
1475 case NATIVE_WINDOW_DEFAULT_HEIGHT:
1476 *value = static_cast<int>(
1477 mUserHeight ? mUserHeight : mDefaultHeight);
1478 return NO_ERROR;
1479 case NATIVE_WINDOW_TRANSFORM_HINT:
1480 *value = static_cast<int>(getTransformHint());
1481 return NO_ERROR;
1482 case NATIVE_WINDOW_CONSUMER_RUNNING_BEHIND: {
1483 status_t err = NO_ERROR;
1484 if (!mConsumerRunningBehind) {
1485 *value = 0;
1486 } else {
1487 err = mGraphicBufferProducer->query(what, value);
1488 if (err == NO_ERROR) {
1489 mConsumerRunningBehind = *value;
1490 }
1491 }
1492 return err;
1493 }
1494 case NATIVE_WINDOW_BUFFER_AGE: {
1495 if (mBufferAge > INT32_MAX) {
1496 *value = 0;
1497 } else {
1498 *value = static_cast<int32_t>(mBufferAge);
1499 }
1500 return NO_ERROR;
1501 }
1502 case NATIVE_WINDOW_LAST_DEQUEUE_DURATION: {
1503 int64_t durationUs = mLastDequeueDuration / 1000;
1504 *value = durationUs > std::numeric_limits<int>::max() ?
1505 std::numeric_limits<int>::max() :
1506 static_cast<int>(durationUs);
1507 return NO_ERROR;
1508 }
1509 case NATIVE_WINDOW_LAST_QUEUE_DURATION: {
1510 int64_t durationUs = mLastQueueDuration / 1000;
1511 *value = durationUs > std::numeric_limits<int>::max() ?
1512 std::numeric_limits<int>::max() :
1513 static_cast<int>(durationUs);
1514 return NO_ERROR;
1515 }
1516 case NATIVE_WINDOW_FRAME_TIMESTAMPS_SUPPORTS_PRESENT: {
1517 querySupportedTimestampsLocked();
1518 *value = mFrameTimestampsSupportsPresent ? 1 : 0;
1519 return NO_ERROR;
1520 }
1521 case NATIVE_WINDOW_IS_VALID: {
1522 *value = mGraphicBufferProducer != nullptr ? 1 : 0;
1523 return NO_ERROR;
1524 }
1525 case NATIVE_WINDOW_DATASPACE: {
1526 *value = static_cast<int>(mDataSpace);
1527 return NO_ERROR;
1528 }
1529 case NATIVE_WINDOW_MAX_BUFFER_COUNT: {
1530 *value = mMaxBufferCount;
1531 return NO_ERROR;
1532 }
1533 }
1534 }
1535 return mGraphicBufferProducer->query(what, value);
1536 }
1537
perform(int operation,va_list args)1538 int Surface::perform(int operation, va_list args)
1539 {
1540 int res = NO_ERROR;
1541 switch (operation) {
1542 case NATIVE_WINDOW_CONNECT:
1543 // deprecated. must return NO_ERROR.
1544 break;
1545 case NATIVE_WINDOW_DISCONNECT:
1546 // deprecated. must return NO_ERROR.
1547 break;
1548 case NATIVE_WINDOW_SET_USAGE:
1549 res = dispatchSetUsage(args);
1550 break;
1551 case NATIVE_WINDOW_SET_CROP:
1552 res = dispatchSetCrop(args);
1553 break;
1554 case NATIVE_WINDOW_SET_BUFFER_COUNT:
1555 res = dispatchSetBufferCount(args);
1556 break;
1557 case NATIVE_WINDOW_SET_BUFFERS_GEOMETRY:
1558 res = dispatchSetBuffersGeometry(args);
1559 break;
1560 case NATIVE_WINDOW_SET_BUFFERS_TRANSFORM:
1561 res = dispatchSetBuffersTransform(args);
1562 break;
1563 case NATIVE_WINDOW_SET_BUFFERS_STICKY_TRANSFORM:
1564 res = dispatchSetBuffersStickyTransform(args);
1565 break;
1566 case NATIVE_WINDOW_SET_BUFFERS_TIMESTAMP:
1567 res = dispatchSetBuffersTimestamp(args);
1568 break;
1569 case NATIVE_WINDOW_SET_BUFFERS_DIMENSIONS:
1570 res = dispatchSetBuffersDimensions(args);
1571 break;
1572 case NATIVE_WINDOW_SET_BUFFERS_USER_DIMENSIONS:
1573 res = dispatchSetBuffersUserDimensions(args);
1574 break;
1575 case NATIVE_WINDOW_SET_BUFFERS_FORMAT:
1576 res = dispatchSetBuffersFormat(args);
1577 break;
1578 case NATIVE_WINDOW_LOCK:
1579 res = dispatchLock(args);
1580 break;
1581 case NATIVE_WINDOW_UNLOCK_AND_POST:
1582 res = dispatchUnlockAndPost(args);
1583 break;
1584 case NATIVE_WINDOW_SET_SCALING_MODE:
1585 res = dispatchSetScalingMode(args);
1586 break;
1587 case NATIVE_WINDOW_API_CONNECT:
1588 res = dispatchConnect(args);
1589 break;
1590 case NATIVE_WINDOW_API_DISCONNECT:
1591 res = dispatchDisconnect(args);
1592 break;
1593 case NATIVE_WINDOW_SET_SIDEBAND_STREAM:
1594 res = dispatchSetSidebandStream(args);
1595 break;
1596 case NATIVE_WINDOW_SET_BUFFERS_DATASPACE:
1597 res = dispatchSetBuffersDataSpace(args);
1598 break;
1599 case NATIVE_WINDOW_SET_BUFFERS_SMPTE2086_METADATA:
1600 res = dispatchSetBuffersSmpte2086Metadata(args);
1601 break;
1602 case NATIVE_WINDOW_SET_BUFFERS_CTA861_3_METADATA:
1603 res = dispatchSetBuffersCta8613Metadata(args);
1604 break;
1605 case NATIVE_WINDOW_SET_BUFFERS_HDR10_PLUS_METADATA:
1606 res = dispatchSetBuffersHdr10PlusMetadata(args);
1607 break;
1608 case NATIVE_WINDOW_SET_SURFACE_DAMAGE:
1609 res = dispatchSetSurfaceDamage(args);
1610 break;
1611 case NATIVE_WINDOW_SET_SHARED_BUFFER_MODE:
1612 res = dispatchSetSharedBufferMode(args);
1613 break;
1614 case NATIVE_WINDOW_SET_AUTO_REFRESH:
1615 res = dispatchSetAutoRefresh(args);
1616 break;
1617 case NATIVE_WINDOW_GET_REFRESH_CYCLE_DURATION:
1618 res = dispatchGetDisplayRefreshCycleDuration(args);
1619 break;
1620 case NATIVE_WINDOW_GET_NEXT_FRAME_ID:
1621 res = dispatchGetNextFrameId(args);
1622 break;
1623 case NATIVE_WINDOW_ENABLE_FRAME_TIMESTAMPS:
1624 res = dispatchEnableFrameTimestamps(args);
1625 break;
1626 case NATIVE_WINDOW_GET_COMPOSITOR_TIMING:
1627 res = dispatchGetCompositorTiming(args);
1628 break;
1629 case NATIVE_WINDOW_GET_FRAME_TIMESTAMPS:
1630 res = dispatchGetFrameTimestamps(args);
1631 break;
1632 case NATIVE_WINDOW_GET_WIDE_COLOR_SUPPORT:
1633 res = dispatchGetWideColorSupport(args);
1634 break;
1635 case NATIVE_WINDOW_GET_HDR_SUPPORT:
1636 res = dispatchGetHdrSupport(args);
1637 break;
1638 case NATIVE_WINDOW_SET_USAGE64:
1639 res = dispatchSetUsage64(args);
1640 break;
1641 case NATIVE_WINDOW_GET_CONSUMER_USAGE64:
1642 res = dispatchGetConsumerUsage64(args);
1643 break;
1644 case NATIVE_WINDOW_SET_AUTO_PREROTATION:
1645 res = dispatchSetAutoPrerotation(args);
1646 break;
1647 case NATIVE_WINDOW_GET_LAST_DEQUEUE_START:
1648 res = dispatchGetLastDequeueStartTime(args);
1649 break;
1650 case NATIVE_WINDOW_SET_DEQUEUE_TIMEOUT:
1651 res = dispatchSetDequeueTimeout(args);
1652 break;
1653 case NATIVE_WINDOW_GET_LAST_DEQUEUE_DURATION:
1654 res = dispatchGetLastDequeueDuration(args);
1655 break;
1656 case NATIVE_WINDOW_GET_LAST_QUEUE_DURATION:
1657 res = dispatchGetLastQueueDuration(args);
1658 break;
1659 case NATIVE_WINDOW_SET_FRAME_RATE:
1660 res = dispatchSetFrameRate(args);
1661 break;
1662 case NATIVE_WINDOW_SET_CANCEL_INTERCEPTOR:
1663 res = dispatchAddCancelInterceptor(args);
1664 break;
1665 case NATIVE_WINDOW_SET_DEQUEUE_INTERCEPTOR:
1666 res = dispatchAddDequeueInterceptor(args);
1667 break;
1668 case NATIVE_WINDOW_SET_PERFORM_INTERCEPTOR:
1669 res = dispatchAddPerformInterceptor(args);
1670 break;
1671 case NATIVE_WINDOW_SET_QUEUE_INTERCEPTOR:
1672 res = dispatchAddQueueInterceptor(args);
1673 break;
1674 case NATIVE_WINDOW_SET_QUERY_INTERCEPTOR:
1675 res = dispatchAddQueryInterceptor(args);
1676 break;
1677 case NATIVE_WINDOW_ALLOCATE_BUFFERS:
1678 allocateBuffers();
1679 res = NO_ERROR;
1680 break;
1681 case NATIVE_WINDOW_GET_LAST_QUEUED_BUFFER:
1682 res = dispatchGetLastQueuedBuffer(args);
1683 break;
1684 case NATIVE_WINDOW_GET_LAST_QUEUED_BUFFER2:
1685 res = dispatchGetLastQueuedBuffer2(args);
1686 break;
1687 case NATIVE_WINDOW_SET_FRAME_TIMELINE_INFO:
1688 res = dispatchSetFrameTimelineInfo(args);
1689 break;
1690 case NATIVE_WINDOW_SET_BUFFERS_ADDITIONAL_OPTIONS:
1691 res = dispatchSetAdditionalOptions(args);
1692 break;
1693 default:
1694 res = NAME_NOT_FOUND;
1695 break;
1696 }
1697 return res;
1698 }
1699
dispatchConnect(va_list args)1700 int Surface::dispatchConnect(va_list args) {
1701 int api = va_arg(args, int);
1702 return connect(api);
1703 }
1704
dispatchDisconnect(va_list args)1705 int Surface::dispatchDisconnect(va_list args) {
1706 int api = va_arg(args, int);
1707 return disconnect(api);
1708 }
1709
dispatchSetUsage(va_list args)1710 int Surface::dispatchSetUsage(va_list args) {
1711 uint64_t usage = va_arg(args, uint32_t);
1712 return setUsage(usage);
1713 }
1714
dispatchSetUsage64(va_list args)1715 int Surface::dispatchSetUsage64(va_list args) {
1716 uint64_t usage = va_arg(args, uint64_t);
1717 return setUsage(usage);
1718 }
1719
dispatchSetCrop(va_list args)1720 int Surface::dispatchSetCrop(va_list args) {
1721 android_native_rect_t const* rect = va_arg(args, android_native_rect_t*);
1722 return setCrop(reinterpret_cast<Rect const*>(rect));
1723 }
1724
dispatchSetBufferCount(va_list args)1725 int Surface::dispatchSetBufferCount(va_list args) {
1726 size_t bufferCount = va_arg(args, size_t);
1727 return setBufferCount(static_cast<int32_t>(bufferCount));
1728 }
1729
dispatchSetBuffersGeometry(va_list args)1730 int Surface::dispatchSetBuffersGeometry(va_list args) {
1731 uint32_t width = va_arg(args, uint32_t);
1732 uint32_t height = va_arg(args, uint32_t);
1733 PixelFormat format = va_arg(args, PixelFormat);
1734 int err = setBuffersDimensions(width, height);
1735 if (err != 0) {
1736 return err;
1737 }
1738 return setBuffersFormat(format);
1739 }
1740
dispatchSetBuffersDimensions(va_list args)1741 int Surface::dispatchSetBuffersDimensions(va_list args) {
1742 uint32_t width = va_arg(args, uint32_t);
1743 uint32_t height = va_arg(args, uint32_t);
1744 return setBuffersDimensions(width, height);
1745 }
1746
dispatchSetBuffersUserDimensions(va_list args)1747 int Surface::dispatchSetBuffersUserDimensions(va_list args) {
1748 uint32_t width = va_arg(args, uint32_t);
1749 uint32_t height = va_arg(args, uint32_t);
1750 return setBuffersUserDimensions(width, height);
1751 }
1752
dispatchSetBuffersFormat(va_list args)1753 int Surface::dispatchSetBuffersFormat(va_list args) {
1754 PixelFormat format = va_arg(args, PixelFormat);
1755 return setBuffersFormat(format);
1756 }
1757
dispatchSetScalingMode(va_list args)1758 int Surface::dispatchSetScalingMode(va_list args) {
1759 int mode = va_arg(args, int);
1760 return setScalingMode(mode);
1761 }
1762
dispatchSetBuffersTransform(va_list args)1763 int Surface::dispatchSetBuffersTransform(va_list args) {
1764 uint32_t transform = va_arg(args, uint32_t);
1765 return setBuffersTransform(transform);
1766 }
1767
dispatchSetBuffersStickyTransform(va_list args)1768 int Surface::dispatchSetBuffersStickyTransform(va_list args) {
1769 uint32_t transform = va_arg(args, uint32_t);
1770 return setBuffersStickyTransform(transform);
1771 }
1772
dispatchSetBuffersTimestamp(va_list args)1773 int Surface::dispatchSetBuffersTimestamp(va_list args) {
1774 int64_t timestamp = va_arg(args, int64_t);
1775 return setBuffersTimestamp(timestamp);
1776 }
1777
dispatchLock(va_list args)1778 int Surface::dispatchLock(va_list args) {
1779 ANativeWindow_Buffer* outBuffer = va_arg(args, ANativeWindow_Buffer*);
1780 ARect* inOutDirtyBounds = va_arg(args, ARect*);
1781 return lock(outBuffer, inOutDirtyBounds);
1782 }
1783
dispatchUnlockAndPost(va_list args)1784 int Surface::dispatchUnlockAndPost(va_list args __attribute__((unused))) {
1785 return unlockAndPost();
1786 }
1787
dispatchSetSidebandStream(va_list args)1788 int Surface::dispatchSetSidebandStream(va_list args) {
1789 native_handle_t* sH = va_arg(args, native_handle_t*);
1790 sp<NativeHandle> sidebandHandle = NativeHandle::create(sH, false);
1791 setSidebandStream(sidebandHandle);
1792 return OK;
1793 }
1794
dispatchSetBuffersDataSpace(va_list args)1795 int Surface::dispatchSetBuffersDataSpace(va_list args) {
1796 Dataspace dataspace = static_cast<Dataspace>(va_arg(args, int));
1797 return setBuffersDataSpace(dataspace);
1798 }
1799
dispatchSetBuffersSmpte2086Metadata(va_list args)1800 int Surface::dispatchSetBuffersSmpte2086Metadata(va_list args) {
1801 const android_smpte2086_metadata* metadata =
1802 va_arg(args, const android_smpte2086_metadata*);
1803 return setBuffersSmpte2086Metadata(metadata);
1804 }
1805
dispatchSetBuffersCta8613Metadata(va_list args)1806 int Surface::dispatchSetBuffersCta8613Metadata(va_list args) {
1807 const android_cta861_3_metadata* metadata =
1808 va_arg(args, const android_cta861_3_metadata*);
1809 return setBuffersCta8613Metadata(metadata);
1810 }
1811
dispatchSetBuffersHdr10PlusMetadata(va_list args)1812 int Surface::dispatchSetBuffersHdr10PlusMetadata(va_list args) {
1813 const size_t size = va_arg(args, size_t);
1814 const uint8_t* metadata = va_arg(args, const uint8_t*);
1815 return setBuffersHdr10PlusMetadata(size, metadata);
1816 }
1817
dispatchSetSurfaceDamage(va_list args)1818 int Surface::dispatchSetSurfaceDamage(va_list args) {
1819 android_native_rect_t* rects = va_arg(args, android_native_rect_t*);
1820 size_t numRects = va_arg(args, size_t);
1821 setSurfaceDamage(rects, numRects);
1822 return NO_ERROR;
1823 }
1824
dispatchSetSharedBufferMode(va_list args)1825 int Surface::dispatchSetSharedBufferMode(va_list args) {
1826 bool sharedBufferMode = va_arg(args, int);
1827 return setSharedBufferMode(sharedBufferMode);
1828 }
1829
dispatchSetAutoRefresh(va_list args)1830 int Surface::dispatchSetAutoRefresh(va_list args) {
1831 bool autoRefresh = va_arg(args, int);
1832 return setAutoRefresh(autoRefresh);
1833 }
1834
dispatchGetDisplayRefreshCycleDuration(va_list args)1835 int Surface::dispatchGetDisplayRefreshCycleDuration(va_list args) {
1836 nsecs_t* outRefreshDuration = va_arg(args, int64_t*);
1837 return getDisplayRefreshCycleDuration(outRefreshDuration);
1838 }
1839
dispatchGetNextFrameId(va_list args)1840 int Surface::dispatchGetNextFrameId(va_list args) {
1841 uint64_t* nextFrameId = va_arg(args, uint64_t*);
1842 *nextFrameId = getNextFrameNumber();
1843 return NO_ERROR;
1844 }
1845
dispatchEnableFrameTimestamps(va_list args)1846 int Surface::dispatchEnableFrameTimestamps(va_list args) {
1847 bool enable = va_arg(args, int);
1848 enableFrameTimestamps(enable);
1849 return NO_ERROR;
1850 }
1851
dispatchGetCompositorTiming(va_list args)1852 int Surface::dispatchGetCompositorTiming(va_list args) {
1853 nsecs_t* compositeDeadline = va_arg(args, int64_t*);
1854 nsecs_t* compositeInterval = va_arg(args, int64_t*);
1855 nsecs_t* compositeToPresentLatency = va_arg(args, int64_t*);
1856 return getCompositorTiming(compositeDeadline, compositeInterval,
1857 compositeToPresentLatency);
1858 }
1859
dispatchGetFrameTimestamps(va_list args)1860 int Surface::dispatchGetFrameTimestamps(va_list args) {
1861 uint64_t frameId = va_arg(args, uint64_t);
1862 nsecs_t* outRequestedPresentTime = va_arg(args, int64_t*);
1863 nsecs_t* outAcquireTime = va_arg(args, int64_t*);
1864 nsecs_t* outLatchTime = va_arg(args, int64_t*);
1865 nsecs_t* outFirstRefreshStartTime = va_arg(args, int64_t*);
1866 nsecs_t* outLastRefreshStartTime = va_arg(args, int64_t*);
1867 nsecs_t* outGpuCompositionDoneTime = va_arg(args, int64_t*);
1868 nsecs_t* outDisplayPresentTime = va_arg(args, int64_t*);
1869 nsecs_t* outDequeueReadyTime = va_arg(args, int64_t*);
1870 nsecs_t* outReleaseTime = va_arg(args, int64_t*);
1871 return getFrameTimestamps(frameId,
1872 outRequestedPresentTime, outAcquireTime, outLatchTime,
1873 outFirstRefreshStartTime, outLastRefreshStartTime,
1874 outGpuCompositionDoneTime, outDisplayPresentTime,
1875 outDequeueReadyTime, outReleaseTime);
1876 }
1877
dispatchGetWideColorSupport(va_list args)1878 int Surface::dispatchGetWideColorSupport(va_list args) {
1879 bool* outSupport = va_arg(args, bool*);
1880 return getWideColorSupport(outSupport);
1881 }
1882
dispatchGetHdrSupport(va_list args)1883 int Surface::dispatchGetHdrSupport(va_list args) {
1884 bool* outSupport = va_arg(args, bool*);
1885 return getHdrSupport(outSupport);
1886 }
1887
dispatchGetConsumerUsage64(va_list args)1888 int Surface::dispatchGetConsumerUsage64(va_list args) {
1889 uint64_t* usage = va_arg(args, uint64_t*);
1890 return getConsumerUsage(usage);
1891 }
1892
dispatchSetAutoPrerotation(va_list args)1893 int Surface::dispatchSetAutoPrerotation(va_list args) {
1894 bool autoPrerotation = va_arg(args, int);
1895 return setAutoPrerotation(autoPrerotation);
1896 }
1897
dispatchGetLastDequeueStartTime(va_list args)1898 int Surface::dispatchGetLastDequeueStartTime(va_list args) {
1899 int64_t* lastDequeueStartTime = va_arg(args, int64_t*);
1900 *lastDequeueStartTime = mLastDequeueStartTime;
1901 return NO_ERROR;
1902 }
1903
dispatchSetDequeueTimeout(va_list args)1904 int Surface::dispatchSetDequeueTimeout(va_list args) {
1905 nsecs_t timeout = va_arg(args, int64_t);
1906 return setDequeueTimeout(timeout);
1907 }
1908
dispatchGetLastDequeueDuration(va_list args)1909 int Surface::dispatchGetLastDequeueDuration(va_list args) {
1910 int64_t* lastDequeueDuration = va_arg(args, int64_t*);
1911 *lastDequeueDuration = mLastDequeueDuration;
1912 return NO_ERROR;
1913 }
1914
dispatchGetLastQueueDuration(va_list args)1915 int Surface::dispatchGetLastQueueDuration(va_list args) {
1916 int64_t* lastQueueDuration = va_arg(args, int64_t*);
1917 *lastQueueDuration = mLastQueueDuration;
1918 return NO_ERROR;
1919 }
1920
dispatchSetFrameRate(va_list args)1921 int Surface::dispatchSetFrameRate(va_list args) {
1922 float frameRate = static_cast<float>(va_arg(args, double));
1923 int8_t compatibility = static_cast<int8_t>(va_arg(args, int));
1924 int8_t changeFrameRateStrategy = static_cast<int8_t>(va_arg(args, int));
1925 return setFrameRate(frameRate, compatibility, changeFrameRateStrategy);
1926 }
1927
dispatchAddCancelInterceptor(va_list args)1928 int Surface::dispatchAddCancelInterceptor(va_list args) {
1929 ANativeWindow_cancelBufferInterceptor interceptor =
1930 va_arg(args, ANativeWindow_cancelBufferInterceptor);
1931 void* data = va_arg(args, void*);
1932 std::lock_guard<std::shared_mutex> lock(mInterceptorMutex);
1933 mCancelInterceptor = interceptor;
1934 mCancelInterceptorData = data;
1935 return NO_ERROR;
1936 }
1937
dispatchAddDequeueInterceptor(va_list args)1938 int Surface::dispatchAddDequeueInterceptor(va_list args) {
1939 ANativeWindow_dequeueBufferInterceptor interceptor =
1940 va_arg(args, ANativeWindow_dequeueBufferInterceptor);
1941 void* data = va_arg(args, void*);
1942 std::lock_guard<std::shared_mutex> lock(mInterceptorMutex);
1943 mDequeueInterceptor = interceptor;
1944 mDequeueInterceptorData = data;
1945 return NO_ERROR;
1946 }
1947
dispatchAddPerformInterceptor(va_list args)1948 int Surface::dispatchAddPerformInterceptor(va_list args) {
1949 ANativeWindow_performInterceptor interceptor = va_arg(args, ANativeWindow_performInterceptor);
1950 void* data = va_arg(args, void*);
1951 std::lock_guard<std::shared_mutex> lock(mInterceptorMutex);
1952 mPerformInterceptor = interceptor;
1953 mPerformInterceptorData = data;
1954 return NO_ERROR;
1955 }
1956
dispatchAddQueueInterceptor(va_list args)1957 int Surface::dispatchAddQueueInterceptor(va_list args) {
1958 ANativeWindow_queueBufferInterceptor interceptor =
1959 va_arg(args, ANativeWindow_queueBufferInterceptor);
1960 void* data = va_arg(args, void*);
1961 std::lock_guard<std::shared_mutex> lock(mInterceptorMutex);
1962 mQueueInterceptor = interceptor;
1963 mQueueInterceptorData = data;
1964 return NO_ERROR;
1965 }
1966
dispatchAddQueryInterceptor(va_list args)1967 int Surface::dispatchAddQueryInterceptor(va_list args) {
1968 ANativeWindow_queryInterceptor interceptor = va_arg(args, ANativeWindow_queryInterceptor);
1969 void* data = va_arg(args, void*);
1970 std::lock_guard<std::shared_mutex> lock(mInterceptorMutex);
1971 mQueryInterceptor = interceptor;
1972 mQueryInterceptorData = data;
1973 return NO_ERROR;
1974 }
1975
dispatchGetLastQueuedBuffer(va_list args)1976 int Surface::dispatchGetLastQueuedBuffer(va_list args) {
1977 AHardwareBuffer** buffer = va_arg(args, AHardwareBuffer**);
1978 int* fence = va_arg(args, int*);
1979 float* matrix = va_arg(args, float*);
1980 sp<GraphicBuffer> graphicBuffer;
1981 sp<Fence> spFence;
1982
1983 int result = mGraphicBufferProducer->getLastQueuedBuffer(&graphicBuffer, &spFence, matrix);
1984
1985 if (graphicBuffer != nullptr) {
1986 *buffer = graphicBuffer->toAHardwareBuffer();
1987 AHardwareBuffer_acquire(*buffer);
1988 } else {
1989 *buffer = nullptr;
1990 }
1991
1992 if (spFence != nullptr) {
1993 *fence = spFence->dup();
1994 } else {
1995 *fence = -1;
1996 }
1997 return result;
1998 }
1999
dispatchGetLastQueuedBuffer2(va_list args)2000 int Surface::dispatchGetLastQueuedBuffer2(va_list args) {
2001 AHardwareBuffer** buffer = va_arg(args, AHardwareBuffer**);
2002 int* fence = va_arg(args, int*);
2003 ARect* crop = va_arg(args, ARect*);
2004 uint32_t* transform = va_arg(args, uint32_t*);
2005 sp<GraphicBuffer> graphicBuffer;
2006 sp<Fence> spFence;
2007
2008 Rect r;
2009 int result =
2010 mGraphicBufferProducer->getLastQueuedBuffer(&graphicBuffer, &spFence, &r, transform);
2011
2012 if (graphicBuffer != nullptr) {
2013 *buffer = graphicBuffer->toAHardwareBuffer();
2014 AHardwareBuffer_acquire(*buffer);
2015
2016 // Avoid setting crop* unless buffer is valid (matches IGBP behavior)
2017 crop->left = r.left;
2018 crop->top = r.top;
2019 crop->right = r.right;
2020 crop->bottom = r.bottom;
2021 } else {
2022 *buffer = nullptr;
2023 }
2024
2025 if (spFence != nullptr) {
2026 *fence = spFence->dup();
2027 } else {
2028 *fence = -1;
2029 }
2030 return result;
2031 }
2032
dispatchSetFrameTimelineInfo(va_list args)2033 int Surface::dispatchSetFrameTimelineInfo(va_list args) {
2034 ATRACE_CALL();
2035 ALOGV("Surface::%s", __func__);
2036
2037 const auto nativeWindowFtlInfo = static_cast<ANativeWindowFrameTimelineInfo>(
2038 va_arg(args, ANativeWindowFrameTimelineInfo));
2039
2040 FrameTimelineInfo ftlInfo;
2041 ftlInfo.vsyncId = nativeWindowFtlInfo.frameTimelineVsyncId;
2042 ftlInfo.inputEventId = nativeWindowFtlInfo.inputEventId;
2043 ftlInfo.startTimeNanos = nativeWindowFtlInfo.startTimeNanos;
2044 ftlInfo.useForRefreshRateSelection = nativeWindowFtlInfo.useForRefreshRateSelection;
2045 ftlInfo.skippedFrameVsyncId = nativeWindowFtlInfo.skippedFrameVsyncId;
2046 ftlInfo.skippedFrameStartTimeNanos = nativeWindowFtlInfo.skippedFrameStartTimeNanos;
2047
2048 return setFrameTimelineInfo(nativeWindowFtlInfo.frameNumber, ftlInfo);
2049 }
2050
dispatchSetAdditionalOptions(va_list args)2051 int Surface::dispatchSetAdditionalOptions(va_list args) {
2052 ATRACE_CALL();
2053
2054 #if COM_ANDROID_GRAPHICS_LIBGUI_FLAGS(BQ_EXTENDEDALLOCATE)
2055 const AHardwareBufferLongOptions* opts = va_arg(args, const AHardwareBufferLongOptions*);
2056 const size_t optsSize = va_arg(args, size_t);
2057 std::vector<gui::AdditionalOptions> convertedOpts;
2058 convertedOpts.reserve(optsSize);
2059 for (size_t i = 0; i < optsSize; i++) {
2060 convertedOpts.emplace_back(opts[i].name, opts[i].value);
2061 }
2062 return setAdditionalOptions(convertedOpts);
2063 #else
2064 (void)args;
2065 return INVALID_OPERATION;
2066 #endif
2067 }
2068
transformToDisplayInverse() const2069 bool Surface::transformToDisplayInverse() const {
2070 return (mTransform & NATIVE_WINDOW_TRANSFORM_INVERSE_DISPLAY) ==
2071 NATIVE_WINDOW_TRANSFORM_INVERSE_DISPLAY;
2072 }
2073
connect(int api)2074 int Surface::connect(int api) {
2075 static sp<SurfaceListener> listener = new StubSurfaceListener();
2076 return connect(api, listener);
2077 }
2078
connect(int api,const sp<SurfaceListener> & listener,bool reportBufferRemoval)2079 int Surface::connect(int api, const sp<SurfaceListener>& listener, bool reportBufferRemoval) {
2080 ATRACE_CALL();
2081 ALOGV("Surface::connect");
2082 Mutex::Autolock lock(mMutex);
2083 IGraphicBufferProducer::QueueBufferOutput output;
2084 mReportRemovedBuffers = reportBufferRemoval;
2085
2086 if (listener != nullptr) {
2087 mListenerProxy = new ProducerListenerProxy(this, listener);
2088 }
2089
2090 int err =
2091 mGraphicBufferProducer->connect(mListenerProxy, api, mProducerControlledByApp, &output);
2092 if (err == NO_ERROR) {
2093 mDefaultWidth = output.width;
2094 mDefaultHeight = output.height;
2095 mNextFrameNumber = output.nextFrameNumber;
2096 mMaxBufferCount = output.maxBufferCount;
2097
2098 // Ignore transform hint if sticky transform is set or transform to display inverse flag is
2099 // set. Transform hint should be ignored if the client is expected to always submit buffers
2100 // in the same orientation.
2101 if (mStickyTransform == 0 && !transformToDisplayInverse()) {
2102 mTransformHint = output.transformHint;
2103 }
2104
2105 mConsumerRunningBehind = (output.numPendingBuffers >= 2);
2106
2107 #if COM_ANDROID_GRAPHICS_LIBGUI_FLAGS(WB_PLATFORM_API_IMPROVEMENTS)
2108 if (listener && listener->needsDeathNotify()) {
2109 mSurfaceDeathListener = sp<ProducerDeathListenerProxy>::make(listener);
2110 IInterface::asBinder(mGraphicBufferProducer)->linkToDeath(mSurfaceDeathListener);
2111 }
2112 #endif // COM_ANDROID_GRAPHICS_LIBGUI_FLAGS(WB_PLATFORM_API_IMPROVEMENTS)
2113 }
2114 if (!err && api == NATIVE_WINDOW_API_CPU) {
2115 mConnectedToCpu = true;
2116 // Clear the dirty region in case we're switching from a non-CPU API
2117 mDirtyRegion.clear();
2118 } else if (!err) {
2119 // Initialize the dirty region for tracking surface damage
2120 mDirtyRegion = Region::INVALID_REGION;
2121 }
2122
2123 return err;
2124 }
2125
disconnect(int api,IGraphicBufferProducer::DisconnectMode mode)2126 int Surface::disconnect(int api, IGraphicBufferProducer::DisconnectMode mode) {
2127 ATRACE_CALL();
2128 ALOGV("Surface::disconnect");
2129 Mutex::Autolock lock(mMutex);
2130 mRemovedBuffers.clear();
2131 mSharedBufferSlot = BufferItem::INVALID_BUFFER_SLOT;
2132 mSharedBufferHasBeenQueued = false;
2133 freeAllBuffers();
2134 int err = mGraphicBufferProducer->disconnect(api, mode);
2135 if (!err) {
2136 mReqFormat = 0;
2137 mReqWidth = 0;
2138 mReqHeight = 0;
2139 mReqUsage = 0;
2140 mCrop.clear();
2141 mDataSpace = Dataspace::UNKNOWN;
2142 mScalingMode = NATIVE_WINDOW_SCALING_MODE_FREEZE;
2143 mTransform = 0;
2144 mStickyTransform = 0;
2145 mAutoPrerotation = false;
2146 mEnableFrameTimestamps = false;
2147 mMaxBufferCount = NUM_BUFFER_SLOTS;
2148
2149 if (api == NATIVE_WINDOW_API_CPU) {
2150 mConnectedToCpu = false;
2151 }
2152 }
2153
2154 #if COM_ANDROID_GRAPHICS_LIBGUI_FLAGS(WB_PLATFORM_API_IMPROVEMENTS)
2155 if (mSurfaceDeathListener != nullptr) {
2156 IInterface::asBinder(mGraphicBufferProducer)->unlinkToDeath(mSurfaceDeathListener);
2157 mSurfaceDeathListener = nullptr;
2158 }
2159 #endif // COM_ANDROID_GRAPHICS_LIBGUI_FLAGS(WB_PLATFORM_API_IMPROVEMENTS)
2160
2161 return err;
2162 }
2163
detachNextBuffer(sp<GraphicBuffer> * outBuffer,sp<Fence> * outFence)2164 int Surface::detachNextBuffer(sp<GraphicBuffer>* outBuffer,
2165 sp<Fence>* outFence) {
2166 ATRACE_CALL();
2167 ALOGV("Surface::detachNextBuffer");
2168
2169 if (outBuffer == nullptr || outFence == nullptr) {
2170 return BAD_VALUE;
2171 }
2172
2173 Mutex::Autolock lock(mMutex);
2174 if (mReportRemovedBuffers) {
2175 mRemovedBuffers.clear();
2176 }
2177
2178 sp<GraphicBuffer> buffer(nullptr);
2179 sp<Fence> fence(nullptr);
2180 status_t result = mGraphicBufferProducer->detachNextBuffer(
2181 &buffer, &fence);
2182 if (result != NO_ERROR) {
2183 return result;
2184 }
2185
2186 *outBuffer = buffer;
2187 if (fence != nullptr && fence->isValid()) {
2188 *outFence = fence;
2189 } else {
2190 *outFence = Fence::NO_FENCE;
2191 }
2192
2193 for (int i = 0; i < NUM_BUFFER_SLOTS; i++) {
2194 if (mSlots[i].buffer != nullptr &&
2195 mSlots[i].buffer->getId() == buffer->getId()) {
2196 if (mReportRemovedBuffers) {
2197 mRemovedBuffers.push_back(mSlots[i].buffer);
2198 }
2199 mSlots[i].buffer = nullptr;
2200 }
2201 }
2202
2203 return NO_ERROR;
2204 }
2205
attachBuffer(ANativeWindowBuffer * buffer)2206 int Surface::attachBuffer(ANativeWindowBuffer* buffer)
2207 {
2208 ATRACE_CALL();
2209 ALOGV("Surface::attachBuffer");
2210
2211 Mutex::Autolock lock(mMutex);
2212 if (mReportRemovedBuffers) {
2213 mRemovedBuffers.clear();
2214 }
2215
2216 sp<GraphicBuffer> graphicBuffer(static_cast<GraphicBuffer*>(buffer));
2217 uint32_t priorGeneration = graphicBuffer->mGenerationNumber;
2218 graphicBuffer->mGenerationNumber = mGenerationNumber;
2219 int32_t attachedSlot = -1;
2220 status_t result = mGraphicBufferProducer->attachBuffer(&attachedSlot, graphicBuffer);
2221 if (result != NO_ERROR) {
2222 ALOGE("attachBuffer: IGraphicBufferProducer call failed (%d)", result);
2223 graphicBuffer->mGenerationNumber = priorGeneration;
2224 return result;
2225 }
2226 if (mReportRemovedBuffers && (mSlots[attachedSlot].buffer != nullptr)) {
2227 mRemovedBuffers.push_back(mSlots[attachedSlot].buffer);
2228 }
2229 mSlots[attachedSlot].buffer = graphicBuffer;
2230 mDequeuedSlots.insert(attachedSlot);
2231
2232 return NO_ERROR;
2233 }
2234
setUsage(uint64_t reqUsage)2235 int Surface::setUsage(uint64_t reqUsage)
2236 {
2237 ALOGV("Surface::setUsage");
2238 Mutex::Autolock lock(mMutex);
2239 if (reqUsage != mReqUsage) {
2240 mSharedBufferSlot = BufferItem::INVALID_BUFFER_SLOT;
2241 }
2242 mReqUsage = reqUsage;
2243 return OK;
2244 }
2245
setCrop(Rect const * rect)2246 int Surface::setCrop(Rect const* rect)
2247 {
2248 ATRACE_CALL();
2249
2250 Rect realRect(Rect::EMPTY_RECT);
2251 if (rect == nullptr || rect->isEmpty()) {
2252 realRect.clear();
2253 } else {
2254 realRect = *rect;
2255 }
2256
2257 ALOGV("Surface::setCrop rect=[%d %d %d %d]",
2258 realRect.left, realRect.top, realRect.right, realRect.bottom);
2259
2260 Mutex::Autolock lock(mMutex);
2261 mCrop = realRect;
2262 return NO_ERROR;
2263 }
2264
setBufferCount(int bufferCount)2265 int Surface::setBufferCount(int bufferCount)
2266 {
2267 ATRACE_CALL();
2268 ALOGV("Surface::setBufferCount");
2269 Mutex::Autolock lock(mMutex);
2270
2271 status_t err = NO_ERROR;
2272 if (bufferCount == 0) {
2273 err = mGraphicBufferProducer->setMaxDequeuedBufferCount(1);
2274 } else {
2275 int minUndequeuedBuffers = 0;
2276 err = mGraphicBufferProducer->query(
2277 NATIVE_WINDOW_MIN_UNDEQUEUED_BUFFERS, &minUndequeuedBuffers);
2278 if (err == NO_ERROR) {
2279 err = mGraphicBufferProducer->setMaxDequeuedBufferCount(
2280 bufferCount - minUndequeuedBuffers);
2281 }
2282 }
2283
2284 ALOGE_IF(err, "IGraphicBufferProducer::setBufferCount(%d) returned %s",
2285 bufferCount, strerror(-err));
2286
2287 return err;
2288 }
2289
setMaxDequeuedBufferCount(int maxDequeuedBuffers)2290 int Surface::setMaxDequeuedBufferCount(int maxDequeuedBuffers) {
2291 ATRACE_CALL();
2292 ALOGV("Surface::setMaxDequeuedBufferCount");
2293 Mutex::Autolock lock(mMutex);
2294
2295 status_t err = mGraphicBufferProducer->setMaxDequeuedBufferCount(
2296 maxDequeuedBuffers);
2297 ALOGE_IF(err, "IGraphicBufferProducer::setMaxDequeuedBufferCount(%d) "
2298 "returned %s", maxDequeuedBuffers, strerror(-err));
2299
2300 return err;
2301 }
2302
setAsyncMode(bool async)2303 int Surface::setAsyncMode(bool async) {
2304 ATRACE_CALL();
2305 ALOGV("Surface::setAsyncMode");
2306 Mutex::Autolock lock(mMutex);
2307
2308 status_t err = mGraphicBufferProducer->setAsyncMode(async);
2309 ALOGE_IF(err, "IGraphicBufferProducer::setAsyncMode(%d) returned %s",
2310 async, strerror(-err));
2311
2312 return err;
2313 }
2314
setSharedBufferMode(bool sharedBufferMode)2315 int Surface::setSharedBufferMode(bool sharedBufferMode) {
2316 ATRACE_CALL();
2317 ALOGV("Surface::setSharedBufferMode (%d)", sharedBufferMode);
2318 Mutex::Autolock lock(mMutex);
2319
2320 status_t err = mGraphicBufferProducer->setSharedBufferMode(
2321 sharedBufferMode);
2322 if (err == NO_ERROR) {
2323 mSharedBufferMode = sharedBufferMode;
2324 }
2325 ALOGE_IF(err, "IGraphicBufferProducer::setSharedBufferMode(%d) returned"
2326 "%s", sharedBufferMode, strerror(-err));
2327
2328 return err;
2329 }
2330
setAutoRefresh(bool autoRefresh)2331 int Surface::setAutoRefresh(bool autoRefresh) {
2332 ATRACE_CALL();
2333 ALOGV("Surface::setAutoRefresh (%d)", autoRefresh);
2334 Mutex::Autolock lock(mMutex);
2335
2336 status_t err = mGraphicBufferProducer->setAutoRefresh(autoRefresh);
2337 if (err == NO_ERROR) {
2338 mAutoRefresh = autoRefresh;
2339 }
2340 ALOGE_IF(err, "IGraphicBufferProducer::setAutoRefresh(%d) returned %s",
2341 autoRefresh, strerror(-err));
2342 return err;
2343 }
2344
setBuffersDimensions(uint32_t width,uint32_t height)2345 int Surface::setBuffersDimensions(uint32_t width, uint32_t height)
2346 {
2347 ATRACE_CALL();
2348 ALOGV("Surface::setBuffersDimensions");
2349
2350 if ((width && !height) || (!width && height))
2351 return BAD_VALUE;
2352
2353 Mutex::Autolock lock(mMutex);
2354 if (width != mReqWidth || height != mReqHeight) {
2355 mSharedBufferSlot = BufferItem::INVALID_BUFFER_SLOT;
2356 }
2357 mReqWidth = width;
2358 mReqHeight = height;
2359 return NO_ERROR;
2360 }
2361
setBuffersUserDimensions(uint32_t width,uint32_t height)2362 int Surface::setBuffersUserDimensions(uint32_t width, uint32_t height)
2363 {
2364 ATRACE_CALL();
2365 ALOGV("Surface::setBuffersUserDimensions");
2366
2367 if ((width && !height) || (!width && height))
2368 return BAD_VALUE;
2369
2370 Mutex::Autolock lock(mMutex);
2371 if (width != mUserWidth || height != mUserHeight) {
2372 mSharedBufferSlot = BufferItem::INVALID_BUFFER_SLOT;
2373 }
2374 mUserWidth = width;
2375 mUserHeight = height;
2376 return NO_ERROR;
2377 }
2378
setBuffersFormat(PixelFormat format)2379 int Surface::setBuffersFormat(PixelFormat format)
2380 {
2381 ALOGV("Surface::setBuffersFormat");
2382
2383 Mutex::Autolock lock(mMutex);
2384 if (format != mReqFormat) {
2385 mSharedBufferSlot = BufferItem::INVALID_BUFFER_SLOT;
2386 }
2387 mReqFormat = format;
2388 return NO_ERROR;
2389 }
2390
setScalingMode(int mode)2391 int Surface::setScalingMode(int mode)
2392 {
2393 ATRACE_CALL();
2394 ALOGV("Surface::setScalingMode(%d)", mode);
2395
2396 switch (mode) {
2397 case NATIVE_WINDOW_SCALING_MODE_FREEZE:
2398 case NATIVE_WINDOW_SCALING_MODE_SCALE_TO_WINDOW:
2399 case NATIVE_WINDOW_SCALING_MODE_SCALE_CROP:
2400 case NATIVE_WINDOW_SCALING_MODE_NO_SCALE_CROP:
2401 break;
2402 default:
2403 ALOGE("unknown scaling mode: %d", mode);
2404 return BAD_VALUE;
2405 }
2406
2407 Mutex::Autolock lock(mMutex);
2408 mScalingMode = mode;
2409 return NO_ERROR;
2410 }
2411
setBuffersTransform(uint32_t transform)2412 int Surface::setBuffersTransform(uint32_t transform)
2413 {
2414 ATRACE_CALL();
2415 ALOGV("Surface::setBuffersTransform");
2416 Mutex::Autolock lock(mMutex);
2417 // Ensure NATIVE_WINDOW_TRANSFORM_INVERSE_DISPLAY is sticky. If the client sets the flag, do not
2418 // override it until the surface is disconnected. This is a temporary workaround for camera
2419 // until they switch to using Buffer State Layers. Currently if client sets the buffer transform
2420 // it may be overriden by the buffer producer when the producer sets the buffer transform.
2421 if (transformToDisplayInverse()) {
2422 transform |= NATIVE_WINDOW_TRANSFORM_INVERSE_DISPLAY;
2423 }
2424 mTransform = transform;
2425 return NO_ERROR;
2426 }
2427
setBuffersStickyTransform(uint32_t transform)2428 int Surface::setBuffersStickyTransform(uint32_t transform)
2429 {
2430 ATRACE_CALL();
2431 ALOGV("Surface::setBuffersStickyTransform");
2432 Mutex::Autolock lock(mMutex);
2433 mStickyTransform = transform;
2434 return NO_ERROR;
2435 }
2436
setBuffersTimestamp(int64_t timestamp)2437 int Surface::setBuffersTimestamp(int64_t timestamp)
2438 {
2439 ALOGV("Surface::setBuffersTimestamp");
2440 Mutex::Autolock lock(mMutex);
2441 mTimestamp = timestamp;
2442 return NO_ERROR;
2443 }
2444
setBuffersDataSpace(Dataspace dataSpace)2445 int Surface::setBuffersDataSpace(Dataspace dataSpace)
2446 {
2447 ALOGV("Surface::setBuffersDataSpace");
2448 Mutex::Autolock lock(mMutex);
2449 mDataSpace = dataSpace;
2450 return NO_ERROR;
2451 }
2452
setBuffersSmpte2086Metadata(const android_smpte2086_metadata * metadata)2453 int Surface::setBuffersSmpte2086Metadata(const android_smpte2086_metadata* metadata) {
2454 ALOGV("Surface::setBuffersSmpte2086Metadata");
2455 Mutex::Autolock lock(mMutex);
2456 mHdrMetadataIsSet |= HdrMetadata::SMPTE2086;
2457 if (metadata) {
2458 mHdrMetadata.smpte2086 = *metadata;
2459 mHdrMetadata.validTypes |= HdrMetadata::SMPTE2086;
2460 } else {
2461 mHdrMetadata.validTypes &= ~HdrMetadata::SMPTE2086;
2462 }
2463 return NO_ERROR;
2464 }
2465
setBuffersCta8613Metadata(const android_cta861_3_metadata * metadata)2466 int Surface::setBuffersCta8613Metadata(const android_cta861_3_metadata* metadata) {
2467 ALOGV("Surface::setBuffersCta8613Metadata");
2468 Mutex::Autolock lock(mMutex);
2469 mHdrMetadataIsSet |= HdrMetadata::CTA861_3;
2470 if (metadata) {
2471 mHdrMetadata.cta8613 = *metadata;
2472 mHdrMetadata.validTypes |= HdrMetadata::CTA861_3;
2473 } else {
2474 mHdrMetadata.validTypes &= ~HdrMetadata::CTA861_3;
2475 }
2476 return NO_ERROR;
2477 }
2478
setBuffersHdr10PlusMetadata(const size_t size,const uint8_t * metadata)2479 int Surface::setBuffersHdr10PlusMetadata(const size_t size, const uint8_t* metadata) {
2480 ALOGV("Surface::setBuffersBlobMetadata");
2481 Mutex::Autolock lock(mMutex);
2482 mHdrMetadataIsSet |= HdrMetadata::HDR10PLUS;
2483 if (size > 0) {
2484 mHdrMetadata.hdr10plus.assign(metadata, metadata + size);
2485 mHdrMetadata.validTypes |= HdrMetadata::HDR10PLUS;
2486 } else {
2487 mHdrMetadata.validTypes &= ~HdrMetadata::HDR10PLUS;
2488 mHdrMetadata.hdr10plus.clear();
2489 }
2490 return NO_ERROR;
2491 }
2492
getBuffersDataSpace()2493 Dataspace Surface::getBuffersDataSpace() {
2494 ALOGV("Surface::getBuffersDataSpace");
2495 Mutex::Autolock lock(mMutex);
2496 return mDataSpace;
2497 }
2498
freeAllBuffers()2499 void Surface::freeAllBuffers() {
2500 if (!mDequeuedSlots.empty()) {
2501 ALOGE("%s: %zu buffers were freed while being dequeued!",
2502 __FUNCTION__, mDequeuedSlots.size());
2503 }
2504 for (int i = 0; i < NUM_BUFFER_SLOTS; i++) {
2505 mSlots[i].buffer = nullptr;
2506 }
2507 }
2508
getAndFlushBuffersFromSlots(const std::vector<int32_t> & slots,std::vector<sp<GraphicBuffer>> * outBuffers)2509 status_t Surface::getAndFlushBuffersFromSlots(const std::vector<int32_t>& slots,
2510 std::vector<sp<GraphicBuffer>>* outBuffers) {
2511 ALOGV("Surface::getAndFlushBuffersFromSlots");
2512 for (int32_t i : slots) {
2513 if (i < 0 || i >= NUM_BUFFER_SLOTS) {
2514 ALOGE("%s: Invalid slotIndex: %d", __FUNCTION__, i);
2515 return BAD_VALUE;
2516 }
2517 }
2518
2519 Mutex::Autolock lock(mMutex);
2520 for (int32_t i : slots) {
2521 if (mSlots[i].buffer == nullptr) {
2522 ALOGW("%s: Discarded slot %d doesn't contain buffer!", __FUNCTION__, i);
2523 continue;
2524 }
2525 // Don't flush currently dequeued buffers
2526 if (mDequeuedSlots.count(i) > 0) {
2527 continue;
2528 }
2529 outBuffers->push_back(mSlots[i].buffer);
2530 mSlots[i].buffer = nullptr;
2531 }
2532 return OK;
2533 }
2534
setSurfaceDamage(android_native_rect_t * rects,size_t numRects)2535 void Surface::setSurfaceDamage(android_native_rect_t* rects, size_t numRects) {
2536 ATRACE_CALL();
2537 ALOGV("Surface::setSurfaceDamage");
2538 Mutex::Autolock lock(mMutex);
2539
2540 if (mConnectedToCpu || numRects == 0) {
2541 mDirtyRegion = Region::INVALID_REGION;
2542 return;
2543 }
2544
2545 mDirtyRegion.clear();
2546 for (size_t r = 0; r < numRects; ++r) {
2547 // We intentionally flip top and bottom here, since because they're
2548 // specified with a bottom-left origin, top > bottom, which fails
2549 // validation in the Region class. We will fix this up when we flip to a
2550 // top-left origin in queueBuffer.
2551 Rect rect(rects[r].left, rects[r].bottom, rects[r].right, rects[r].top);
2552 mDirtyRegion.orSelf(rect);
2553 }
2554 }
2555
2556 // ----------------------------------------------------------------------
2557 // the lock/unlock APIs must be used from the same thread
2558
copyBlt(const sp<GraphicBuffer> & dst,const sp<GraphicBuffer> & src,const Region & reg,int * dstFenceFd)2559 static status_t copyBlt(
2560 const sp<GraphicBuffer>& dst,
2561 const sp<GraphicBuffer>& src,
2562 const Region& reg,
2563 int *dstFenceFd)
2564 {
2565 if (dst->getId() == src->getId())
2566 return OK;
2567
2568 // src and dst with, height and format must be identical. no verification
2569 // is done here.
2570 status_t err;
2571 uint8_t* src_bits = nullptr;
2572 err = src->lock(GRALLOC_USAGE_SW_READ_OFTEN, reg.bounds(),
2573 reinterpret_cast<void**>(&src_bits));
2574 ALOGE_IF(err, "error locking src buffer %s", strerror(-err));
2575
2576 uint8_t* dst_bits = nullptr;
2577 err = dst->lockAsync(GRALLOC_USAGE_SW_WRITE_OFTEN, reg.bounds(),
2578 reinterpret_cast<void**>(&dst_bits), *dstFenceFd);
2579 ALOGE_IF(err, "error locking dst buffer %s", strerror(-err));
2580 *dstFenceFd = -1;
2581
2582 Region::const_iterator head(reg.begin());
2583 Region::const_iterator tail(reg.end());
2584 if (head != tail && src_bits && dst_bits) {
2585 const size_t bpp = bytesPerPixel(src->format);
2586 const size_t dbpr = static_cast<uint32_t>(dst->stride) * bpp;
2587 const size_t sbpr = static_cast<uint32_t>(src->stride) * bpp;
2588
2589 while (head != tail) {
2590 const Rect& r(*head++);
2591 int32_t h = r.height();
2592 if (h <= 0) continue;
2593 size_t size = static_cast<uint32_t>(r.width()) * bpp;
2594 uint8_t const * s = src_bits +
2595 static_cast<uint32_t>(r.left + src->stride * r.top) * bpp;
2596 uint8_t * d = dst_bits +
2597 static_cast<uint32_t>(r.left + dst->stride * r.top) * bpp;
2598 if (dbpr==sbpr && size==sbpr) {
2599 size *= static_cast<size_t>(h);
2600 h = 1;
2601 }
2602 do {
2603 memcpy(d, s, size);
2604 d += dbpr;
2605 s += sbpr;
2606 } while (--h > 0);
2607 }
2608 }
2609
2610 if (src_bits)
2611 src->unlock();
2612
2613 if (dst_bits)
2614 dst->unlockAsync(dstFenceFd);
2615
2616 return err;
2617 }
2618
2619 // ----------------------------------------------------------------------------
2620
lock(ANativeWindow_Buffer * outBuffer,ARect * inOutDirtyBounds)2621 status_t Surface::lock(
2622 ANativeWindow_Buffer* outBuffer, ARect* inOutDirtyBounds)
2623 {
2624 if (mLockedBuffer != nullptr) {
2625 ALOGE("Surface::lock failed, already locked");
2626 return INVALID_OPERATION;
2627 }
2628
2629 if (!mConnectedToCpu) {
2630 int err = Surface::connect(NATIVE_WINDOW_API_CPU);
2631 if (err) {
2632 return err;
2633 }
2634 // we're intending to do software rendering from this point
2635 setUsage(GRALLOC_USAGE_SW_READ_OFTEN | GRALLOC_USAGE_SW_WRITE_OFTEN);
2636 }
2637
2638 ANativeWindowBuffer* out;
2639 int fenceFd = -1;
2640 status_t err = dequeueBuffer(&out, &fenceFd);
2641 ALOGE_IF(err, "dequeueBuffer failed (%s)", strerror(-err));
2642 if (err == NO_ERROR) {
2643 sp<GraphicBuffer> backBuffer(GraphicBuffer::getSelf(out));
2644 const Rect bounds(backBuffer->width, backBuffer->height);
2645
2646 Region newDirtyRegion;
2647 if (inOutDirtyBounds) {
2648 newDirtyRegion.set(static_cast<Rect const&>(*inOutDirtyBounds));
2649 newDirtyRegion.andSelf(bounds);
2650 } else {
2651 newDirtyRegion.set(bounds);
2652 }
2653
2654 // figure out if we can copy the frontbuffer back
2655 const sp<GraphicBuffer>& frontBuffer(mPostedBuffer);
2656 const bool canCopyBack = (frontBuffer != nullptr &&
2657 backBuffer->width == frontBuffer->width &&
2658 backBuffer->height == frontBuffer->height &&
2659 backBuffer->format == frontBuffer->format);
2660
2661 if (canCopyBack) {
2662 // copy the area that is invalid and not repainted this round
2663 const Region copyback(mDirtyRegion.subtract(newDirtyRegion));
2664 if (!copyback.isEmpty()) {
2665 copyBlt(backBuffer, frontBuffer, copyback, &fenceFd);
2666 }
2667 } else {
2668 // if we can't copy-back anything, modify the user's dirty
2669 // region to make sure they redraw the whole buffer
2670 newDirtyRegion.set(bounds);
2671 mDirtyRegion.clear();
2672 Mutex::Autolock lock(mMutex);
2673 for (size_t i=0 ; i<NUM_BUFFER_SLOTS ; i++) {
2674 mSlots[i].dirtyRegion.clear();
2675 }
2676 }
2677
2678
2679 { // scope for the lock
2680 Mutex::Autolock lock(mMutex);
2681 int backBufferSlot(getSlotFromBufferLocked(backBuffer.get()));
2682 if (backBufferSlot >= 0) {
2683 Region& dirtyRegion(mSlots[backBufferSlot].dirtyRegion);
2684 mDirtyRegion.subtract(dirtyRegion);
2685 dirtyRegion = newDirtyRegion;
2686 }
2687 }
2688
2689 mDirtyRegion.orSelf(newDirtyRegion);
2690 if (inOutDirtyBounds) {
2691 *inOutDirtyBounds = newDirtyRegion.getBounds();
2692 }
2693
2694 void* vaddr;
2695 status_t res = backBuffer->lockAsync(
2696 GRALLOC_USAGE_SW_READ_OFTEN | GRALLOC_USAGE_SW_WRITE_OFTEN,
2697 newDirtyRegion.bounds(), &vaddr, fenceFd);
2698
2699 ALOGW_IF(res, "failed locking buffer (handle = %p)",
2700 backBuffer->handle);
2701
2702 if (res != 0) {
2703 err = INVALID_OPERATION;
2704 } else {
2705 mLockedBuffer = backBuffer;
2706 outBuffer->width = backBuffer->width;
2707 outBuffer->height = backBuffer->height;
2708 outBuffer->stride = backBuffer->stride;
2709 outBuffer->format = backBuffer->format;
2710 outBuffer->bits = vaddr;
2711 }
2712 }
2713 return err;
2714 }
2715
unlockAndPost()2716 status_t Surface::unlockAndPost()
2717 {
2718 if (mLockedBuffer == nullptr) {
2719 ALOGE("Surface::unlockAndPost failed, no locked buffer");
2720 return INVALID_OPERATION;
2721 }
2722
2723 int fd = -1;
2724 status_t err = mLockedBuffer->unlockAsync(&fd);
2725 ALOGE_IF(err, "failed unlocking buffer (%p)", mLockedBuffer->handle);
2726
2727 err = queueBuffer(mLockedBuffer.get(), fd);
2728 ALOGE_IF(err, "queueBuffer (handle=%p) failed (%s)",
2729 mLockedBuffer->handle, strerror(-err));
2730
2731 mPostedBuffer = mLockedBuffer;
2732 mLockedBuffer = nullptr;
2733 return err;
2734 }
2735
waitForNextFrame(uint64_t lastFrame,nsecs_t timeout)2736 bool Surface::waitForNextFrame(uint64_t lastFrame, nsecs_t timeout) {
2737 Mutex::Autolock lock(mMutex);
2738 if (mLastFrameNumber > lastFrame) {
2739 return true;
2740 }
2741 return mQueueBufferCondition.waitRelative(mMutex, timeout) == OK;
2742 }
2743
getUniqueId(uint64_t * outId) const2744 status_t Surface::getUniqueId(uint64_t* outId) const {
2745 Mutex::Autolock lock(mMutex);
2746 return mGraphicBufferProducer->getUniqueId(outId);
2747 }
2748
getConsumerUsage(uint64_t * outUsage) const2749 int Surface::getConsumerUsage(uint64_t* outUsage) const {
2750 Mutex::Autolock lock(mMutex);
2751 return mGraphicBufferProducer->getConsumerUsage(outUsage);
2752 }
2753
getAndFlushRemovedBuffers(std::vector<sp<GraphicBuffer>> * out)2754 status_t Surface::getAndFlushRemovedBuffers(std::vector<sp<GraphicBuffer>>* out) {
2755 if (out == nullptr) {
2756 ALOGE("%s: out must not be null!", __FUNCTION__);
2757 return BAD_VALUE;
2758 }
2759
2760 Mutex::Autolock lock(mMutex);
2761 *out = mRemovedBuffers;
2762 mRemovedBuffers.clear();
2763 return OK;
2764 }
2765
attachAndQueueBufferWithDataspace(Surface * surface,sp<GraphicBuffer> buffer,Dataspace dataspace)2766 status_t Surface::attachAndQueueBufferWithDataspace(Surface* surface, sp<GraphicBuffer> buffer,
2767 Dataspace dataspace) {
2768 if (buffer == nullptr) {
2769 return BAD_VALUE;
2770 }
2771 int err = static_cast<ANativeWindow*>(surface)->perform(surface, NATIVE_WINDOW_API_CONNECT,
2772 NATIVE_WINDOW_API_CPU);
2773 if (err != OK) {
2774 return err;
2775 }
2776 ui::Dataspace tmpDataspace = surface->getBuffersDataSpace();
2777 err = surface->setBuffersDataSpace(dataspace);
2778 if (err != OK) {
2779 return err;
2780 }
2781 err = surface->attachBuffer(buffer->getNativeBuffer());
2782 if (err != OK) {
2783 return err;
2784 }
2785 err = static_cast<ANativeWindow*>(surface)->queueBuffer(surface, buffer->getNativeBuffer(), -1);
2786 if (err != OK) {
2787 return err;
2788 }
2789 err = surface->setBuffersDataSpace(tmpDataspace);
2790 if (err != OK) {
2791 return err;
2792 }
2793 err = surface->disconnect(NATIVE_WINDOW_API_CPU);
2794 return err;
2795 }
2796
setAutoPrerotation(bool autoPrerotation)2797 int Surface::setAutoPrerotation(bool autoPrerotation) {
2798 ATRACE_CALL();
2799 ALOGV("Surface::setAutoPrerotation (%d)", autoPrerotation);
2800 Mutex::Autolock lock(mMutex);
2801
2802 if (mAutoPrerotation == autoPrerotation) {
2803 return OK;
2804 }
2805
2806 status_t err = mGraphicBufferProducer->setAutoPrerotation(autoPrerotation);
2807 if (err == NO_ERROR) {
2808 mAutoPrerotation = autoPrerotation;
2809 }
2810 ALOGE_IF(err, "IGraphicBufferProducer::setAutoPrerotation(%d) returned %s", autoPrerotation,
2811 strerror(-err));
2812 return err;
2813 }
2814
onBuffersDiscarded(const std::vector<int32_t> & slots)2815 void Surface::ProducerListenerProxy::onBuffersDiscarded(const std::vector<int32_t>& slots) {
2816 ATRACE_CALL();
2817 sp<Surface> parent = mParent.promote();
2818 if (parent == nullptr) {
2819 return;
2820 }
2821
2822 std::vector<sp<GraphicBuffer>> discardedBufs;
2823 status_t res = parent->getAndFlushBuffersFromSlots(slots, &discardedBufs);
2824 if (res != OK) {
2825 ALOGE("%s: Failed to get buffers from slots: %s(%d)", __FUNCTION__,
2826 strerror(-res), res);
2827 return;
2828 }
2829
2830 mSurfaceListener->onBuffersDiscarded(discardedBufs);
2831 }
2832
setFrameRate(float frameRate,int8_t compatibility,int8_t changeFrameRateStrategy)2833 status_t Surface::setFrameRate(float frameRate, int8_t compatibility,
2834 int8_t changeFrameRateStrategy) {
2835 #if COM_ANDROID_GRAPHICS_LIBGUI_FLAGS(BQ_SETFRAMERATE)
2836 if (flags::bq_setframerate()) {
2837 status_t err = mGraphicBufferProducer->setFrameRate(frameRate, compatibility,
2838 changeFrameRateStrategy);
2839 ALOGE_IF(err, "IGraphicBufferProducer::setFrameRate(%.2f) returned %s", frameRate,
2840 strerror(-err));
2841 return err;
2842 }
2843 #else
2844 static_cast<void>(frameRate);
2845 static_cast<void>(compatibility);
2846 static_cast<void>(changeFrameRateStrategy);
2847 #endif
2848
2849 ALOGI("Surface::setFrameRate is deprecated, setFrameRate hint is dropped as destination is not "
2850 "SurfaceFlinger");
2851 // ISurfaceComposer no longer supports setFrameRate, we will return NO_ERROR when the api is
2852 // called to avoid apps crashing, as BAD_VALUE can generate fatal exception in apps.
2853 return NO_ERROR;
2854 }
2855
setFrameTimelineInfo(uint64_t,const FrameTimelineInfo &)2856 status_t Surface::setFrameTimelineInfo(uint64_t /*frameNumber*/,
2857 const FrameTimelineInfo& /*frameTimelineInfo*/) {
2858 // ISurfaceComposer no longer supports setFrameTimelineInfo
2859 return BAD_VALUE;
2860 }
2861
2862 #if COM_ANDROID_GRAPHICS_LIBGUI_FLAGS(BQ_EXTENDEDALLOCATE)
setAdditionalOptions(const std::vector<gui::AdditionalOptions> & options)2863 status_t Surface::setAdditionalOptions(const std::vector<gui::AdditionalOptions>& options) {
2864 if (!GraphicBufferAllocator::get().supportsAdditionalOptions()) {
2865 return INVALID_OPERATION;
2866 }
2867
2868 Mutex::Autolock lock(mMutex);
2869 return mGraphicBufferProducer->setAdditionalOptions(options);
2870 }
2871 #endif
2872
getSurfaceControlHandle() const2873 sp<IBinder> Surface::getSurfaceControlHandle() const {
2874 Mutex::Autolock lock(mMutex);
2875 return mSurfaceControlHandle;
2876 }
2877
destroy()2878 void Surface::destroy() {
2879 Mutex::Autolock lock(mMutex);
2880 mSurfaceControlHandle = nullptr;
2881 }
2882
getDebugName()2883 const char* Surface::getDebugName() {
2884 std::unique_lock lock{mNameMutex};
2885 if (mName.empty()) {
2886 mName = getConsumerName();
2887 }
2888 return mName.c_str();
2889 }
2890
2891 }; // namespace android
2892