1 /*
2 * Copyright (C) 2008 The Android Open Source Project
3 * Copyright (c) 2010-2013, The Linux Foundation. All rights reserved.
4 *
5 * Not a Contribution, Apache license notifications and license are retained
6 * for attribution purposes only.
7 *
8 * Licensed under the Apache License, Version 2.0 (the "License");
9 * you may not use this file except in compliance with the License.
10 * You may obtain a copy of the License at
11 *
12 * http://www.apache.org/licenses/LICENSE-2.0
13 *
14 * Unless required by applicable law or agreed to in writing, software
15 * distributed under the License is distributed on an "AS IS" BASIS,
16 * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
17 * See the License for the specific language governing permissions and
18 * limitations under the License.
19 */
20 #include <cutils/log.h>
21 #include <sys/resource.h>
22 #include <sys/prctl.h>
23
24 #include <stdint.h>
25 #include <string.h>
26 #include <unistd.h>
27 #include <errno.h>
28 #include <fcntl.h>
29
30 #include <sys/ioctl.h>
31 #include <sys/types.h>
32 #include <sys/mman.h>
33
34 #include <linux/msm_kgsl.h>
35
36 #include <EGL/eglplatform.h>
37 #include <cutils/native_handle.h>
38 #include <cutils/ashmem.h>
39 #include <linux/ashmem.h>
40 #include <gralloc_priv.h>
41
42 #include <copybit.h>
43 #include <alloc_controller.h>
44 #include <memalloc.h>
45
46 #include "c2d2.h"
47 #include "software_converter.h"
48
49 #include <dlfcn.h>
50
51 using gralloc::IMemAlloc;
52 using gralloc::IonController;
53 using gralloc::alloc_data;
54
55 C2D_STATUS (*LINK_c2dCreateSurface)( uint32 *surface_id,
56 uint32 surface_bits,
57 C2D_SURFACE_TYPE surface_type,
58 void *surface_definition );
59
60 C2D_STATUS (*LINK_c2dUpdateSurface)( uint32 surface_id,
61 uint32 surface_bits,
62 C2D_SURFACE_TYPE surface_type,
63 void *surface_definition );
64
65 C2D_STATUS (*LINK_c2dReadSurface)( uint32 surface_id,
66 C2D_SURFACE_TYPE surface_type,
67 void *surface_definition,
68 int32 x, int32 y );
69
70 C2D_STATUS (*LINK_c2dDraw)( uint32 target_id,
71 uint32 target_config, C2D_RECT *target_scissor,
72 uint32 target_mask_id, uint32 target_color_key,
73 C2D_OBJECT *objects_list, uint32 num_objects );
74
75 C2D_STATUS (*LINK_c2dFinish)( uint32 target_id);
76
77 C2D_STATUS (*LINK_c2dFlush)( uint32 target_id, c2d_ts_handle *timestamp);
78
79 C2D_STATUS (*LINK_c2dWaitTimestamp)( c2d_ts_handle timestamp );
80
81 C2D_STATUS (*LINK_c2dDestroySurface)( uint32 surface_id );
82
83 C2D_STATUS (*LINK_c2dMapAddr) ( int mem_fd, void * hostptr, uint32 len,
84 uint32 offset, uint32 flags, void ** gpuaddr);
85
86 C2D_STATUS (*LINK_c2dUnMapAddr) ( void * gpuaddr);
87
88 C2D_STATUS (*LINK_c2dGetDriverCapabilities) ( C2D_DRIVER_INFO * driver_info);
89
90 /* create a fence fd for the timestamp */
91 C2D_STATUS (*LINK_c2dCreateFenceFD) ( uint32 target_id, c2d_ts_handle timestamp,
92 int32 *fd);
93
94 C2D_STATUS (*LINK_c2dFillSurface) ( uint32 surface_id, uint32 fill_color,
95 C2D_RECT * fill_rect);
96
97 /******************************************************************************/
98
99 #if defined(COPYBIT_Z180)
100 #define MAX_SCALE_FACTOR (4096)
101 #define MAX_DIMENSION (4096)
102 #else
103 #error "Unsupported HW version"
104 #endif
105
106 // The following defines can be changed as required i.e. as we encounter
107 // complex use cases.
108 #define MAX_RGB_SURFACES 32 // Max. RGB layers currently supported per draw
109 #define MAX_YUV_2_PLANE_SURFACES 4// Max. 2-plane YUV layers currently supported per draw
110 #define MAX_YUV_3_PLANE_SURFACES 1// Max. 3-plane YUV layers currently supported per draw
111 // +1 for the destination surface. We cannot have multiple destination surfaces.
112 #define MAX_SURFACES (MAX_RGB_SURFACES + MAX_YUV_2_PLANE_SURFACES + MAX_YUV_3_PLANE_SURFACES + 1)
113 #define NUM_SURFACE_TYPES 3 // RGB_SURFACE + YUV_SURFACE_2_PLANES + YUV_SURFACE_3_PLANES
114 #define MAX_BLIT_OBJECT_COUNT 50 // Max. blit objects that can be passed per draw
115
116 enum {
117 RGB_SURFACE,
118 YUV_SURFACE_2_PLANES,
119 YUV_SURFACE_3_PLANES
120 };
121
122 enum eConversionType {
123 CONVERT_TO_ANDROID_FORMAT,
124 CONVERT_TO_C2D_FORMAT
125 };
126
127 enum eC2DFlags {
128 FLAGS_PREMULTIPLIED_ALPHA = 1<<0,
129 FLAGS_YUV_DESTINATION = 1<<1,
130 FLAGS_TEMP_SRC_DST = 1<<2
131 };
132
133 static gralloc::IAllocController* sAlloc = 0;
134 /******************************************************************************/
135
136 /** State information for each device instance */
137 struct copybit_context_t {
138 struct copybit_device_t device;
139 // Templates for the various source surfaces. These templates are created
140 // to avoid the expensive create/destroy C2D Surfaces
141 C2D_OBJECT_STR blit_rgb_object[MAX_RGB_SURFACES];
142 C2D_OBJECT_STR blit_yuv_2_plane_object[MAX_YUV_2_PLANE_SURFACES];
143 C2D_OBJECT_STR blit_yuv_3_plane_object[MAX_YUV_3_PLANE_SURFACES];
144 C2D_OBJECT_STR blit_list[MAX_BLIT_OBJECT_COUNT]; // Z-ordered list of blit objects
145 C2D_DRIVER_INFO c2d_driver_info;
146 void *libc2d2;
147 alloc_data temp_src_buffer;
148 alloc_data temp_dst_buffer;
149 unsigned int dst[NUM_SURFACE_TYPES]; // dst surfaces
150 unsigned int mapped_gpu_addr[MAX_SURFACES]; // GPU addresses mapped inside copybit
151 int blit_rgb_count; // Total RGB surfaces being blit
152 int blit_yuv_2_plane_count; // Total 2 plane YUV surfaces being
153 int blit_yuv_3_plane_count; // Total 3 plane YUV surfaces being blit
154 int blit_count; // Total blit objects.
155 unsigned int trg_transform; /* target transform */
156 int fb_width;
157 int fb_height;
158 int src_global_alpha;
159 int config_mask;
160 int dst_surface_type;
161 bool is_premultiplied_alpha;
162 void* time_stamp;
163 bool dst_surface_mapped; // Set when dst surface is mapped to GPU addr
164 void* dst_surface_base; // Stores the dst surface addr
165
166 // used for signaling the wait thread
167 bool wait_timestamp;
168 pthread_t wait_thread_id;
169 bool stop_thread;
170 pthread_mutex_t wait_cleanup_lock;
171 pthread_cond_t wait_cleanup_cond;
172
173 };
174
175 struct bufferInfo {
176 int width;
177 int height;
178 int format;
179 };
180
181 struct yuvPlaneInfo {
182 int yStride; //luma stride
183 int plane1_stride;
184 int plane2_stride;
185 int plane1_offset;
186 int plane2_offset;
187 };
188
189 /**
190 * Common hardware methods
191 */
192
193 static int open_copybit(const struct hw_module_t* module, const char* name,
194 struct hw_device_t** device);
195
196 static struct hw_module_methods_t copybit_module_methods = {
197 open: open_copybit
198 };
199
200 /*
201 * The COPYBIT Module
202 */
203 struct copybit_module_t HAL_MODULE_INFO_SYM = {
204 common: {
205 tag: HARDWARE_MODULE_TAG,
206 version_major: 1,
207 version_minor: 0,
208 id: COPYBIT_HARDWARE_MODULE_ID,
209 name: "QCT COPYBIT C2D 2.0 Module",
210 author: "Qualcomm",
211 methods: ©bit_module_methods
212 }
213 };
214
215
216 /* thread function which waits on the timeStamp and cleans up the surfaces */
c2d_wait_loop(void * ptr)217 static void* c2d_wait_loop(void* ptr) {
218 copybit_context_t* ctx = (copybit_context_t*)(ptr);
219 char thread_name[64] = "copybitWaitThr";
220 prctl(PR_SET_NAME, (unsigned long) &thread_name, 0, 0, 0);
221 setpriority(PRIO_PROCESS, 0, HAL_PRIORITY_URGENT_DISPLAY);
222
223 while(ctx->stop_thread == false) {
224 pthread_mutex_lock(&ctx->wait_cleanup_lock);
225 while(ctx->wait_timestamp == false && !ctx->stop_thread) {
226 pthread_cond_wait(&(ctx->wait_cleanup_cond),
227 &(ctx->wait_cleanup_lock));
228 }
229 if(ctx->wait_timestamp) {
230 if(LINK_c2dWaitTimestamp(ctx->time_stamp)) {
231 ALOGE("%s: LINK_c2dWaitTimeStamp ERROR!!", __FUNCTION__);
232 }
233 ctx->wait_timestamp = false;
234 // Unmap any mapped addresses.
235 for (int i = 0; i < MAX_SURFACES; i++) {
236 if (ctx->mapped_gpu_addr[i]) {
237 LINK_c2dUnMapAddr( (void*)ctx->mapped_gpu_addr[i]);
238 ctx->mapped_gpu_addr[i] = 0;
239 }
240 }
241 // Reset the counts after the draw.
242 ctx->blit_rgb_count = 0;
243 ctx->blit_yuv_2_plane_count = 0;
244 ctx->blit_yuv_3_plane_count = 0;
245 ctx->blit_count = 0;
246 ctx->dst_surface_mapped = false;
247 ctx->dst_surface_base = 0;
248 }
249 pthread_mutex_unlock(&ctx->wait_cleanup_lock);
250 if(ctx->stop_thread)
251 break;
252 }
253 pthread_exit(NULL);
254 return NULL;
255 }
256
257
258 /* convert COPYBIT_FORMAT to C2D format */
get_format(int format)259 static int get_format(int format) {
260 switch (format) {
261 case HAL_PIXEL_FORMAT_RGB_565: return C2D_COLOR_FORMAT_565_RGB;
262 case HAL_PIXEL_FORMAT_RGBX_8888: return C2D_COLOR_FORMAT_8888_ARGB |
263 C2D_FORMAT_SWAP_RB |
264 C2D_FORMAT_DISABLE_ALPHA;
265 case HAL_PIXEL_FORMAT_RGBA_8888: return C2D_COLOR_FORMAT_8888_ARGB |
266 C2D_FORMAT_SWAP_RB;
267 case HAL_PIXEL_FORMAT_BGRA_8888: return C2D_COLOR_FORMAT_8888_ARGB;
268 case HAL_PIXEL_FORMAT_YCbCr_420_SP: return C2D_COLOR_FORMAT_420_NV12;
269 case HAL_PIXEL_FORMAT_NV12_ENCODEABLE:return C2D_COLOR_FORMAT_420_NV12;
270 case HAL_PIXEL_FORMAT_YCrCb_420_SP: return C2D_COLOR_FORMAT_420_NV21;
271 case HAL_PIXEL_FORMAT_YCbCr_420_SP_TILED: return C2D_COLOR_FORMAT_420_NV12 |
272 C2D_FORMAT_MACROTILED;
273 default: ALOGE("%s: invalid format (0x%x",
274 __FUNCTION__, format);
275 return -EINVAL;
276 }
277 return -EINVAL;
278 }
279
280 /* Get the C2D formats needed for conversion to YUV */
get_c2d_format_for_yuv_destination(int halFormat)281 static int get_c2d_format_for_yuv_destination(int halFormat) {
282 switch (halFormat) {
283 // We do not swap the RB when the target is YUV
284 case HAL_PIXEL_FORMAT_RGBX_8888: return C2D_COLOR_FORMAT_8888_ARGB |
285 C2D_FORMAT_DISABLE_ALPHA;
286 case HAL_PIXEL_FORMAT_RGBA_8888: return C2D_COLOR_FORMAT_8888_ARGB;
287 // The U and V need to be interchanged when the target is YUV
288 case HAL_PIXEL_FORMAT_YCbCr_420_SP: return C2D_COLOR_FORMAT_420_NV21;
289 case HAL_PIXEL_FORMAT_NV12_ENCODEABLE:return C2D_COLOR_FORMAT_420_NV21;
290 case HAL_PIXEL_FORMAT_YCrCb_420_SP: return C2D_COLOR_FORMAT_420_NV12;
291 default: return get_format(halFormat);
292 }
293 return -EINVAL;
294 }
295
296 /* ------------------------------------------------------------------- *//*!
297 * \internal
298 * \brief Get the bpp for a particular color format
299 * \param color format
300 * \return bits per pixel
301 *//* ------------------------------------------------------------------- */
c2diGetBpp(int32 colorformat)302 int c2diGetBpp(int32 colorformat)
303 {
304
305 int c2dBpp = 0;
306
307 switch(colorformat&0xFF)
308 {
309 case C2D_COLOR_FORMAT_4444_RGBA:
310 case C2D_COLOR_FORMAT_4444_ARGB:
311 case C2D_COLOR_FORMAT_1555_ARGB:
312 case C2D_COLOR_FORMAT_565_RGB:
313 case C2D_COLOR_FORMAT_5551_RGBA:
314 c2dBpp = 16;
315 break;
316 case C2D_COLOR_FORMAT_8888_RGBA:
317 case C2D_COLOR_FORMAT_8888_ARGB:
318 c2dBpp = 32;
319 break;
320 case C2D_COLOR_FORMAT_8_L:
321 case C2D_COLOR_FORMAT_8_A:
322 c2dBpp = 8;
323 break;
324 case C2D_COLOR_FORMAT_4_A:
325 c2dBpp = 4;
326 break;
327 case C2D_COLOR_FORMAT_1:
328 c2dBpp = 1;
329 break;
330 default:
331 ALOGE("%s ERROR", __func__);
332 break;
333 }
334 return c2dBpp;
335 }
336
c2d_get_gpuaddr(copybit_context_t * ctx,struct private_handle_t * handle,int & mapped_idx)337 static uint32 c2d_get_gpuaddr(copybit_context_t* ctx,
338 struct private_handle_t *handle, int &mapped_idx)
339 {
340 uint32 memtype, *gpuaddr = 0;
341 C2D_STATUS rc;
342 int freeindex = 0;
343 bool mapaddr = false;
344
345 if(!handle)
346 return 0;
347
348 if (handle->flags & (private_handle_t::PRIV_FLAGS_USES_PMEM |
349 private_handle_t::PRIV_FLAGS_USES_PMEM_ADSP))
350 memtype = KGSL_USER_MEM_TYPE_PMEM;
351 else if (handle->flags & private_handle_t::PRIV_FLAGS_USES_ASHMEM)
352 memtype = KGSL_USER_MEM_TYPE_ASHMEM;
353 else if (handle->flags & private_handle_t::PRIV_FLAGS_USES_ION)
354 memtype = KGSL_USER_MEM_TYPE_ION;
355 else {
356 ALOGE("Invalid handle flags: 0x%x", handle->flags);
357 return 0;
358 }
359
360 // Check for a freeindex in the mapped_gpu_addr list
361 for (freeindex = 0; freeindex < MAX_SURFACES; freeindex++) {
362 if (ctx->mapped_gpu_addr[freeindex] == 0) {
363 // free index is available
364 // map GPU addr and use this as mapped_idx
365 mapaddr = true;
366 break;
367 }
368 }
369
370 if(mapaddr) {
371 rc = LINK_c2dMapAddr(handle->fd, (void*)handle->base, handle->size,
372 handle->offset, memtype, (void**)&gpuaddr);
373
374 if (rc == C2D_STATUS_OK) {
375 // We have mapped the GPU address inside copybit. We need to unmap
376 // this address after the blit. Store this address
377 ctx->mapped_gpu_addr[freeindex] = (uint32) gpuaddr;
378 mapped_idx = freeindex;
379 }
380 }
381 return (uint32) gpuaddr;
382 }
383
unmap_gpuaddr(copybit_context_t * ctx,int mapped_idx)384 static void unmap_gpuaddr(copybit_context_t* ctx, int mapped_idx)
385 {
386 if (!ctx || (mapped_idx == -1))
387 return;
388
389 if (ctx->mapped_gpu_addr[mapped_idx]) {
390 LINK_c2dUnMapAddr( (void*)ctx->mapped_gpu_addr[mapped_idx]);
391 ctx->mapped_gpu_addr[mapped_idx] = 0;
392 }
393 }
394
is_supported_rgb_format(int format)395 static int is_supported_rgb_format(int format)
396 {
397 switch(format) {
398 case HAL_PIXEL_FORMAT_RGBA_8888:
399 case HAL_PIXEL_FORMAT_RGBX_8888:
400 case HAL_PIXEL_FORMAT_RGB_565:
401 case HAL_PIXEL_FORMAT_BGRA_8888: {
402 return COPYBIT_SUCCESS;
403 }
404 default:
405 return COPYBIT_FAILURE;
406 }
407 }
408
get_num_planes(int format)409 static int get_num_planes(int format)
410 {
411 switch(format) {
412 case HAL_PIXEL_FORMAT_YCbCr_420_SP:
413 case HAL_PIXEL_FORMAT_YCrCb_420_SP:
414 case HAL_PIXEL_FORMAT_NV12_ENCODEABLE:
415 case HAL_PIXEL_FORMAT_YCbCr_420_SP_TILED: {
416 return 2;
417 }
418 case HAL_PIXEL_FORMAT_YV12: {
419 return 3;
420 }
421 default:
422 return COPYBIT_FAILURE;
423 }
424 }
425
is_supported_yuv_format(int format)426 static int is_supported_yuv_format(int format)
427 {
428 switch(format) {
429 case HAL_PIXEL_FORMAT_YCbCr_420_SP:
430 case HAL_PIXEL_FORMAT_YCrCb_420_SP:
431 case HAL_PIXEL_FORMAT_NV12_ENCODEABLE:
432 case HAL_PIXEL_FORMAT_YCbCr_420_SP_TILED: {
433 return COPYBIT_SUCCESS;
434 }
435 default:
436 return COPYBIT_FAILURE;
437 }
438 }
439
is_valid_destination_format(int format)440 static int is_valid_destination_format(int format)
441 {
442 if (format == HAL_PIXEL_FORMAT_YCbCr_420_SP_TILED) {
443 // C2D does not support NV12Tile as a destination format.
444 return COPYBIT_FAILURE;
445 }
446 return COPYBIT_SUCCESS;
447 }
448
calculate_yuv_offset_and_stride(const bufferInfo & info,yuvPlaneInfo & yuvInfo)449 static int calculate_yuv_offset_and_stride(const bufferInfo& info,
450 yuvPlaneInfo& yuvInfo)
451 {
452 int width = info.width;
453 int height = info.height;
454 int format = info.format;
455
456 int aligned_height = 0;
457 int aligned_width = 0, size = 0;
458
459 switch (format) {
460 case HAL_PIXEL_FORMAT_YCbCr_420_SP_TILED: {
461 /* NV12 Tile buffers have their luma height aligned to 32bytes and width
462 * aligned to 128 bytes. The chroma offset starts at an 8K boundary
463 */
464 aligned_height = ALIGN(height, 32);
465 aligned_width = ALIGN(width, 128);
466 size = aligned_width * aligned_height;
467 yuvInfo.plane1_offset = ALIGN(size,8192);
468 yuvInfo.yStride = aligned_width;
469 yuvInfo.plane1_stride = aligned_width;
470 break;
471 }
472 case HAL_PIXEL_FORMAT_YCbCr_420_SP:
473 case HAL_PIXEL_FORMAT_NV12_ENCODEABLE:
474 case HAL_PIXEL_FORMAT_YCrCb_420_SP: {
475 aligned_width = ALIGN(width, 32);
476 yuvInfo.yStride = aligned_width;
477 yuvInfo.plane1_stride = aligned_width;
478 if (HAL_PIXEL_FORMAT_NV12_ENCODEABLE == format) {
479 // The encoder requires a 2K aligned chroma offset
480 yuvInfo.plane1_offset = ALIGN(aligned_width * height, 2048);
481 } else
482 yuvInfo.plane1_offset = aligned_width * height;
483
484 break;
485 }
486 default: {
487 return COPYBIT_FAILURE;
488 }
489 }
490 return COPYBIT_SUCCESS;
491 }
492
493 /** create C2D surface from copybit image */
set_image(copybit_context_t * ctx,uint32 surfaceId,const struct copybit_image_t * rhs,const eC2DFlags flags,int & mapped_idx)494 static int set_image(copybit_context_t* ctx, uint32 surfaceId,
495 const struct copybit_image_t *rhs,
496 const eC2DFlags flags, int &mapped_idx)
497 {
498 struct private_handle_t* handle = (struct private_handle_t*)rhs->handle;
499 C2D_SURFACE_TYPE surfaceType;
500 int status = COPYBIT_SUCCESS;
501 uint32 gpuaddr = 0;
502 int c2d_format;
503 mapped_idx = -1;
504
505 if (flags & FLAGS_YUV_DESTINATION) {
506 c2d_format = get_c2d_format_for_yuv_destination(rhs->format);
507 } else {
508 c2d_format = get_format(rhs->format);
509 }
510
511 if(c2d_format == -EINVAL) {
512 ALOGE("%s: invalid format", __FUNCTION__);
513 return -EINVAL;
514 }
515
516 if(handle == NULL) {
517 ALOGE("%s: invalid handle", __func__);
518 return -EINVAL;
519 }
520
521 if (handle->gpuaddr == 0) {
522 gpuaddr = c2d_get_gpuaddr(ctx, handle, mapped_idx);
523 if(!gpuaddr) {
524 ALOGE("%s: c2d_get_gpuaddr failed", __FUNCTION__);
525 return COPYBIT_FAILURE;
526 }
527 } else {
528 gpuaddr = handle->gpuaddr;
529 }
530
531 /* create C2D surface */
532 if(is_supported_rgb_format(rhs->format) == COPYBIT_SUCCESS) {
533 /* RGB */
534 C2D_RGB_SURFACE_DEF surfaceDef;
535
536 surfaceType = (C2D_SURFACE_TYPE) (C2D_SURFACE_RGB_HOST | C2D_SURFACE_WITH_PHYS);
537
538 surfaceDef.phys = (void*) gpuaddr;
539 surfaceDef.buffer = (void*) (handle->base);
540
541 surfaceDef.format = c2d_format |
542 ((flags & FLAGS_PREMULTIPLIED_ALPHA) ? C2D_FORMAT_PREMULTIPLIED : 0);
543 surfaceDef.width = rhs->w;
544 surfaceDef.height = rhs->h;
545 int aligned_width = ALIGN(surfaceDef.width,32);
546 surfaceDef.stride = (aligned_width * c2diGetBpp(surfaceDef.format))>>3;
547
548 if(LINK_c2dUpdateSurface( surfaceId,C2D_TARGET | C2D_SOURCE, surfaceType,
549 &surfaceDef)) {
550 ALOGE("%s: RGB Surface c2dUpdateSurface ERROR", __FUNCTION__);
551 unmap_gpuaddr(ctx, mapped_idx);
552 status = COPYBIT_FAILURE;
553 }
554 } else if (is_supported_yuv_format(rhs->format) == COPYBIT_SUCCESS) {
555 C2D_YUV_SURFACE_DEF surfaceDef;
556 memset(&surfaceDef, 0, sizeof(surfaceDef));
557 surfaceType = (C2D_SURFACE_TYPE)(C2D_SURFACE_YUV_HOST | C2D_SURFACE_WITH_PHYS);
558 surfaceDef.format = c2d_format;
559
560 bufferInfo info;
561 info.width = rhs->w;
562 info.height = rhs->h;
563 info.format = rhs->format;
564
565 yuvPlaneInfo yuvInfo = {0};
566 status = calculate_yuv_offset_and_stride(info, yuvInfo);
567 if(status != COPYBIT_SUCCESS) {
568 ALOGE("%s: calculate_yuv_offset_and_stride error", __FUNCTION__);
569 unmap_gpuaddr(ctx, mapped_idx);
570 }
571
572 surfaceDef.width = rhs->w;
573 surfaceDef.height = rhs->h;
574 surfaceDef.plane0 = (void*) (handle->base);
575 surfaceDef.phys0 = (void*) (gpuaddr);
576 surfaceDef.stride0 = yuvInfo.yStride;
577
578 surfaceDef.plane1 = (void*) (handle->base + yuvInfo.plane1_offset);
579 surfaceDef.phys1 = (void*) (gpuaddr + yuvInfo.plane1_offset);
580 surfaceDef.stride1 = yuvInfo.plane1_stride;
581 if (3 == get_num_planes(rhs->format)) {
582 surfaceDef.plane2 = (void*) (handle->base + yuvInfo.plane2_offset);
583 surfaceDef.phys2 = (void*) (gpuaddr + yuvInfo.plane2_offset);
584 surfaceDef.stride2 = yuvInfo.plane2_stride;
585 }
586
587 if(LINK_c2dUpdateSurface( surfaceId,C2D_TARGET | C2D_SOURCE, surfaceType,
588 &surfaceDef)) {
589 ALOGE("%s: YUV Surface c2dUpdateSurface ERROR", __FUNCTION__);
590 unmap_gpuaddr(ctx, mapped_idx);
591 status = COPYBIT_FAILURE;
592 }
593 } else {
594 ALOGE("%s: invalid format 0x%x", __FUNCTION__, rhs->format);
595 unmap_gpuaddr(ctx, mapped_idx);
596 status = COPYBIT_FAILURE;
597 }
598
599 return status;
600 }
601
602 /** copy the bits */
msm_copybit(struct copybit_context_t * ctx,unsigned int target)603 static int msm_copybit(struct copybit_context_t *ctx, unsigned int target)
604 {
605 if (ctx->blit_count == 0) {
606 return COPYBIT_SUCCESS;
607 }
608
609 for (int i = 0; i < ctx->blit_count; i++)
610 {
611 ctx->blit_list[i].next = &(ctx->blit_list[i+1]);
612 }
613 ctx->blit_list[ctx->blit_count-1].next = NULL;
614 uint32_t target_transform = ctx->trg_transform;
615 if (ctx->c2d_driver_info.capabilities_mask &
616 C2D_DRIVER_SUPPORTS_OVERRIDE_TARGET_ROTATE_OP) {
617 // For A3xx - set 0x0 as the transform is set in the config_mask
618 target_transform = 0x0;
619 }
620 if(LINK_c2dDraw(target, target_transform, 0x0, 0, 0, ctx->blit_list,
621 ctx->blit_count)) {
622 ALOGE("%s: LINK_c2dDraw ERROR", __FUNCTION__);
623 return COPYBIT_FAILURE;
624 }
625 return COPYBIT_SUCCESS;
626 }
627
628
629
flush_get_fence_copybit(struct copybit_device_t * dev,int * fd)630 static int flush_get_fence_copybit (struct copybit_device_t *dev, int* fd)
631 {
632 struct copybit_context_t* ctx = (struct copybit_context_t*)dev;
633 int status = COPYBIT_FAILURE;
634 if (!ctx)
635 return COPYBIT_FAILURE;
636 pthread_mutex_lock(&ctx->wait_cleanup_lock);
637 status = msm_copybit(ctx, ctx->dst[ctx->dst_surface_type]);
638
639 if(LINK_c2dFlush(ctx->dst[ctx->dst_surface_type], &ctx->time_stamp)) {
640 ALOGE("%s: LINK_c2dFlush ERROR", __FUNCTION__);
641 // unlock the mutex and return failure
642 pthread_mutex_unlock(&ctx->wait_cleanup_lock);
643 return COPYBIT_FAILURE;
644 }
645 if(LINK_c2dCreateFenceFD(ctx->dst[ctx->dst_surface_type], ctx->time_stamp,
646 fd)) {
647 ALOGE("%s: LINK_c2dCreateFenceFD ERROR", __FUNCTION__);
648 status = COPYBIT_FAILURE;
649 }
650 if(status == COPYBIT_SUCCESS) {
651 //signal the wait_thread
652 ctx->wait_timestamp = true;
653 pthread_cond_signal(&ctx->wait_cleanup_cond);
654 }
655 pthread_mutex_unlock(&ctx->wait_cleanup_lock);
656 return status;
657 }
658
finish_copybit(struct copybit_device_t * dev)659 static int finish_copybit(struct copybit_device_t *dev)
660 {
661 struct copybit_context_t* ctx = (struct copybit_context_t*)dev;
662 if (!ctx)
663 return COPYBIT_FAILURE;
664
665 int status = msm_copybit(ctx, ctx->dst[ctx->dst_surface_type]);
666
667 if(LINK_c2dFinish(ctx->dst[ctx->dst_surface_type])) {
668 ALOGE("%s: LINK_c2dFinish ERROR", __FUNCTION__);
669 return COPYBIT_FAILURE;
670 }
671
672 // Unmap any mapped addresses.
673 for (int i = 0; i < MAX_SURFACES; i++) {
674 if (ctx->mapped_gpu_addr[i]) {
675 LINK_c2dUnMapAddr( (void*)ctx->mapped_gpu_addr[i]);
676 ctx->mapped_gpu_addr[i] = 0;
677 }
678 }
679
680 // Reset the counts after the draw.
681 ctx->blit_rgb_count = 0;
682 ctx->blit_yuv_2_plane_count = 0;
683 ctx->blit_yuv_3_plane_count = 0;
684 ctx->blit_count = 0;
685 ctx->dst_surface_mapped = false;
686 ctx->dst_surface_base = 0;
687
688 return status;
689 }
690
clear_copybit(struct copybit_device_t * dev,struct copybit_image_t const * buf,struct copybit_rect_t * rect)691 static int clear_copybit(struct copybit_device_t *dev,
692 struct copybit_image_t const *buf,
693 struct copybit_rect_t *rect)
694 {
695 int ret = COPYBIT_SUCCESS;
696 int flags = FLAGS_PREMULTIPLIED_ALPHA;
697 int mapped_dst_idx = -1;
698 struct copybit_context_t* ctx = (struct copybit_context_t*)dev;
699 C2D_RECT c2drect = {rect->l, rect->t, rect->r - rect->l, rect->b - rect->t};
700 pthread_mutex_lock(&ctx->wait_cleanup_lock);
701 if(!ctx->dst_surface_mapped) {
702 ret = set_image(ctx, ctx->dst[RGB_SURFACE], buf,
703 (eC2DFlags)flags, mapped_dst_idx);
704 if(ret) {
705 ALOGE("%s: set_image error", __FUNCTION__);
706 unmap_gpuaddr(ctx, mapped_dst_idx);
707 pthread_mutex_unlock(&ctx->wait_cleanup_lock);
708 return COPYBIT_FAILURE;
709 }
710 //clear_copybit is the first call made by HWC for each composition
711 //with the dest surface, hence set dst_surface_mapped.
712 ctx->dst_surface_mapped = true;
713 ctx->dst_surface_base = buf->base;
714 ret = LINK_c2dFillSurface(ctx->dst[RGB_SURFACE], 0x0, &c2drect);
715 }
716 pthread_mutex_unlock(&ctx->wait_cleanup_lock);
717 return ret;
718 }
719
720
721 /** setup rectangles */
set_rects(struct copybit_context_t * ctx,C2D_OBJECT * c2dObject,const struct copybit_rect_t * dst,const struct copybit_rect_t * src,const struct copybit_rect_t * scissor)722 static void set_rects(struct copybit_context_t *ctx,
723 C2D_OBJECT *c2dObject,
724 const struct copybit_rect_t *dst,
725 const struct copybit_rect_t *src,
726 const struct copybit_rect_t *scissor)
727 {
728 // Set the target rect.
729 if((ctx->trg_transform & C2D_TARGET_ROTATE_90) &&
730 (ctx->trg_transform & C2D_TARGET_ROTATE_180)) {
731 /* target rotation is 270 */
732 c2dObject->target_rect.x = (dst->t)<<16;
733 c2dObject->target_rect.y = ctx->fb_width?(ALIGN(ctx->fb_width,32)- dst->r):dst->r;
734 c2dObject->target_rect.y = c2dObject->target_rect.y<<16;
735 c2dObject->target_rect.height = ((dst->r) - (dst->l))<<16;
736 c2dObject->target_rect.width = ((dst->b) - (dst->t))<<16;
737 } else if(ctx->trg_transform & C2D_TARGET_ROTATE_90) {
738 c2dObject->target_rect.x = ctx->fb_height?(ctx->fb_height - dst->b):dst->b;
739 c2dObject->target_rect.x = c2dObject->target_rect.x<<16;
740 c2dObject->target_rect.y = (dst->l)<<16;
741 c2dObject->target_rect.height = ((dst->r) - (dst->l))<<16;
742 c2dObject->target_rect.width = ((dst->b) - (dst->t))<<16;
743 } else if(ctx->trg_transform & C2D_TARGET_ROTATE_180) {
744 c2dObject->target_rect.y = ctx->fb_height?(ctx->fb_height - dst->b):dst->b;
745 c2dObject->target_rect.y = c2dObject->target_rect.y<<16;
746 c2dObject->target_rect.x = ctx->fb_width?(ALIGN(ctx->fb_width,32) - dst->r):dst->r;
747 c2dObject->target_rect.x = c2dObject->target_rect.x<<16;
748 c2dObject->target_rect.height = ((dst->b) - (dst->t))<<16;
749 c2dObject->target_rect.width = ((dst->r) - (dst->l))<<16;
750 } else {
751 c2dObject->target_rect.x = (dst->l)<<16;
752 c2dObject->target_rect.y = (dst->t)<<16;
753 c2dObject->target_rect.height = ((dst->b) - (dst->t))<<16;
754 c2dObject->target_rect.width = ((dst->r) - (dst->l))<<16;
755 }
756 c2dObject->config_mask |= C2D_TARGET_RECT_BIT;
757
758 // Set the source rect
759 c2dObject->source_rect.x = (src->l)<<16;
760 c2dObject->source_rect.y = (src->t)<<16;
761 c2dObject->source_rect.height = ((src->b) - (src->t))<<16;
762 c2dObject->source_rect.width = ((src->r) - (src->l))<<16;
763 c2dObject->config_mask |= C2D_SOURCE_RECT_BIT;
764
765 // Set the scissor rect
766 c2dObject->scissor_rect.x = scissor->l;
767 c2dObject->scissor_rect.y = scissor->t;
768 c2dObject->scissor_rect.height = (scissor->b) - (scissor->t);
769 c2dObject->scissor_rect.width = (scissor->r) - (scissor->l);
770 c2dObject->config_mask |= C2D_SCISSOR_RECT_BIT;
771 }
772
773 /*****************************************************************************/
774
775 /** Set a parameter to value */
set_parameter_copybit(struct copybit_device_t * dev,int name,int value)776 static int set_parameter_copybit(
777 struct copybit_device_t *dev,
778 int name,
779 int value)
780 {
781 struct copybit_context_t* ctx = (struct copybit_context_t*)dev;
782 int status = COPYBIT_SUCCESS;
783 if (!ctx) {
784 ALOGE("%s: null context", __FUNCTION__);
785 return -EINVAL;
786 }
787
788 pthread_mutex_lock(&ctx->wait_cleanup_lock);
789 switch(name) {
790 case COPYBIT_PLANE_ALPHA:
791 {
792 if (value < 0) value = 0;
793 if (value >= 256) value = 255;
794
795 ctx->src_global_alpha = value;
796 if (value < 255)
797 ctx->config_mask |= C2D_GLOBAL_ALPHA_BIT;
798 else
799 ctx->config_mask &= ~C2D_GLOBAL_ALPHA_BIT;
800 }
801 break;
802 case COPYBIT_BLEND_MODE:
803 {
804 if (value == COPYBIT_BLENDING_NONE) {
805 ctx->config_mask |= C2D_ALPHA_BLEND_NONE;
806 ctx->is_premultiplied_alpha = true;
807 } else if (value == COPYBIT_BLENDING_PREMULT) {
808 ctx->is_premultiplied_alpha = true;
809 } else {
810 ctx->config_mask &= ~C2D_ALPHA_BLEND_NONE;
811 }
812 }
813 break;
814 case COPYBIT_TRANSFORM:
815 {
816 unsigned int transform = 0;
817 uint32 config_mask = 0;
818 config_mask |= C2D_OVERRIDE_GLOBAL_TARGET_ROTATE_CONFIG;
819 if((value & 0x7) == COPYBIT_TRANSFORM_ROT_180) {
820 transform = C2D_TARGET_ROTATE_180;
821 config_mask |= C2D_OVERRIDE_TARGET_ROTATE_180;
822 } else if((value & 0x7) == COPYBIT_TRANSFORM_ROT_270) {
823 transform = C2D_TARGET_ROTATE_90;
824 config_mask |= C2D_OVERRIDE_TARGET_ROTATE_90;
825 } else if(value == COPYBIT_TRANSFORM_ROT_90) {
826 transform = C2D_TARGET_ROTATE_270;
827 config_mask |= C2D_OVERRIDE_TARGET_ROTATE_270;
828 } else {
829 config_mask |= C2D_OVERRIDE_TARGET_ROTATE_0;
830 if(value & COPYBIT_TRANSFORM_FLIP_H) {
831 config_mask |= C2D_MIRROR_H_BIT;
832 } else if(value & COPYBIT_TRANSFORM_FLIP_V) {
833 config_mask |= C2D_MIRROR_V_BIT;
834 }
835 }
836
837 if (ctx->c2d_driver_info.capabilities_mask &
838 C2D_DRIVER_SUPPORTS_OVERRIDE_TARGET_ROTATE_OP) {
839 ctx->config_mask |= config_mask;
840 } else {
841 // The transform for this surface does not match the current
842 // target transform. Draw all previous surfaces. This will be
843 // changed once we have a new mechanism to send different
844 // target rotations to c2d.
845 finish_copybit(dev);
846 }
847 ctx->trg_transform = transform;
848 }
849 break;
850 case COPYBIT_FRAMEBUFFER_WIDTH:
851 ctx->fb_width = value;
852 break;
853 case COPYBIT_FRAMEBUFFER_HEIGHT:
854 ctx->fb_height = value;
855 break;
856 case COPYBIT_ROTATION_DEG:
857 case COPYBIT_DITHER:
858 case COPYBIT_BLUR:
859 case COPYBIT_BLIT_TO_FRAMEBUFFER:
860 // Do nothing
861 break;
862 default:
863 ALOGE("%s: default case param=0x%x", __FUNCTION__, name);
864 status = -EINVAL;
865 break;
866 }
867 pthread_mutex_unlock(&ctx->wait_cleanup_lock);
868 return status;
869 }
870
871 /** Get a static info value */
get(struct copybit_device_t * dev,int name)872 static int get(struct copybit_device_t *dev, int name)
873 {
874 struct copybit_context_t* ctx = (struct copybit_context_t*)dev;
875 int value;
876
877 if (!ctx) {
878 ALOGE("%s: null context error", __FUNCTION__);
879 return -EINVAL;
880 }
881
882 switch(name) {
883 case COPYBIT_MINIFICATION_LIMIT:
884 value = MAX_SCALE_FACTOR;
885 break;
886 case COPYBIT_MAGNIFICATION_LIMIT:
887 value = MAX_SCALE_FACTOR;
888 break;
889 case COPYBIT_SCALING_FRAC_BITS:
890 value = 32;
891 break;
892 case COPYBIT_ROTATION_STEP_DEG:
893 value = 1;
894 break;
895 default:
896 ALOGE("%s: default case param=0x%x", __FUNCTION__, name);
897 value = -EINVAL;
898 }
899 return value;
900 }
901
is_alpha(int cformat)902 static int is_alpha(int cformat)
903 {
904 int alpha = 0;
905 switch (cformat & 0xFF) {
906 case C2D_COLOR_FORMAT_8888_ARGB:
907 case C2D_COLOR_FORMAT_8888_RGBA:
908 case C2D_COLOR_FORMAT_5551_RGBA:
909 case C2D_COLOR_FORMAT_4444_ARGB:
910 alpha = 1;
911 break;
912 default:
913 alpha = 0;
914 break;
915 }
916
917 if(alpha && (cformat&C2D_FORMAT_DISABLE_ALPHA))
918 alpha = 0;
919
920 return alpha;
921 }
922
923 /* Function to check if we need a temporary buffer for the blit.
924 * This would happen if the requested destination stride and the
925 * C2D stride do not match. We ignore RGB buffers, since their
926 * stride is always aligned to 32.
927 */
need_temp_buffer(struct copybit_image_t const * img)928 static bool need_temp_buffer(struct copybit_image_t const *img)
929 {
930 if (COPYBIT_SUCCESS == is_supported_rgb_format(img->format))
931 return false;
932
933 struct private_handle_t* handle = (struct private_handle_t*)img->handle;
934
935 // The width parameter in the handle contains the aligned_w. We check if we
936 // need to convert based on this param. YUV formats have bpp=1, so checking
937 // if the requested stride is aligned should suffice.
938 if (0 == (handle->width)%32) {
939 return false;
940 }
941
942 return true;
943 }
944
945 /* Function to extract the information from the copybit image and set the corresponding
946 * values in the bufferInfo struct.
947 */
populate_buffer_info(struct copybit_image_t const * img,bufferInfo & info)948 static void populate_buffer_info(struct copybit_image_t const *img, bufferInfo& info)
949 {
950 info.width = img->w;
951 info.height = img->h;
952 info.format = img->format;
953 }
954
955 /* Function to get the required size for a particular format, inorder for C2D to perform
956 * the blit operation.
957 */
get_size(const bufferInfo & info)958 static size_t get_size(const bufferInfo& info)
959 {
960 size_t size = 0;
961 int w = info.width;
962 int h = info.height;
963 int aligned_w = ALIGN(w, 32);
964 switch(info.format) {
965 case HAL_PIXEL_FORMAT_NV12_ENCODEABLE:
966 {
967 // Chroma for this format is aligned to 2K.
968 size = ALIGN((aligned_w*h), 2048) +
969 ALIGN(aligned_w/2, 32) * (h/2) *2;
970 size = ALIGN(size, 4096);
971 } break;
972 case HAL_PIXEL_FORMAT_YCbCr_420_SP:
973 case HAL_PIXEL_FORMAT_YCrCb_420_SP:
974 {
975 size = aligned_w * h +
976 ALIGN(aligned_w/2, 32) * (h/2) * 2;
977 size = ALIGN(size, 4096);
978 } break;
979 default: break;
980 }
981 return size;
982 }
983
984 /* Function to allocate memory for the temporary buffer. This memory is
985 * allocated from Ashmem. It is the caller's responsibility to free this
986 * memory.
987 */
get_temp_buffer(const bufferInfo & info,alloc_data & data)988 static int get_temp_buffer(const bufferInfo& info, alloc_data& data)
989 {
990 ALOGD("%s E", __FUNCTION__);
991 // Alloc memory from system heap
992 data.base = 0;
993 data.fd = -1;
994 data.offset = 0;
995 data.size = get_size(info);
996 data.align = getpagesize();
997 data.uncached = true;
998 int allocFlags = GRALLOC_USAGE_PRIVATE_SYSTEM_HEAP;
999
1000 if (sAlloc == 0) {
1001 sAlloc = gralloc::IAllocController::getInstance();
1002 }
1003
1004 if (sAlloc == 0) {
1005 ALOGE("%s: sAlloc is still NULL", __FUNCTION__);
1006 return COPYBIT_FAILURE;
1007 }
1008
1009 int err = sAlloc->allocate(data, allocFlags);
1010 if (0 != err) {
1011 ALOGE("%s: allocate failed", __FUNCTION__);
1012 return COPYBIT_FAILURE;
1013 }
1014
1015 ALOGD("%s X", __FUNCTION__);
1016 return err;
1017 }
1018
1019 /* Function to free the temporary allocated memory.*/
free_temp_buffer(alloc_data & data)1020 static void free_temp_buffer(alloc_data &data)
1021 {
1022 if (-1 != data.fd) {
1023 IMemAlloc* memalloc = sAlloc->getAllocator(data.allocType);
1024 memalloc->free_buffer(data.base, data.size, 0, data.fd);
1025 }
1026 }
1027
1028 /* Function to perform the software color conversion. Convert the
1029 * C2D compatible format to the Android compatible format
1030 */
copy_image(private_handle_t * src_handle,struct copybit_image_t const * rhs,eConversionType conversionType)1031 static int copy_image(private_handle_t *src_handle,
1032 struct copybit_image_t const *rhs,
1033 eConversionType conversionType)
1034 {
1035 if (src_handle->fd == -1) {
1036 ALOGE("%s: src_handle fd is invalid", __FUNCTION__);
1037 return COPYBIT_FAILURE;
1038 }
1039
1040 // Copy the info.
1041 int ret = COPYBIT_SUCCESS;
1042 switch(rhs->format) {
1043 case HAL_PIXEL_FORMAT_NV12_ENCODEABLE:
1044 case HAL_PIXEL_FORMAT_YCbCr_420_SP:
1045 case HAL_PIXEL_FORMAT_YCrCb_420_SP:
1046 {
1047 if (CONVERT_TO_ANDROID_FORMAT == conversionType) {
1048 return convert_yuv_c2d_to_yuv_android(src_handle, rhs);
1049 } else {
1050 return convert_yuv_android_to_yuv_c2d(src_handle, rhs);
1051 }
1052
1053 } break;
1054 default: {
1055 ALOGE("%s: invalid format 0x%x", __FUNCTION__, rhs->format);
1056 ret = COPYBIT_FAILURE;
1057 } break;
1058 }
1059 return ret;
1060 }
1061
delete_handle(private_handle_t * handle)1062 static void delete_handle(private_handle_t *handle)
1063 {
1064 if (handle) {
1065 delete handle;
1066 handle = 0;
1067 }
1068 }
1069
need_to_execute_draw(struct copybit_context_t * ctx,eC2DFlags flags)1070 static bool need_to_execute_draw(struct copybit_context_t* ctx,
1071 eC2DFlags flags)
1072 {
1073 if (flags & FLAGS_TEMP_SRC_DST) {
1074 return true;
1075 }
1076 if (flags & FLAGS_YUV_DESTINATION) {
1077 return true;
1078 }
1079 return false;
1080 }
1081
1082 /** do a stretch blit type operation */
stretch_copybit_internal(struct copybit_device_t * dev,struct copybit_image_t const * dst,struct copybit_image_t const * src,struct copybit_rect_t const * dst_rect,struct copybit_rect_t const * src_rect,struct copybit_region_t const * region,bool enableBlend)1083 static int stretch_copybit_internal(
1084 struct copybit_device_t *dev,
1085 struct copybit_image_t const *dst,
1086 struct copybit_image_t const *src,
1087 struct copybit_rect_t const *dst_rect,
1088 struct copybit_rect_t const *src_rect,
1089 struct copybit_region_t const *region,
1090 bool enableBlend)
1091 {
1092 struct copybit_context_t* ctx = (struct copybit_context_t*)dev;
1093 int status = COPYBIT_SUCCESS;
1094 int flags = 0;
1095 int src_surface_type;
1096 int mapped_src_idx = -1, mapped_dst_idx = -1;
1097 C2D_OBJECT_STR src_surface;
1098
1099 if (!ctx) {
1100 ALOGE("%s: null context error", __FUNCTION__);
1101 return -EINVAL;
1102 }
1103
1104 if (src->w > MAX_DIMENSION || src->h > MAX_DIMENSION) {
1105 ALOGE("%s: src dimension error", __FUNCTION__);
1106 return -EINVAL;
1107 }
1108
1109 if (dst->w > MAX_DIMENSION || dst->h > MAX_DIMENSION) {
1110 ALOGE("%s : dst dimension error dst w %d h %d", __FUNCTION__, dst->w,
1111 dst->h);
1112 return -EINVAL;
1113 }
1114
1115 if (is_valid_destination_format(dst->format) == COPYBIT_FAILURE) {
1116 ALOGE("%s: Invalid destination format format = 0x%x", __FUNCTION__,
1117 dst->format);
1118 return COPYBIT_FAILURE;
1119 }
1120
1121 int dst_surface_type;
1122 if (is_supported_rgb_format(dst->format) == COPYBIT_SUCCESS) {
1123 dst_surface_type = RGB_SURFACE;
1124 flags |= FLAGS_PREMULTIPLIED_ALPHA;
1125 } else if (is_supported_yuv_format(dst->format) == COPYBIT_SUCCESS) {
1126 int num_planes = get_num_planes(dst->format);
1127 flags |= FLAGS_YUV_DESTINATION;
1128 if (num_planes == 2) {
1129 dst_surface_type = YUV_SURFACE_2_PLANES;
1130 } else if (num_planes == 3) {
1131 dst_surface_type = YUV_SURFACE_3_PLANES;
1132 } else {
1133 ALOGE("%s: dst number of YUV planes is invalid dst format = 0x%x",
1134 __FUNCTION__, dst->format);
1135 return COPYBIT_FAILURE;
1136 }
1137 } else {
1138 ALOGE("%s: Invalid dst surface format 0x%x", __FUNCTION__,
1139 dst->format);
1140 return COPYBIT_FAILURE;
1141 }
1142
1143 if (ctx->blit_rgb_count == MAX_RGB_SURFACES ||
1144 ctx->blit_yuv_2_plane_count == MAX_YUV_2_PLANE_SURFACES ||
1145 ctx->blit_yuv_3_plane_count == MAX_YUV_2_PLANE_SURFACES ||
1146 ctx->blit_count == MAX_BLIT_OBJECT_COUNT ||
1147 ctx->dst_surface_type != dst_surface_type) {
1148 // we have reached the max. limits of our internal structures or
1149 // changed the target.
1150 // Draw the remaining surfaces. We need to do the finish here since
1151 // we need to free up the surface templates.
1152 finish_copybit(dev);
1153 }
1154
1155 ctx->dst_surface_type = dst_surface_type;
1156
1157 // Update the destination
1158 copybit_image_t dst_image;
1159 dst_image.w = dst->w;
1160 dst_image.h = dst->h;
1161 dst_image.format = dst->format;
1162 dst_image.handle = dst->handle;
1163 // Check if we need a temp. copy for the destination. We'd need this the destination
1164 // width is not aligned to 32. This case occurs for YUV formats. RGB formats are
1165 // aligned to 32.
1166 bool need_temp_dst = need_temp_buffer(dst);
1167 bufferInfo dst_info;
1168 populate_buffer_info(dst, dst_info);
1169 private_handle_t* dst_hnd = new private_handle_t(-1, 0, 0, 0, dst_info.format,
1170 dst_info.width, dst_info.height);
1171 if (dst_hnd == NULL) {
1172 ALOGE("%s: dst_hnd is null", __FUNCTION__);
1173 return COPYBIT_FAILURE;
1174 }
1175 if (need_temp_dst) {
1176 if (get_size(dst_info) != ctx->temp_dst_buffer.size) {
1177 free_temp_buffer(ctx->temp_dst_buffer);
1178 // Create a temp buffer and set that as the destination.
1179 if (COPYBIT_FAILURE == get_temp_buffer(dst_info, ctx->temp_dst_buffer)) {
1180 ALOGE("%s: get_temp_buffer(dst) failed", __FUNCTION__);
1181 delete_handle(dst_hnd);
1182 return COPYBIT_FAILURE;
1183 }
1184 }
1185 dst_hnd->fd = ctx->temp_dst_buffer.fd;
1186 dst_hnd->size = ctx->temp_dst_buffer.size;
1187 dst_hnd->flags = ctx->temp_dst_buffer.allocType;
1188 dst_hnd->base = (int)(ctx->temp_dst_buffer.base);
1189 dst_hnd->offset = ctx->temp_dst_buffer.offset;
1190 dst_hnd->gpuaddr = 0;
1191 dst_image.handle = dst_hnd;
1192 }
1193 if(!ctx->dst_surface_mapped) {
1194 //map the destination surface to GPU address
1195 status = set_image(ctx, ctx->dst[ctx->dst_surface_type], &dst_image,
1196 (eC2DFlags)flags, mapped_dst_idx);
1197 if(status) {
1198 ALOGE("%s: dst: set_image error", __FUNCTION__);
1199 delete_handle(dst_hnd);
1200 unmap_gpuaddr(ctx, mapped_dst_idx);
1201 return COPYBIT_FAILURE;
1202 }
1203 ctx->dst_surface_mapped = true;
1204 ctx->dst_surface_base = dst->base;
1205 } else if(ctx->dst_surface_mapped && ctx->dst_surface_base != dst->base) {
1206 // Destination surface for the operation should be same for multiple
1207 // requests, this check is catch if there is any case when the
1208 // destination changes
1209 ALOGE("%s: a different destination surface!!", __FUNCTION__);
1210 }
1211
1212 // Update the source
1213 flags = 0;
1214 if(is_supported_rgb_format(src->format) == COPYBIT_SUCCESS) {
1215 src_surface_type = RGB_SURFACE;
1216 src_surface = ctx->blit_rgb_object[ctx->blit_rgb_count];
1217 } else if (is_supported_yuv_format(src->format) == COPYBIT_SUCCESS) {
1218 int num_planes = get_num_planes(src->format);
1219 if (num_planes == 2) {
1220 src_surface_type = YUV_SURFACE_2_PLANES;
1221 src_surface = ctx->blit_yuv_2_plane_object[ctx->blit_yuv_2_plane_count];
1222 } else if (num_planes == 3) {
1223 src_surface_type = YUV_SURFACE_3_PLANES;
1224 src_surface = ctx->blit_yuv_3_plane_object[ctx->blit_yuv_2_plane_count];
1225 } else {
1226 ALOGE("%s: src number of YUV planes is invalid src format = 0x%x",
1227 __FUNCTION__, src->format);
1228 delete_handle(dst_hnd);
1229 unmap_gpuaddr(ctx, mapped_dst_idx);
1230 return -EINVAL;
1231 }
1232 } else {
1233 ALOGE("%s: Invalid source surface format 0x%x", __FUNCTION__,
1234 src->format);
1235 delete_handle(dst_hnd);
1236 unmap_gpuaddr(ctx, mapped_dst_idx);
1237 return -EINVAL;
1238 }
1239
1240 copybit_image_t src_image;
1241 src_image.w = src->w;
1242 src_image.h = src->h;
1243 src_image.format = src->format;
1244 src_image.handle = src->handle;
1245
1246 bool need_temp_src = need_temp_buffer(src);
1247 bufferInfo src_info;
1248 populate_buffer_info(src, src_info);
1249 private_handle_t* src_hnd = new private_handle_t(-1, 0, 0, 0, src_info.format,
1250 src_info.width, src_info.height);
1251 if (NULL == src_hnd) {
1252 ALOGE("%s: src_hnd is null", __FUNCTION__);
1253 delete_handle(dst_hnd);
1254 unmap_gpuaddr(ctx, mapped_dst_idx);
1255 return COPYBIT_FAILURE;
1256 }
1257 if (need_temp_src) {
1258 if (get_size(src_info) != ctx->temp_src_buffer.size) {
1259 free_temp_buffer(ctx->temp_src_buffer);
1260 // Create a temp buffer and set that as the destination.
1261 if (COPYBIT_SUCCESS != get_temp_buffer(src_info,
1262 ctx->temp_src_buffer)) {
1263 ALOGE("%s: get_temp_buffer(src) failed", __FUNCTION__);
1264 delete_handle(dst_hnd);
1265 delete_handle(src_hnd);
1266 unmap_gpuaddr(ctx, mapped_dst_idx);
1267 return COPYBIT_FAILURE;
1268 }
1269 }
1270 src_hnd->fd = ctx->temp_src_buffer.fd;
1271 src_hnd->size = ctx->temp_src_buffer.size;
1272 src_hnd->flags = ctx->temp_src_buffer.allocType;
1273 src_hnd->base = (int)(ctx->temp_src_buffer.base);
1274 src_hnd->offset = ctx->temp_src_buffer.offset;
1275 src_hnd->gpuaddr = 0;
1276 src_image.handle = src_hnd;
1277
1278 // Copy the source.
1279 status = copy_image((private_handle_t *)src->handle, &src_image,
1280 CONVERT_TO_C2D_FORMAT);
1281 if (status == COPYBIT_FAILURE) {
1282 ALOGE("%s:copy_image failed in temp source",__FUNCTION__);
1283 delete_handle(dst_hnd);
1284 delete_handle(src_hnd);
1285 unmap_gpuaddr(ctx, mapped_dst_idx);
1286 return status;
1287 }
1288
1289 // Clean the cache
1290 IMemAlloc* memalloc = sAlloc->getAllocator(src_hnd->flags);
1291 if (memalloc->clean_buffer((void *)(src_hnd->base), src_hnd->size,
1292 src_hnd->offset, src_hnd->fd,
1293 gralloc::CACHE_CLEAN)) {
1294 ALOGE("%s: clean_buffer failed", __FUNCTION__);
1295 delete_handle(dst_hnd);
1296 delete_handle(src_hnd);
1297 unmap_gpuaddr(ctx, mapped_dst_idx);
1298 return COPYBIT_FAILURE;
1299 }
1300 }
1301
1302 flags |= (ctx->is_premultiplied_alpha) ? FLAGS_PREMULTIPLIED_ALPHA : 0;
1303 flags |= (ctx->dst_surface_type != RGB_SURFACE) ? FLAGS_YUV_DESTINATION : 0;
1304 status = set_image(ctx, src_surface.surface_id, &src_image,
1305 (eC2DFlags)flags, mapped_src_idx);
1306 if(status) {
1307 ALOGE("%s: set_image (src) error", __FUNCTION__);
1308 delete_handle(dst_hnd);
1309 delete_handle(src_hnd);
1310 unmap_gpuaddr(ctx, mapped_dst_idx);
1311 unmap_gpuaddr(ctx, mapped_src_idx);
1312 return COPYBIT_FAILURE;
1313 }
1314
1315 src_surface.config_mask = C2D_NO_ANTIALIASING_BIT | ctx->config_mask;
1316 src_surface.global_alpha = ctx->src_global_alpha;
1317 if (enableBlend) {
1318 if(src_surface.config_mask & C2D_GLOBAL_ALPHA_BIT) {
1319 src_surface.config_mask &= ~C2D_ALPHA_BLEND_NONE;
1320 if(!(src_surface.global_alpha)) {
1321 // src alpha is zero
1322 delete_handle(dst_hnd);
1323 delete_handle(src_hnd);
1324 unmap_gpuaddr(ctx, mapped_dst_idx);
1325 unmap_gpuaddr(ctx, mapped_src_idx);
1326 return COPYBIT_FAILURE;
1327 }
1328 }
1329 } else {
1330 src_surface.config_mask |= C2D_ALPHA_BLEND_NONE;
1331 }
1332
1333 if (src_surface_type == RGB_SURFACE) {
1334 ctx->blit_rgb_object[ctx->blit_rgb_count] = src_surface;
1335 ctx->blit_rgb_count++;
1336 } else if (src_surface_type == YUV_SURFACE_2_PLANES) {
1337 ctx->blit_yuv_2_plane_object[ctx->blit_yuv_2_plane_count] = src_surface;
1338 ctx->blit_yuv_2_plane_count++;
1339 } else {
1340 ctx->blit_yuv_3_plane_object[ctx->blit_yuv_3_plane_count] = src_surface;
1341 ctx->blit_yuv_3_plane_count++;
1342 }
1343
1344 struct copybit_rect_t clip;
1345 while ((status == 0) && region->next(region, &clip)) {
1346 set_rects(ctx, &(src_surface), dst_rect, src_rect, &clip);
1347 if (ctx->blit_count == MAX_BLIT_OBJECT_COUNT) {
1348 ALOGW("Reached end of blit count");
1349 finish_copybit(dev);
1350 }
1351 ctx->blit_list[ctx->blit_count] = src_surface;
1352 ctx->blit_count++;
1353 }
1354
1355 // Check if we need to perform an early draw-finish.
1356 flags |= (need_temp_dst || need_temp_src) ? FLAGS_TEMP_SRC_DST : 0;
1357 if (need_to_execute_draw(ctx, (eC2DFlags)flags))
1358 {
1359 finish_copybit(dev);
1360 }
1361
1362 if (need_temp_dst) {
1363 // copy the temp. destination without the alignment to the actual
1364 // destination.
1365 status = copy_image(dst_hnd, dst, CONVERT_TO_ANDROID_FORMAT);
1366 if (status == COPYBIT_FAILURE) {
1367 ALOGE("%s:copy_image failed in temp Dest",__FUNCTION__);
1368 delete_handle(dst_hnd);
1369 delete_handle(src_hnd);
1370 unmap_gpuaddr(ctx, mapped_dst_idx);
1371 unmap_gpuaddr(ctx, mapped_src_idx);
1372 return status;
1373 }
1374 // Clean the cache.
1375 IMemAlloc* memalloc = sAlloc->getAllocator(dst_hnd->flags);
1376 memalloc->clean_buffer((void *)(dst_hnd->base), dst_hnd->size,
1377 dst_hnd->offset, dst_hnd->fd,
1378 gralloc::CACHE_CLEAN);
1379 }
1380 delete_handle(dst_hnd);
1381 delete_handle(src_hnd);
1382
1383 ctx->is_premultiplied_alpha = false;
1384 ctx->fb_width = 0;
1385 ctx->fb_height = 0;
1386 ctx->config_mask = 0;
1387 return status;
1388 }
1389
set_sync_copybit(struct copybit_device_t * dev,int acquireFenceFd)1390 static int set_sync_copybit(struct copybit_device_t *dev,
1391 int acquireFenceFd)
1392 {
1393 return 0;
1394 }
1395
stretch_copybit(struct copybit_device_t * dev,struct copybit_image_t const * dst,struct copybit_image_t const * src,struct copybit_rect_t const * dst_rect,struct copybit_rect_t const * src_rect,struct copybit_region_t const * region)1396 static int stretch_copybit(
1397 struct copybit_device_t *dev,
1398 struct copybit_image_t const *dst,
1399 struct copybit_image_t const *src,
1400 struct copybit_rect_t const *dst_rect,
1401 struct copybit_rect_t const *src_rect,
1402 struct copybit_region_t const *region)
1403 {
1404 struct copybit_context_t* ctx = (struct copybit_context_t*)dev;
1405 int status = COPYBIT_SUCCESS;
1406 bool needsBlending = (ctx->src_global_alpha != 0);
1407 pthread_mutex_lock(&ctx->wait_cleanup_lock);
1408 status = stretch_copybit_internal(dev, dst, src, dst_rect, src_rect,
1409 region, needsBlending);
1410 pthread_mutex_unlock(&ctx->wait_cleanup_lock);
1411 return status;
1412 }
1413
1414 /** Perform a blit type operation */
blit_copybit(struct copybit_device_t * dev,struct copybit_image_t const * dst,struct copybit_image_t const * src,struct copybit_region_t const * region)1415 static int blit_copybit(
1416 struct copybit_device_t *dev,
1417 struct copybit_image_t const *dst,
1418 struct copybit_image_t const *src,
1419 struct copybit_region_t const *region)
1420 {
1421 int status = COPYBIT_SUCCESS;
1422 struct copybit_context_t* ctx = (struct copybit_context_t*)dev;
1423 struct copybit_rect_t dr = { 0, 0, (int)dst->w, (int)dst->h };
1424 struct copybit_rect_t sr = { 0, 0, (int)src->w, (int)src->h };
1425 pthread_mutex_lock(&ctx->wait_cleanup_lock);
1426 status = stretch_copybit_internal(dev, dst, src, &dr, &sr, region, false);
1427 pthread_mutex_unlock(&ctx->wait_cleanup_lock);
1428 return status;
1429 }
1430
1431 /*****************************************************************************/
1432
clean_up(copybit_context_t * ctx)1433 static void clean_up(copybit_context_t* ctx)
1434 {
1435 void* ret;
1436 if (!ctx)
1437 return;
1438
1439 // stop the wait_cleanup_thread
1440 pthread_mutex_lock(&ctx->wait_cleanup_lock);
1441 ctx->stop_thread = true;
1442 // Signal waiting thread
1443 pthread_cond_signal(&ctx->wait_cleanup_cond);
1444 pthread_mutex_unlock(&ctx->wait_cleanup_lock);
1445 // waits for the cleanup thread to exit
1446 pthread_join(ctx->wait_thread_id, &ret);
1447 pthread_mutex_destroy(&ctx->wait_cleanup_lock);
1448 pthread_cond_destroy (&ctx->wait_cleanup_cond);
1449
1450 for (int i = 0; i < NUM_SURFACE_TYPES; i++) {
1451 if (ctx->dst[i])
1452 LINK_c2dDestroySurface(ctx->dst[i]);
1453 }
1454
1455 for (int i = 0; i < MAX_RGB_SURFACES; i++) {
1456 if (ctx->blit_rgb_object[i].surface_id)
1457 LINK_c2dDestroySurface(ctx->blit_rgb_object[i].surface_id);
1458 }
1459
1460 for (int i = 0; i < MAX_YUV_2_PLANE_SURFACES; i++) {
1461 if (ctx->blit_yuv_2_plane_object[i].surface_id)
1462 LINK_c2dDestroySurface(ctx->blit_yuv_2_plane_object[i].surface_id);
1463 }
1464
1465 for (int i = 0; i < MAX_YUV_3_PLANE_SURFACES; i++) {
1466 if (ctx->blit_yuv_3_plane_object[i].surface_id)
1467 LINK_c2dDestroySurface(ctx->blit_yuv_3_plane_object[i].surface_id);
1468 }
1469
1470 if (ctx->libc2d2) {
1471 ::dlclose(ctx->libc2d2);
1472 ALOGV("dlclose(libc2d2)");
1473 }
1474
1475 free(ctx);
1476 }
1477
1478 /** Close the copybit device */
close_copybit(struct hw_device_t * dev)1479 static int close_copybit(struct hw_device_t *dev)
1480 {
1481 struct copybit_context_t* ctx = (struct copybit_context_t*)dev;
1482 if (ctx) {
1483 free_temp_buffer(ctx->temp_src_buffer);
1484 free_temp_buffer(ctx->temp_dst_buffer);
1485 }
1486 clean_up(ctx);
1487 return 0;
1488 }
1489
1490 /** Open a new instance of a copybit device using name */
open_copybit(const struct hw_module_t * module,const char * name,struct hw_device_t ** device)1491 static int open_copybit(const struct hw_module_t* module, const char* name,
1492 struct hw_device_t** device)
1493 {
1494 int status = COPYBIT_SUCCESS;
1495 C2D_RGB_SURFACE_DEF surfDefinition = {0};
1496 C2D_YUV_SURFACE_DEF yuvSurfaceDef = {0} ;
1497 struct copybit_context_t *ctx;
1498 char fbName[64];
1499
1500 ctx = (struct copybit_context_t *)malloc(sizeof(struct copybit_context_t));
1501 if(!ctx) {
1502 ALOGE("%s: malloc failed", __FUNCTION__);
1503 return COPYBIT_FAILURE;
1504 }
1505
1506 /* initialize drawstate */
1507 memset(ctx, 0, sizeof(*ctx));
1508 ctx->libc2d2 = ::dlopen("libC2D2.so", RTLD_NOW);
1509 if (!ctx->libc2d2) {
1510 ALOGE("FATAL ERROR: could not dlopen libc2d2.so: %s", dlerror());
1511 clean_up(ctx);
1512 status = COPYBIT_FAILURE;
1513 *device = NULL;
1514 return status;
1515 }
1516 *(void **)&LINK_c2dCreateSurface = ::dlsym(ctx->libc2d2,
1517 "c2dCreateSurface");
1518 *(void **)&LINK_c2dUpdateSurface = ::dlsym(ctx->libc2d2,
1519 "c2dUpdateSurface");
1520 *(void **)&LINK_c2dReadSurface = ::dlsym(ctx->libc2d2,
1521 "c2dReadSurface");
1522 *(void **)&LINK_c2dDraw = ::dlsym(ctx->libc2d2, "c2dDraw");
1523 *(void **)&LINK_c2dFlush = ::dlsym(ctx->libc2d2, "c2dFlush");
1524 *(void **)&LINK_c2dFinish = ::dlsym(ctx->libc2d2, "c2dFinish");
1525 *(void **)&LINK_c2dWaitTimestamp = ::dlsym(ctx->libc2d2,
1526 "c2dWaitTimestamp");
1527 *(void **)&LINK_c2dDestroySurface = ::dlsym(ctx->libc2d2,
1528 "c2dDestroySurface");
1529 *(void **)&LINK_c2dMapAddr = ::dlsym(ctx->libc2d2,
1530 "c2dMapAddr");
1531 *(void **)&LINK_c2dUnMapAddr = ::dlsym(ctx->libc2d2,
1532 "c2dUnMapAddr");
1533 *(void **)&LINK_c2dGetDriverCapabilities = ::dlsym(ctx->libc2d2,
1534 "c2dGetDriverCapabilities");
1535 *(void **)&LINK_c2dCreateFenceFD = ::dlsym(ctx->libc2d2,
1536 "c2dCreateFenceFD");
1537 *(void **)&LINK_c2dFillSurface = ::dlsym(ctx->libc2d2,
1538 "c2dFillSurface");
1539
1540 if (!LINK_c2dCreateSurface || !LINK_c2dUpdateSurface || !LINK_c2dReadSurface
1541 || !LINK_c2dDraw || !LINK_c2dFlush || !LINK_c2dWaitTimestamp ||
1542 !LINK_c2dFinish || !LINK_c2dDestroySurface ||
1543 !LINK_c2dGetDriverCapabilities || !LINK_c2dCreateFenceFD ||
1544 !LINK_c2dFillSurface) {
1545 ALOGE("%s: dlsym ERROR", __FUNCTION__);
1546 clean_up(ctx);
1547 status = COPYBIT_FAILURE;
1548 *device = NULL;
1549 return status;
1550 }
1551
1552 ctx->device.common.tag = HARDWARE_DEVICE_TAG;
1553 ctx->device.common.version = 1;
1554 ctx->device.common.module = (hw_module_t*)(module);
1555 ctx->device.common.close = close_copybit;
1556 ctx->device.set_parameter = set_parameter_copybit;
1557 ctx->device.get = get;
1558 ctx->device.blit = blit_copybit;
1559 ctx->device.set_sync = set_sync_copybit;
1560 ctx->device.stretch = stretch_copybit;
1561 ctx->device.finish = finish_copybit;
1562 ctx->device.flush_get_fence = flush_get_fence_copybit;
1563 ctx->device.clear = clear_copybit;
1564
1565 /* Create RGB Surface */
1566 surfDefinition.buffer = (void*)0xdddddddd;
1567 surfDefinition.phys = (void*)0xdddddddd;
1568 surfDefinition.stride = 1 * 4;
1569 surfDefinition.width = 1;
1570 surfDefinition.height = 1;
1571 surfDefinition.format = C2D_COLOR_FORMAT_8888_ARGB;
1572 if (LINK_c2dCreateSurface(&(ctx->dst[RGB_SURFACE]), C2D_TARGET | C2D_SOURCE,
1573 (C2D_SURFACE_TYPE)(C2D_SURFACE_RGB_HOST |
1574 C2D_SURFACE_WITH_PHYS |
1575 C2D_SURFACE_WITH_PHYS_DUMMY ),
1576 &surfDefinition)) {
1577 ALOGE("%s: create ctx->dst_surface[RGB_SURFACE] failed", __FUNCTION__);
1578 ctx->dst[RGB_SURFACE] = 0;
1579 clean_up(ctx);
1580 status = COPYBIT_FAILURE;
1581 *device = NULL;
1582 return status;
1583 }
1584
1585 unsigned int surface_id = 0;
1586 for (int i = 0; i < MAX_RGB_SURFACES; i++)
1587 {
1588 if (LINK_c2dCreateSurface(&surface_id, C2D_TARGET | C2D_SOURCE,
1589 (C2D_SURFACE_TYPE)(C2D_SURFACE_RGB_HOST |
1590 C2D_SURFACE_WITH_PHYS |
1591 C2D_SURFACE_WITH_PHYS_DUMMY ),
1592 &surfDefinition)) {
1593 ALOGE("%s: create RGB source surface %d failed", __FUNCTION__, i);
1594 ctx->blit_rgb_object[i].surface_id = 0;
1595 status = COPYBIT_FAILURE;
1596 break;
1597 } else {
1598 ctx->blit_rgb_object[i].surface_id = surface_id;
1599 ALOGW("%s i = %d surface_id=%d", __FUNCTION__, i,
1600 ctx->blit_rgb_object[i].surface_id);
1601 }
1602 }
1603
1604 if (status == COPYBIT_FAILURE) {
1605 clean_up(ctx);
1606 status = COPYBIT_FAILURE;
1607 *device = NULL;
1608 return status;
1609 }
1610
1611 // Create 2 plane YUV surfaces
1612 yuvSurfaceDef.format = C2D_COLOR_FORMAT_420_NV12;
1613 yuvSurfaceDef.width = 4;
1614 yuvSurfaceDef.height = 4;
1615 yuvSurfaceDef.plane0 = (void*)0xaaaaaaaa;
1616 yuvSurfaceDef.phys0 = (void*) 0xaaaaaaaa;
1617 yuvSurfaceDef.stride0 = 4;
1618
1619 yuvSurfaceDef.plane1 = (void*)0xaaaaaaaa;
1620 yuvSurfaceDef.phys1 = (void*) 0xaaaaaaaa;
1621 yuvSurfaceDef.stride1 = 4;
1622 if (LINK_c2dCreateSurface(&(ctx->dst[YUV_SURFACE_2_PLANES]),
1623 C2D_TARGET | C2D_SOURCE,
1624 (C2D_SURFACE_TYPE)(C2D_SURFACE_YUV_HOST |
1625 C2D_SURFACE_WITH_PHYS |
1626 C2D_SURFACE_WITH_PHYS_DUMMY),
1627 &yuvSurfaceDef)) {
1628 ALOGE("%s: create ctx->dst[YUV_SURFACE_2_PLANES] failed", __FUNCTION__);
1629 ctx->dst[YUV_SURFACE_2_PLANES] = 0;
1630 clean_up(ctx);
1631 status = COPYBIT_FAILURE;
1632 *device = NULL;
1633 return status;
1634 }
1635
1636 for (int i=0; i < MAX_YUV_2_PLANE_SURFACES; i++)
1637 {
1638 if (LINK_c2dCreateSurface(&surface_id, C2D_TARGET | C2D_SOURCE,
1639 (C2D_SURFACE_TYPE)(C2D_SURFACE_YUV_HOST |
1640 C2D_SURFACE_WITH_PHYS |
1641 C2D_SURFACE_WITH_PHYS_DUMMY ),
1642 &yuvSurfaceDef)) {
1643 ALOGE("%s: create YUV source %d failed", __FUNCTION__, i);
1644 ctx->blit_yuv_2_plane_object[i].surface_id = 0;
1645 status = COPYBIT_FAILURE;
1646 break;
1647 } else {
1648 ctx->blit_yuv_2_plane_object[i].surface_id = surface_id;
1649 ALOGW("%s: 2 Plane YUV i=%d surface_id=%d", __FUNCTION__, i,
1650 ctx->blit_yuv_2_plane_object[i].surface_id);
1651 }
1652 }
1653
1654 if (status == COPYBIT_FAILURE) {
1655 clean_up(ctx);
1656 status = COPYBIT_FAILURE;
1657 *device = NULL;
1658 return status;
1659 }
1660
1661 // Create YUV 3 plane surfaces
1662 yuvSurfaceDef.format = C2D_COLOR_FORMAT_420_YV12;
1663 yuvSurfaceDef.plane2 = (void*)0xaaaaaaaa;
1664 yuvSurfaceDef.phys2 = (void*) 0xaaaaaaaa;
1665 yuvSurfaceDef.stride2 = 4;
1666
1667 if (LINK_c2dCreateSurface(&(ctx->dst[YUV_SURFACE_3_PLANES]),
1668 C2D_TARGET | C2D_SOURCE,
1669 (C2D_SURFACE_TYPE)(C2D_SURFACE_YUV_HOST |
1670 C2D_SURFACE_WITH_PHYS |
1671 C2D_SURFACE_WITH_PHYS_DUMMY),
1672 &yuvSurfaceDef)) {
1673 ALOGE("%s: create ctx->dst[YUV_SURFACE_3_PLANES] failed", __FUNCTION__);
1674 ctx->dst[YUV_SURFACE_3_PLANES] = 0;
1675 clean_up(ctx);
1676 status = COPYBIT_FAILURE;
1677 *device = NULL;
1678 return status;
1679 }
1680
1681 for (int i=0; i < MAX_YUV_3_PLANE_SURFACES; i++)
1682 {
1683 if (LINK_c2dCreateSurface(&(surface_id),
1684 C2D_TARGET | C2D_SOURCE,
1685 (C2D_SURFACE_TYPE)(C2D_SURFACE_YUV_HOST |
1686 C2D_SURFACE_WITH_PHYS |
1687 C2D_SURFACE_WITH_PHYS_DUMMY),
1688 &yuvSurfaceDef)) {
1689 ALOGE("%s: create 3 plane YUV surface %d failed", __FUNCTION__, i);
1690 ctx->blit_yuv_3_plane_object[i].surface_id = 0;
1691 status = COPYBIT_FAILURE;
1692 break;
1693 } else {
1694 ctx->blit_yuv_3_plane_object[i].surface_id = surface_id;
1695 ALOGW("%s: 3 Plane YUV i=%d surface_id=%d", __FUNCTION__, i,
1696 ctx->blit_yuv_3_plane_object[i].surface_id);
1697 }
1698 }
1699
1700 if (status == COPYBIT_FAILURE) {
1701 clean_up(ctx);
1702 status = COPYBIT_FAILURE;
1703 *device = NULL;
1704 return status;
1705 }
1706
1707 if (LINK_c2dGetDriverCapabilities(&(ctx->c2d_driver_info))) {
1708 ALOGE("%s: LINK_c2dGetDriverCapabilities failed", __FUNCTION__);
1709 clean_up(ctx);
1710 status = COPYBIT_FAILURE;
1711 *device = NULL;
1712 return status;
1713 }
1714 // Initialize context variables.
1715 ctx->trg_transform = C2D_TARGET_ROTATE_0;
1716
1717 ctx->temp_src_buffer.fd = -1;
1718 ctx->temp_src_buffer.base = 0;
1719 ctx->temp_src_buffer.size = 0;
1720
1721 ctx->temp_dst_buffer.fd = -1;
1722 ctx->temp_dst_buffer.base = 0;
1723 ctx->temp_dst_buffer.size = 0;
1724
1725 ctx->fb_width = 0;
1726 ctx->fb_height = 0;
1727
1728 ctx->blit_rgb_count = 0;
1729 ctx->blit_yuv_2_plane_count = 0;
1730 ctx->blit_yuv_3_plane_count = 0;
1731 ctx->blit_count = 0;
1732
1733 ctx->wait_timestamp = false;
1734 ctx->stop_thread = false;
1735 pthread_mutex_init(&(ctx->wait_cleanup_lock), NULL);
1736 pthread_cond_init(&(ctx->wait_cleanup_cond), NULL);
1737 /* Start the wait thread */
1738 pthread_attr_t attr;
1739 pthread_attr_init(&attr);
1740 pthread_attr_setdetachstate(&attr, PTHREAD_CREATE_JOINABLE);
1741
1742 pthread_create(&ctx->wait_thread_id, &attr, &c2d_wait_loop,
1743 (void *)ctx);
1744 pthread_attr_destroy(&attr);
1745
1746 *device = &ctx->device.common;
1747 return status;
1748 }
1749