)]}' { "commit": "497ba0e08503806571b52ebe27cc7eee4c0e71a7", "tree": "0edeb7b6cce3fa669fb45be3ef3a1dd6febde936", "parents": [ "40da5283ebc6b5cf1e3820740dc274c47cc55f6d" ], "author": { "name": "Jesse Hall", "email": "jessehall@google.com", "time": "Mon Nov 04 16:43:03 2013 -0800" }, "committer": { "name": "Jesse Hall", "email": "jessehall@google.com", "time": "Mon Nov 04 16:43:03 2013 -0800" }, "message": "Don\u0027t use implementation-defined format with CPU consumers\n\nIf the virtual display surface is being consumed by the CPU, it can\u0027t\nbe allowed with HAL_PIXEL_FORMAT_IMPLEMENTATION_DEFINED since there is\nno way for the CPU consumer to find out what format gralloc chose. So\nfor CPU-consumer surfaces, just use the BufferQueue\u0027s default format,\nwhich can be set by the consumer.\n\nA better but more invasive change would be to let the consumer require\na certain format (or set of formats?), and disallow the producer from\nrequesting a different format.\n\nBug: 11479817\nChange-Id: I5b20ee6ac1146550e8799b806e14661d279670c0\n", "tree_diff": [ { "type": "modify", "old_id": "bbe8d68a74b01293a4c2a5afbbdc4415e94c9811", "old_mode": 33188, "old_path": "services/surfaceflinger/DisplayHardware/VirtualDisplaySurface.cpp", "new_id": "2bf7d21412eaeef8c156429d794573e2d35493b1", "new_mode": 33188, "new_path": "services/surfaceflinger/DisplayHardware/VirtualDisplaySurface.cpp" }, { "type": "modify", "old_id": "57b55543d310871c382fb355a07ea62d7eb029bd", "old_mode": 33188, "old_path": "services/surfaceflinger/DisplayHardware/VirtualDisplaySurface.h", "new_id": "1e85ac4bdb029f01b9c73a248eb8d68f95148502", "new_mode": 33188, "new_path": "services/surfaceflinger/DisplayHardware/VirtualDisplaySurface.h" } ] }