I have 2 questions:
1)How can i draw on screen, something which i created on pBuffer…ie an image is attached to a pBuffer…and if i want to render it …how can i do it… i guess there are only 2 ways…using glReadPixels or using glCopyPixels.Am i right?
What does the api eglCreatePBufferFromClientBuffer() do?
Is the surface returned by it an offscreen buffer connected surface or framebuffer connected surface ? It seems it returns a surface which is connected to an offscreen buffer…
eglCreatePbufferFromClientBuffer allows you to take a buffer from an EGL client API and encapsulate it as a pbuffer surface. Currently the only supported buffer type is an OpenVG VGimage. This enables such use cases as rendering to a VGimage with OpenGL ES.
VGimage is an off-screen buffer, and so are pbuffers.
I’m not sure I understand your first question. Could you please rephrase it? Where does the image data come from and where do you want to put it?