I think that I don't quite get the Unity rendering engine.
I use RenderTexture to generate the screenshot (I need to manage it later on):
screenshotRenderTexture = new RenderTexture(screenshot.width, screenshot.height, depthBufferBits, RenderTextureFormat.Default);
screenshotRenderTexture.Create();
RenderTexture currentRenderTexture = RenderTexture.active;
RenderTexture.active = screenshotRenderTexture;
Camera[] cams = Camera.allCameras;
System.Array.Sort(
cams,
delegate(Camera cam1, Camera cam2)
{
// It's easier than write float to int conversion that won't floor
// depth deltas under 1 to zero and will correctly work with NaNs
if (cam1.depth < cam2.depth)
return -1;
else if (cam1.depth > cam2.depth)
return 1;
else return 0;
}
);
foreach(Camera cam in cams)
{
cam.targetTexture = screenshotRenderTexture;
cam.Render();
cam.targetTexture = null;
}
screenshot.ReadPixels(new Rect(0, 0, textureWidth, textureHeight), 0, 0);
screenshot.Apply();
RenderTexture.active = currentRenderTexture;
However, if depthBufferBits is 0, the render results with all kind of z-buffer bugs (things rendered in wrong order).
I understand what is depth buffer in a general sense. However, what I don't understand is — if RenderTexture is used to combine render results of individual cameras, why is depth buffer needed in it? How do this abstractions work, exactly — does camera create image on it's own and then gives it to RenderTexture, or does camera use RenderTexture's depth buffer? It seems like it's the latter, because of bugs I experience (the things in wrong order are all rendered with the same camera, so the issues is with ordering stuff inside one camera, not ordering the stuff between different cameras), but at the same time it kind of contradicts the common sense of how these abstractions are structured on C# level.
And, finally — can I somehow use the default depth buffer that is used for normal rendering on this one? Because 16 bits per pixel on mobile device is pretty painful.
Update:
Here's what I attempted to do:
screenshotRenderTexture = new RenderTexture(
screenshot.width,
screenshot.height,
0,
RenderTextureFormat.Default
);
screenshotRenderTexture.Create();
RenderBuffer currentColorBuffer = Graphics.activeColorBuffer;
Graphics.SetRenderTarget(screenshotRenderTexture.colorBuffer, Graphics.activeDepthBuffer);
yield return new WaitForEndOfFrame();
Graphics.SetRenderTarget(currentColorBuffer, Graphics.activeDepthBuffer);
And here's what I got:
SetRenderTarget can only mix color & depth buffers from RenderTextures. You're trying to set depth buffer from the screen.
UnityEngine.Graphics:SetRenderTarget(RenderBuffer, RenderBuffer)
<ScreenshotTaking>c__Iterator21:MoveNext() (at Assets/Scripts/Managers/ScreenshotManager.cs:126)
Why can't it mix depth buffer from the screen and color buffer from RenderTexture?
A depth render texture format. Depth format is used to render high precision "depth" value into a render texture. Which format is actually used depends on the platform. On OpenGL it is the native "depth component" format (usually 24 or 16 bits), on Direct3D9 it is the 32 bit floating point ("R32F") format.
A depth buffer, also known as a z-buffer, is a type of data buffer used in computer graphics to represent depth information of objects in 3D space from a particular perspective. Depth buffers are an aid to rendering a scene to ensure that the correct polygons properly occlude other polygons.
Description. Copies source texture into destination render texture with a shader. This is mostly used for implementing post-processing effects. Blit sets dest as the render target, sets source _MainTex property on the material, and draws a full-screen quad.
Render textures are textures that can be rendered to. They can be used to implement image based rendering effects, dynamic shadows, projectors, reflections or surveillance cameras. One typical usage of render textures is setting them as the "target texture" property of a Camera (Camera.
I don't know Unity well, but I know their base layer, and if they can map over say, D3D9, D3d10 and OpenGL then their abstraction has to use a common denominator.
In which case here, the D3D10 is the most limitative, you cannot share a depth surface between render targets of different sizes. If you have the same size of screen and render targets all over the place then you can indeed bind a unique depth buffer to different render targets.
The depth buffer is not stricto sensu necessary like you have observed you can render without a depth buffer but then the result is just the rendering in the order of which the draw commands are issued. (draw call = DrawPrimitive in D3D, or glDrawBuffers and the likes) It is even guaranteed that the order will be consistent at the triangle level by the spec, even if the graphic cards are very parallel they are refused to issue primitives in parallel for the sake of consistency of drawing through different runs of one draw call.
If you use a depth buffer the objects that happen to have their draw call after objects that have been drawn at a lower depth, will overwrite these close objects (in view space) and give wrong results, the depth buffer helps to discard pixel by pixel, the pixels of an object that would have its depth (in view space) deeper than something that has already been rendered at this pixel before, but with a closer depth.
binding the depth buffer also helps performance because if every pixel in one block have minimum depth of a certain value, the primitive rasterizer knows after the exit of the vertex buffer that a whole part of your primitive will never be rendered on this block, and discards the whole block altogether. This is called early Z culling and is a huge help in performance. So it is very preferable to keep the depth buffer plugged.
The camera has no concept in low level graphics theory, it is just represented by a view matrix that is an inverse transformation applied to the whole world, to shift the world from world space to view space as part of any single vertex transformation calculation. That is why in a classic vertex shader the position is taken from object space (in the vertex buffer stream), then multiplied by an object matrix transformation, then by a view matrix transform, and then a projection matrix transform, and then the rasterizer devides everything by 'w' to make the perspective divide.
That is by exploiting this pipeline behavior that you "create" the concept of a camera. Unity must abstract all of these by exposing a camera class. Maybe the camera class even have a "texture target" member to explain where the rendering from this camera will be stored.
And yes, the rendering is made DIRECTLY to the render texture that is specified, there is no intermediary front buffer or whatever, a render target is a hardware supported feature, and needs no copy at the end of a rendering.
The render target is a whole configuration in itself, because of hardware multisampling it may actually bind multiple buffers of the size of the required resolution.
There is the color buffer, which is for example 4 RGBA surfaces in MSAA4x, the depth buffer, which is generally a fixed point 24 bit representation, and 8 bits for stencil surface. All of these surfaces represent the render target configuration and are necessary to rendering.
I hope this helps
If you love us? You can donate to us via Paypal or buy me a coffee so we can maintain and grow! Thank you!
Donate Us With