mirror of
https://github.com/gfx-rs/wgpu.git
synced 2025-12-08 21:26:17 +00:00
2868 lines
112 KiB
Rust
2868 lines
112 KiB
Rust
use crate::resource::Resource;
|
|
use crate::snatch::SnatchGuard;
|
|
use crate::{
|
|
api_log,
|
|
binding_model::BindError,
|
|
command::{
|
|
self,
|
|
bind::Binder,
|
|
end_occlusion_query, end_pipeline_statistics_query,
|
|
memory_init::{fixup_discarded_surfaces, SurfacesInDiscardState},
|
|
BasePass, BasePassRef, BindGroupStateChange, CommandBuffer, CommandEncoderError,
|
|
CommandEncoderStatus, DrawError, ExecutionError, MapPassErr, PassErrorScope, QueryUseError,
|
|
RenderCommand, RenderCommandError, StateChange,
|
|
},
|
|
device::{
|
|
AttachmentData, Device, DeviceError, MissingDownlevelFlags, MissingFeatures,
|
|
RenderPassCompatibilityCheckType, RenderPassCompatibilityError, RenderPassContext,
|
|
},
|
|
error::{ErrorFormatter, PrettyError},
|
|
global::Global,
|
|
hal_api::HalApi,
|
|
hal_label, id,
|
|
identity::GlobalIdentityHandlerFactory,
|
|
init_tracker::{MemoryInitKind, TextureInitRange, TextureInitTrackerAction},
|
|
pipeline::{self, PipelineFlags},
|
|
resource::{Buffer, QuerySet, Texture, TextureView, TextureViewNotRenderableReason},
|
|
storage::Storage,
|
|
track::{TextureSelector, Tracker, UsageConflict, UsageScope},
|
|
validation::{
|
|
check_buffer_usage, check_texture_usage, MissingBufferUsageError, MissingTextureUsageError,
|
|
},
|
|
Label,
|
|
};
|
|
|
|
use arrayvec::ArrayVec;
|
|
use hal::CommandEncoder as _;
|
|
use thiserror::Error;
|
|
use wgt::{
|
|
BufferAddress, BufferSize, BufferUsages, Color, IndexFormat, TextureUsages,
|
|
TextureViewDimension, VertexStepMode,
|
|
};
|
|
|
|
#[cfg(feature = "serde")]
|
|
use serde::Deserialize;
|
|
#[cfg(feature = "serde")]
|
|
use serde::Serialize;
|
|
|
|
use std::sync::Arc;
|
|
use std::{borrow::Cow, fmt, iter, marker::PhantomData, mem, num::NonZeroU32, ops::Range, str};
|
|
|
|
use super::{
|
|
memory_init::TextureSurfaceDiscard, CommandBufferTextureMemoryActions, CommandEncoder,
|
|
QueryResetMap,
|
|
};
|
|
|
|
/// Operation to perform to the output attachment at the start of a renderpass.
|
|
#[repr(C)]
|
|
#[derive(Copy, Clone, Debug, Hash, Eq, PartialEq)]
|
|
#[cfg_attr(feature = "serde", derive(Serialize, Deserialize))]
|
|
#[cfg_attr(feature = "serde", serde(rename_all = "kebab-case"))]
|
|
pub enum LoadOp {
|
|
/// Clear the output attachment with the clear color. Clearing is faster than loading.
|
|
Clear = 0,
|
|
/// Do not clear output attachment.
|
|
Load = 1,
|
|
}
|
|
|
|
/// Operation to perform to the output attachment at the end of a renderpass.
|
|
#[repr(C)]
|
|
#[derive(Copy, Clone, Debug, Hash, Eq, PartialEq)]
|
|
#[cfg_attr(feature = "serde", derive(Serialize, Deserialize))]
|
|
#[cfg_attr(feature = "serde", serde(rename_all = "kebab-case"))]
|
|
pub enum StoreOp {
|
|
/// Discards the content of the render target.
|
|
///
|
|
/// If you don't care about the contents of the target, this can be faster.
|
|
Discard = 0,
|
|
/// Store the result of the renderpass.
|
|
Store = 1,
|
|
}
|
|
|
|
/// Describes an individual channel within a render pass, such as color, depth, or stencil.
|
|
#[repr(C)]
|
|
#[derive(Clone, Debug, Eq, PartialEq)]
|
|
#[cfg_attr(feature = "serde", derive(Serialize, Deserialize))]
|
|
pub struct PassChannel<V> {
|
|
/// Operation to perform to the output attachment at the start of a
|
|
/// renderpass.
|
|
///
|
|
/// This must be clear if it is the first renderpass rendering to a swap
|
|
/// chain image.
|
|
pub load_op: LoadOp,
|
|
/// Operation to perform to the output attachment at the end of a renderpass.
|
|
pub store_op: StoreOp,
|
|
/// If load_op is [`LoadOp::Clear`], the attachment will be cleared to this
|
|
/// color.
|
|
pub clear_value: V,
|
|
/// If true, the relevant channel is not changed by a renderpass, and the
|
|
/// corresponding attachment can be used inside the pass by other read-only
|
|
/// usages.
|
|
pub read_only: bool,
|
|
}
|
|
|
|
impl<V> PassChannel<V> {
|
|
fn hal_ops(&self) -> hal::AttachmentOps {
|
|
let mut ops = hal::AttachmentOps::empty();
|
|
match self.load_op {
|
|
LoadOp::Load => ops |= hal::AttachmentOps::LOAD,
|
|
LoadOp::Clear => (),
|
|
};
|
|
match self.store_op {
|
|
StoreOp::Store => ops |= hal::AttachmentOps::STORE,
|
|
StoreOp::Discard => (),
|
|
};
|
|
ops
|
|
}
|
|
}
|
|
|
|
/// Describes a color attachment to a render pass.
|
|
#[repr(C)]
|
|
#[derive(Clone, Debug, PartialEq)]
|
|
#[cfg_attr(feature = "serde", derive(Serialize, Deserialize))]
|
|
pub struct RenderPassColorAttachment {
|
|
/// The view to use as an attachment.
|
|
pub view: id::TextureViewId,
|
|
/// The view that will receive the resolved output if multisampling is used.
|
|
pub resolve_target: Option<id::TextureViewId>,
|
|
/// What operations will be performed on this color attachment.
|
|
pub channel: PassChannel<Color>,
|
|
}
|
|
|
|
/// Describes a depth/stencil attachment to a render pass.
|
|
#[repr(C)]
|
|
#[derive(Clone, Debug, PartialEq)]
|
|
#[cfg_attr(feature = "serde", derive(Serialize, Deserialize))]
|
|
pub struct RenderPassDepthStencilAttachment {
|
|
/// The view to use as an attachment.
|
|
pub view: id::TextureViewId,
|
|
/// What operations will be performed on the depth part of the attachment.
|
|
pub depth: PassChannel<f32>,
|
|
/// What operations will be performed on the stencil part of the attachment.
|
|
pub stencil: PassChannel<u32>,
|
|
}
|
|
|
|
impl RenderPassDepthStencilAttachment {
|
|
/// Validate the given aspects' read-only flags against their load
|
|
/// and store ops.
|
|
///
|
|
/// When an aspect is read-only, its load and store ops must be
|
|
/// `LoadOp::Load` and `StoreOp::Store`.
|
|
///
|
|
/// On success, return a pair `(depth, stencil)` indicating
|
|
/// whether the depth and stencil passes are read-only.
|
|
fn depth_stencil_read_only(
|
|
&self,
|
|
aspects: hal::FormatAspects,
|
|
) -> Result<(bool, bool), RenderPassErrorInner> {
|
|
let mut depth_read_only = true;
|
|
let mut stencil_read_only = true;
|
|
|
|
if aspects.contains(hal::FormatAspects::DEPTH) {
|
|
if self.depth.read_only
|
|
&& (self.depth.load_op, self.depth.store_op) != (LoadOp::Load, StoreOp::Store)
|
|
{
|
|
return Err(RenderPassErrorInner::InvalidDepthOps);
|
|
}
|
|
depth_read_only = self.depth.read_only;
|
|
}
|
|
|
|
if aspects.contains(hal::FormatAspects::STENCIL) {
|
|
if self.stencil.read_only
|
|
&& (self.stencil.load_op, self.stencil.store_op) != (LoadOp::Load, StoreOp::Store)
|
|
{
|
|
return Err(RenderPassErrorInner::InvalidStencilOps);
|
|
}
|
|
stencil_read_only = self.stencil.read_only;
|
|
}
|
|
|
|
Ok((depth_read_only, stencil_read_only))
|
|
}
|
|
}
|
|
|
|
/// Location to write a timestamp to (beginning or end of the pass).
|
|
#[repr(C)]
|
|
#[derive(Copy, Clone, Debug, Hash, Eq, PartialEq)]
|
|
#[cfg_attr(feature = "serde", derive(Serialize, Deserialize))]
|
|
#[cfg_attr(feature = "serde", serde(rename_all = "kebab-case"))]
|
|
pub enum RenderPassTimestampLocation {
|
|
Beginning = 0,
|
|
End = 1,
|
|
}
|
|
|
|
/// Describes the writing of timestamp values in a render pass.
|
|
#[repr(C)]
|
|
#[derive(Clone, Debug, PartialEq, Eq)]
|
|
#[cfg_attr(feature = "serde", derive(Serialize, Deserialize))]
|
|
pub struct RenderPassTimestampWrites {
|
|
/// The query set to write the timestamp to.
|
|
pub query_set: id::QuerySetId,
|
|
/// The index of the query set at which a start timestamp of this pass is written, if any.
|
|
pub beginning_of_pass_write_index: Option<u32>,
|
|
/// The index of the query set at which an end timestamp of this pass is written, if any.
|
|
pub end_of_pass_write_index: Option<u32>,
|
|
}
|
|
|
|
/// Describes the attachments of a render pass.
|
|
#[derive(Clone, Debug, Default, PartialEq)]
|
|
pub struct RenderPassDescriptor<'a> {
|
|
pub label: Label<'a>,
|
|
/// The color attachments of the render pass.
|
|
pub color_attachments: Cow<'a, [Option<RenderPassColorAttachment>]>,
|
|
/// The depth and stencil attachment of the render pass, if any.
|
|
pub depth_stencil_attachment: Option<&'a RenderPassDepthStencilAttachment>,
|
|
/// Defines where and when timestamp values will be written for this pass.
|
|
pub timestamp_writes: Option<&'a RenderPassTimestampWrites>,
|
|
/// Defines where the occlusion query results will be stored for this pass.
|
|
pub occlusion_query_set: Option<id::QuerySetId>,
|
|
}
|
|
|
|
#[cfg_attr(feature = "serde", derive(Deserialize, Serialize))]
|
|
pub struct RenderPass {
|
|
base: BasePass<RenderCommand>,
|
|
parent_id: id::CommandEncoderId,
|
|
color_targets: ArrayVec<Option<RenderPassColorAttachment>, { hal::MAX_COLOR_ATTACHMENTS }>,
|
|
depth_stencil_target: Option<RenderPassDepthStencilAttachment>,
|
|
timestamp_writes: Option<RenderPassTimestampWrites>,
|
|
occlusion_query_set_id: Option<id::QuerySetId>,
|
|
|
|
// Resource binding dedupe state.
|
|
#[cfg_attr(feature = "serde", serde(skip))]
|
|
current_bind_groups: BindGroupStateChange,
|
|
#[cfg_attr(feature = "serde", serde(skip))]
|
|
current_pipeline: StateChange<id::RenderPipelineId>,
|
|
}
|
|
|
|
impl RenderPass {
|
|
pub fn new(parent_id: id::CommandEncoderId, desc: &RenderPassDescriptor) -> Self {
|
|
Self {
|
|
base: BasePass::new(&desc.label),
|
|
parent_id,
|
|
color_targets: desc.color_attachments.iter().cloned().collect(),
|
|
depth_stencil_target: desc.depth_stencil_attachment.cloned(),
|
|
timestamp_writes: desc.timestamp_writes.cloned(),
|
|
occlusion_query_set_id: desc.occlusion_query_set,
|
|
|
|
current_bind_groups: BindGroupStateChange::new(),
|
|
current_pipeline: StateChange::new(),
|
|
}
|
|
}
|
|
|
|
pub fn parent_id(&self) -> id::CommandEncoderId {
|
|
self.parent_id
|
|
}
|
|
|
|
#[cfg(feature = "trace")]
|
|
pub fn into_command(self) -> crate::device::trace::Command {
|
|
crate::device::trace::Command::RunRenderPass {
|
|
base: self.base,
|
|
target_colors: self.color_targets.into_iter().collect(),
|
|
target_depth_stencil: self.depth_stencil_target,
|
|
timestamp_writes: self.timestamp_writes,
|
|
occlusion_query_set_id: self.occlusion_query_set_id,
|
|
}
|
|
}
|
|
|
|
pub fn set_index_buffer(
|
|
&mut self,
|
|
buffer_id: id::BufferId,
|
|
index_format: IndexFormat,
|
|
offset: BufferAddress,
|
|
size: Option<BufferSize>,
|
|
) {
|
|
self.base.commands.push(RenderCommand::SetIndexBuffer {
|
|
buffer_id,
|
|
index_format,
|
|
offset,
|
|
size,
|
|
});
|
|
}
|
|
}
|
|
|
|
impl fmt::Debug for RenderPass {
|
|
fn fmt(&self, f: &mut fmt::Formatter<'_>) -> fmt::Result {
|
|
f.debug_struct("RenderPass")
|
|
.field("encoder_id", &self.parent_id)
|
|
.field("color_targets", &self.color_targets)
|
|
.field("depth_stencil_target", &self.depth_stencil_target)
|
|
.field("command count", &self.base.commands.len())
|
|
.field("dynamic offset count", &self.base.dynamic_offsets.len())
|
|
.field(
|
|
"push constant u32 count",
|
|
&self.base.push_constant_data.len(),
|
|
)
|
|
.finish()
|
|
}
|
|
}
|
|
|
|
#[derive(Debug, PartialEq)]
|
|
enum OptionalState {
|
|
Unused,
|
|
Required,
|
|
Set,
|
|
}
|
|
|
|
impl OptionalState {
|
|
fn require(&mut self, require: bool) {
|
|
if require && *self == Self::Unused {
|
|
*self = Self::Required;
|
|
}
|
|
}
|
|
}
|
|
|
|
#[derive(Debug, Default)]
|
|
struct IndexState {
|
|
bound_buffer_view: Option<(id::BufferId, Range<BufferAddress>)>,
|
|
format: Option<IndexFormat>,
|
|
pipeline_format: Option<IndexFormat>,
|
|
limit: u64,
|
|
}
|
|
|
|
impl IndexState {
|
|
fn update_limit(&mut self) {
|
|
self.limit = match self.bound_buffer_view {
|
|
Some((_, ref range)) => {
|
|
let format = self
|
|
.format
|
|
.expect("IndexState::update_limit must be called after a index buffer is set");
|
|
let shift = match format {
|
|
IndexFormat::Uint16 => 1,
|
|
IndexFormat::Uint32 => 2,
|
|
};
|
|
|
|
(range.end - range.start) >> shift
|
|
}
|
|
None => 0,
|
|
}
|
|
}
|
|
|
|
fn reset(&mut self) {
|
|
self.bound_buffer_view = None;
|
|
self.limit = 0;
|
|
}
|
|
}
|
|
|
|
#[derive(Clone, Copy, Debug)]
|
|
struct VertexBufferState {
|
|
total_size: BufferAddress,
|
|
step: pipeline::VertexStep,
|
|
bound: bool,
|
|
}
|
|
|
|
impl VertexBufferState {
|
|
const EMPTY: Self = Self {
|
|
total_size: 0,
|
|
step: pipeline::VertexStep {
|
|
stride: 0,
|
|
last_stride: 0,
|
|
mode: VertexStepMode::Vertex,
|
|
},
|
|
bound: false,
|
|
};
|
|
}
|
|
|
|
#[derive(Debug, Default)]
|
|
struct VertexState {
|
|
inputs: ArrayVec<VertexBufferState, { hal::MAX_VERTEX_BUFFERS }>,
|
|
/// Length of the shortest vertex rate vertex buffer
|
|
vertex_limit: u64,
|
|
/// Buffer slot which the shortest vertex rate vertex buffer is bound to
|
|
vertex_limit_slot: u32,
|
|
/// Length of the shortest instance rate vertex buffer
|
|
instance_limit: u64,
|
|
/// Buffer slot which the shortest instance rate vertex buffer is bound to
|
|
instance_limit_slot: u32,
|
|
/// Total amount of buffers required by the pipeline.
|
|
buffers_required: u32,
|
|
}
|
|
|
|
impl VertexState {
|
|
fn update_limits(&mut self) {
|
|
// Implements the validation from https://gpuweb.github.io/gpuweb/#dom-gpurendercommandsmixin-draw
|
|
// Except that the formula is shuffled to extract the number of vertices in order
|
|
// to carry the bulk of the computation when changing states intead of when producing
|
|
// draws. Draw calls tend to happen at a higher frequency. Here we determine vertex
|
|
// limits that can be cheaply checked for each draw call.
|
|
self.vertex_limit = u32::MAX as u64;
|
|
self.instance_limit = u32::MAX as u64;
|
|
for (idx, vbs) in self.inputs.iter().enumerate() {
|
|
if !vbs.bound {
|
|
continue;
|
|
}
|
|
|
|
let limit = if vbs.total_size < vbs.step.last_stride {
|
|
// The buffer cannot fit the last vertex.
|
|
0
|
|
} else {
|
|
if vbs.step.stride == 0 {
|
|
// We already checked that the last stride fits, the same
|
|
// vertex will be repeated so this slot can accomodate any number of
|
|
// vertices.
|
|
continue;
|
|
}
|
|
|
|
// The general case.
|
|
(vbs.total_size - vbs.step.last_stride) / vbs.step.stride + 1
|
|
};
|
|
|
|
match vbs.step.mode {
|
|
VertexStepMode::Vertex => {
|
|
if limit < self.vertex_limit {
|
|
self.vertex_limit = limit;
|
|
self.vertex_limit_slot = idx as _;
|
|
}
|
|
}
|
|
VertexStepMode::Instance => {
|
|
if limit < self.instance_limit {
|
|
self.instance_limit = limit;
|
|
self.instance_limit_slot = idx as _;
|
|
}
|
|
}
|
|
}
|
|
}
|
|
}
|
|
|
|
fn reset(&mut self) {
|
|
self.inputs.clear();
|
|
self.vertex_limit = 0;
|
|
self.instance_limit = 0;
|
|
}
|
|
}
|
|
|
|
#[derive(Debug)]
|
|
struct State<A: HalApi> {
|
|
pipeline_flags: PipelineFlags,
|
|
binder: Binder<A>,
|
|
blend_constant: OptionalState,
|
|
stencil_reference: u32,
|
|
pipeline: Option<id::RenderPipelineId>,
|
|
index: IndexState,
|
|
vertex: VertexState,
|
|
debug_scope_depth: u32,
|
|
}
|
|
|
|
impl<A: HalApi> State<A> {
|
|
fn is_ready(&self, indexed: bool) -> Result<(), DrawError> {
|
|
// Determine how many vertex buffers have already been bound
|
|
let vertex_buffer_count = self.vertex.inputs.iter().take_while(|v| v.bound).count() as u32;
|
|
// Compare with the needed quantity
|
|
if vertex_buffer_count < self.vertex.buffers_required {
|
|
return Err(DrawError::MissingVertexBuffer {
|
|
index: vertex_buffer_count,
|
|
});
|
|
}
|
|
|
|
let bind_mask = self.binder.invalid_mask();
|
|
if bind_mask != 0 {
|
|
//let (expected, provided) = self.binder.entries[index as usize].info();
|
|
return Err(DrawError::IncompatibleBindGroup {
|
|
index: bind_mask.trailing_zeros(),
|
|
diff: self.binder.bgl_diff(),
|
|
});
|
|
}
|
|
if self.pipeline.is_none() {
|
|
return Err(DrawError::MissingPipeline);
|
|
}
|
|
if self.blend_constant == OptionalState::Required {
|
|
return Err(DrawError::MissingBlendConstant);
|
|
}
|
|
|
|
if indexed {
|
|
// Pipeline expects an index buffer
|
|
if let Some(pipeline_index_format) = self.index.pipeline_format {
|
|
// We have a buffer bound
|
|
let buffer_index_format = self.index.format.ok_or(DrawError::MissingIndexBuffer)?;
|
|
|
|
// The buffers are different formats
|
|
if pipeline_index_format != buffer_index_format {
|
|
return Err(DrawError::UnmatchedIndexFormats {
|
|
pipeline: pipeline_index_format,
|
|
buffer: buffer_index_format,
|
|
});
|
|
}
|
|
}
|
|
}
|
|
|
|
self.binder.check_late_buffer_bindings()?;
|
|
|
|
Ok(())
|
|
}
|
|
|
|
/// Reset the `RenderBundle`-related states.
|
|
fn reset_bundle(&mut self) {
|
|
self.binder.reset();
|
|
self.pipeline = None;
|
|
self.index.reset();
|
|
self.vertex.reset();
|
|
}
|
|
}
|
|
|
|
/// Describes an attachment location in words.
|
|
///
|
|
/// Can be used as "the {loc} has..." or "{loc} has..."
|
|
#[derive(Debug, Copy, Clone)]
|
|
pub enum AttachmentErrorLocation {
|
|
Color { index: usize, resolve: bool },
|
|
Depth,
|
|
}
|
|
|
|
impl fmt::Display for AttachmentErrorLocation {
|
|
fn fmt(&self, f: &mut fmt::Formatter<'_>) -> fmt::Result {
|
|
match *self {
|
|
AttachmentErrorLocation::Color {
|
|
index,
|
|
resolve: false,
|
|
} => write!(f, "color attachment at index {index}'s texture view"),
|
|
AttachmentErrorLocation::Color {
|
|
index,
|
|
resolve: true,
|
|
} => write!(
|
|
f,
|
|
"color attachment at index {index}'s resolve texture view"
|
|
),
|
|
AttachmentErrorLocation::Depth => write!(f, "depth attachment's texture view"),
|
|
}
|
|
}
|
|
}
|
|
|
|
#[derive(Clone, Debug, Error)]
|
|
#[non_exhaustive]
|
|
pub enum ColorAttachmentError {
|
|
#[error("Attachment format {0:?} is not a color format")]
|
|
InvalidFormat(wgt::TextureFormat),
|
|
#[error("The number of color attachments {given} exceeds the limit {limit}")]
|
|
TooMany { given: usize, limit: usize },
|
|
}
|
|
|
|
/// Error encountered when performing a render pass.
|
|
#[derive(Clone, Debug, Error)]
|
|
pub enum RenderPassErrorInner {
|
|
#[error(transparent)]
|
|
Device(DeviceError),
|
|
#[error(transparent)]
|
|
ColorAttachment(#[from] ColorAttachmentError),
|
|
#[error(transparent)]
|
|
Encoder(#[from] CommandEncoderError),
|
|
#[error("Attachment texture view {0:?} is invalid")]
|
|
InvalidAttachment(id::TextureViewId),
|
|
#[error("Attachment texture view {0:?} is invalid")]
|
|
InvalidResolveTarget(id::TextureViewId),
|
|
#[error("The format of the depth-stencil attachment ({0:?}) is not a depth-stencil format")]
|
|
InvalidDepthStencilAttachmentFormat(wgt::TextureFormat),
|
|
#[error("The format of the {location} ({format:?}) is not resolvable")]
|
|
UnsupportedResolveTargetFormat {
|
|
location: AttachmentErrorLocation,
|
|
format: wgt::TextureFormat,
|
|
},
|
|
#[error("No color attachments or depth attachments were provided, at least one attachment of any kind must be provided")]
|
|
MissingAttachments,
|
|
#[error("The {location} is not renderable:")]
|
|
TextureViewIsNotRenderable {
|
|
location: AttachmentErrorLocation,
|
|
#[source]
|
|
reason: TextureViewNotRenderableReason,
|
|
},
|
|
#[error("Attachments have differing sizes: the {expected_location} has extent {expected_extent:?} but is followed by the {actual_location} which has {actual_extent:?}")]
|
|
AttachmentsDimensionMismatch {
|
|
expected_location: AttachmentErrorLocation,
|
|
expected_extent: wgt::Extent3d,
|
|
actual_location: AttachmentErrorLocation,
|
|
actual_extent: wgt::Extent3d,
|
|
},
|
|
#[error("Attachments have differing sample counts: the {expected_location} has count {expected_samples:?} but is followed by the {actual_location} which has count {actual_samples:?}")]
|
|
AttachmentSampleCountMismatch {
|
|
expected_location: AttachmentErrorLocation,
|
|
expected_samples: u32,
|
|
actual_location: AttachmentErrorLocation,
|
|
actual_samples: u32,
|
|
},
|
|
#[error("The resolve source, {location}, must be multi-sampled (has {src} samples) while the resolve destination must not be multisampled (has {dst} samples)")]
|
|
InvalidResolveSampleCounts {
|
|
location: AttachmentErrorLocation,
|
|
src: u32,
|
|
dst: u32,
|
|
},
|
|
#[error(
|
|
"Resource source, {location}, format ({src:?}) must match the resolve destination format ({dst:?})"
|
|
)]
|
|
MismatchedResolveTextureFormat {
|
|
location: AttachmentErrorLocation,
|
|
src: wgt::TextureFormat,
|
|
dst: wgt::TextureFormat,
|
|
},
|
|
#[error("Surface texture is dropped before the render pass is finished")]
|
|
SurfaceTextureDropped,
|
|
#[error("Not enough memory left for render pass")]
|
|
OutOfMemory,
|
|
#[error("The bind group at index {0:?} is invalid")]
|
|
InvalidBindGroup(usize),
|
|
#[error("Unable to clear non-present/read-only depth")]
|
|
InvalidDepthOps,
|
|
#[error("Unable to clear non-present/read-only stencil")]
|
|
InvalidStencilOps,
|
|
#[error("Setting `values_offset` to be `None` is only for internal use in render bundles")]
|
|
InvalidValuesOffset,
|
|
#[error(transparent)]
|
|
MissingFeatures(#[from] MissingFeatures),
|
|
#[error(transparent)]
|
|
MissingDownlevelFlags(#[from] MissingDownlevelFlags),
|
|
#[error("Indirect draw uses bytes {offset}..{end_offset} {} which overruns indirect buffer of size {buffer_size}",
|
|
count.map_or_else(String::new, |v| format!("(using count {v})")))]
|
|
IndirectBufferOverrun {
|
|
count: Option<NonZeroU32>,
|
|
offset: u64,
|
|
end_offset: u64,
|
|
buffer_size: u64,
|
|
},
|
|
#[error("Indirect draw uses bytes {begin_count_offset}..{end_count_offset} which overruns indirect buffer of size {count_buffer_size}")]
|
|
IndirectCountBufferOverrun {
|
|
begin_count_offset: u64,
|
|
end_count_offset: u64,
|
|
count_buffer_size: u64,
|
|
},
|
|
#[error("Cannot pop debug group, because number of pushed debug groups is zero")]
|
|
InvalidPopDebugGroup,
|
|
#[error(transparent)]
|
|
ResourceUsageConflict(#[from] UsageConflict),
|
|
#[error("Render bundle has incompatible targets, {0}")]
|
|
IncompatibleBundleTargets(#[from] RenderPassCompatibilityError),
|
|
#[error(
|
|
"Render bundle has incompatible read-only flags: \
|
|
bundle has flags depth = {bundle_depth} and stencil = {bundle_stencil}, \
|
|
while the pass has flags depth = {pass_depth} and stencil = {pass_stencil}. \
|
|
Read-only renderpasses are only compatible with read-only bundles for that aspect."
|
|
)]
|
|
IncompatibleBundleReadOnlyDepthStencil {
|
|
pass_depth: bool,
|
|
pass_stencil: bool,
|
|
bundle_depth: bool,
|
|
bundle_stencil: bool,
|
|
},
|
|
#[error(transparent)]
|
|
RenderCommand(#[from] RenderCommandError),
|
|
#[error(transparent)]
|
|
Draw(#[from] DrawError),
|
|
#[error(transparent)]
|
|
Bind(#[from] BindError),
|
|
#[error(transparent)]
|
|
QueryUse(#[from] QueryUseError),
|
|
#[error("Multiview layer count must match")]
|
|
MultiViewMismatch,
|
|
#[error(
|
|
"Multiview pass texture views with more than one array layer must have D2Array dimension"
|
|
)]
|
|
MultiViewDimensionMismatch,
|
|
#[error("QuerySet {0:?} is invalid")]
|
|
InvalidQuerySet(id::QuerySetId),
|
|
#[error("missing occlusion query set")]
|
|
MissingOcclusionQuerySet,
|
|
}
|
|
|
|
impl PrettyError for RenderPassErrorInner {
|
|
fn fmt_pretty(&self, fmt: &mut ErrorFormatter) {
|
|
fmt.error(self);
|
|
if let Self::InvalidAttachment(id) = *self {
|
|
fmt.texture_view_label_with_key(&id, "attachment");
|
|
};
|
|
if let Self::Draw(DrawError::IncompatibleBindGroup { diff, .. }) = self {
|
|
for d in diff {
|
|
fmt.note(&d);
|
|
}
|
|
};
|
|
}
|
|
}
|
|
|
|
impl From<MissingBufferUsageError> for RenderPassErrorInner {
|
|
fn from(error: MissingBufferUsageError) -> Self {
|
|
Self::RenderCommand(error.into())
|
|
}
|
|
}
|
|
|
|
impl From<MissingTextureUsageError> for RenderPassErrorInner {
|
|
fn from(error: MissingTextureUsageError) -> Self {
|
|
Self::RenderCommand(error.into())
|
|
}
|
|
}
|
|
|
|
impl From<DeviceError> for RenderPassErrorInner {
|
|
fn from(error: DeviceError) -> Self {
|
|
Self::Device(error)
|
|
}
|
|
}
|
|
|
|
/// Error encountered when performing a render pass.
|
|
#[derive(Clone, Debug, Error)]
|
|
#[error("{scope}")]
|
|
pub struct RenderPassError {
|
|
pub scope: PassErrorScope,
|
|
#[source]
|
|
inner: RenderPassErrorInner,
|
|
}
|
|
impl PrettyError for RenderPassError {
|
|
fn fmt_pretty(&self, fmt: &mut ErrorFormatter) {
|
|
// This error is wrapper for the inner error,
|
|
// but the scope has useful labels
|
|
fmt.error(self);
|
|
self.scope.fmt_pretty(fmt);
|
|
}
|
|
}
|
|
|
|
impl<T, E> MapPassErr<T, RenderPassError> for Result<T, E>
|
|
where
|
|
E: Into<RenderPassErrorInner>,
|
|
{
|
|
fn map_pass_err(self, scope: PassErrorScope) -> Result<T, RenderPassError> {
|
|
self.map_err(|inner| RenderPassError {
|
|
scope,
|
|
inner: inner.into(),
|
|
})
|
|
}
|
|
}
|
|
|
|
struct RenderAttachment<'a, A: HalApi> {
|
|
texture: Arc<Texture<A>>,
|
|
selector: &'a TextureSelector,
|
|
usage: hal::TextureUses,
|
|
}
|
|
|
|
impl<A: HalApi> TextureView<A> {
|
|
fn to_render_attachment(&self, usage: hal::TextureUses) -> RenderAttachment<A> {
|
|
RenderAttachment {
|
|
texture: self.parent.clone(),
|
|
selector: &self.selector,
|
|
usage,
|
|
}
|
|
}
|
|
}
|
|
|
|
const MAX_TOTAL_ATTACHMENTS: usize = hal::MAX_COLOR_ATTACHMENTS + hal::MAX_COLOR_ATTACHMENTS + 1;
|
|
type AttachmentDataVec<T> = ArrayVec<T, MAX_TOTAL_ATTACHMENTS>;
|
|
|
|
struct RenderPassInfo<'a, A: HalApi> {
|
|
context: RenderPassContext,
|
|
usage_scope: UsageScope<A>,
|
|
/// All render attachments, including depth/stencil
|
|
render_attachments: AttachmentDataVec<RenderAttachment<'a, A>>,
|
|
is_depth_read_only: bool,
|
|
is_stencil_read_only: bool,
|
|
extent: wgt::Extent3d,
|
|
_phantom: PhantomData<A>,
|
|
|
|
pending_discard_init_fixups: SurfacesInDiscardState<A>,
|
|
divergent_discarded_depth_stencil_aspect: Option<(wgt::TextureAspect, &'a TextureView<A>)>,
|
|
multiview: Option<NonZeroU32>,
|
|
}
|
|
|
|
impl<'a, A: HalApi> RenderPassInfo<'a, A> {
|
|
fn add_pass_texture_init_actions<V>(
|
|
channel: &PassChannel<V>,
|
|
texture_memory_actions: &mut CommandBufferTextureMemoryActions<A>,
|
|
view: &TextureView<A>,
|
|
pending_discard_init_fixups: &mut SurfacesInDiscardState<A>,
|
|
) {
|
|
if channel.load_op == LoadOp::Load {
|
|
pending_discard_init_fixups.extend(texture_memory_actions.register_init_action(
|
|
&TextureInitTrackerAction {
|
|
texture: view.parent.clone(),
|
|
range: TextureInitRange::from(view.selector.clone()),
|
|
// Note that this is needed even if the target is discarded,
|
|
kind: MemoryInitKind::NeedsInitializedMemory,
|
|
},
|
|
));
|
|
} else if channel.store_op == StoreOp::Store {
|
|
// Clear + Store
|
|
texture_memory_actions.register_implicit_init(
|
|
&view.parent,
|
|
TextureInitRange::from(view.selector.clone()),
|
|
);
|
|
}
|
|
if channel.store_op == StoreOp::Discard {
|
|
// the discard happens at the *end* of a pass, but recording the
|
|
// discard right away be alright since the texture can't be used
|
|
// during the pass anyways
|
|
texture_memory_actions.discard(TextureSurfaceDiscard {
|
|
texture: view.parent.clone(),
|
|
mip_level: view.selector.mips.start,
|
|
layer: view.selector.layers.start,
|
|
});
|
|
}
|
|
}
|
|
|
|
fn start(
|
|
device: &Device<A>,
|
|
label: Option<&str>,
|
|
color_attachments: &[Option<RenderPassColorAttachment>],
|
|
depth_stencil_attachment: Option<&RenderPassDepthStencilAttachment>,
|
|
timestamp_writes: Option<&RenderPassTimestampWrites>,
|
|
occlusion_query_set: Option<id::QuerySetId>,
|
|
encoder: &mut CommandEncoder<A>,
|
|
trackers: &mut Tracker<A>,
|
|
texture_memory_actions: &mut CommandBufferTextureMemoryActions<A>,
|
|
pending_query_resets: &mut QueryResetMap<A>,
|
|
view_guard: &'a Storage<TextureView<A>>,
|
|
buffer_guard: &'a Storage<Buffer<A>>,
|
|
texture_guard: &'a Storage<Texture<A>>,
|
|
query_set_guard: &'a Storage<QuerySet<A>>,
|
|
snatch_guard: &SnatchGuard<'a>,
|
|
) -> Result<Self, RenderPassErrorInner> {
|
|
profiling::scope!("RenderPassInfo::start");
|
|
|
|
// We default to false intentionally, even if depth-stencil isn't used at all.
|
|
// This allows us to use the primary raw pipeline in `RenderPipeline`,
|
|
// instead of the special read-only one, which would be `None`.
|
|
let mut is_depth_read_only = false;
|
|
let mut is_stencil_read_only = false;
|
|
|
|
let mut render_attachments = AttachmentDataVec::<RenderAttachment<A>>::new();
|
|
let mut discarded_surfaces = AttachmentDataVec::new();
|
|
let mut pending_discard_init_fixups = SurfacesInDiscardState::new();
|
|
let mut divergent_discarded_depth_stencil_aspect = None;
|
|
|
|
let mut attachment_location = AttachmentErrorLocation::Color {
|
|
index: usize::MAX,
|
|
resolve: false,
|
|
};
|
|
let mut extent = None;
|
|
let mut sample_count = 0;
|
|
|
|
let mut detected_multiview: Option<Option<NonZeroU32>> = None;
|
|
|
|
let mut check_multiview = |view: &TextureView<A>| {
|
|
// Get the multiview configuration for this texture view
|
|
let layers = view.selector.layers.end - view.selector.layers.start;
|
|
let this_multiview = if layers >= 2 {
|
|
// Trivially proven by the if above
|
|
Some(unsafe { NonZeroU32::new_unchecked(layers) })
|
|
} else {
|
|
None
|
|
};
|
|
|
|
// Make sure that if this view is a multiview, it is set to be an array
|
|
if this_multiview.is_some() && view.desc.dimension != TextureViewDimension::D2Array {
|
|
return Err(RenderPassErrorInner::MultiViewDimensionMismatch);
|
|
}
|
|
|
|
// Validate matching first, or store the first one
|
|
if let Some(multiview) = detected_multiview {
|
|
if multiview != this_multiview {
|
|
return Err(RenderPassErrorInner::MultiViewMismatch);
|
|
}
|
|
} else {
|
|
// Multiview is only supported if the feature is enabled
|
|
if this_multiview.is_some() {
|
|
device.require_features(wgt::Features::MULTIVIEW)?;
|
|
}
|
|
|
|
detected_multiview = Some(this_multiview);
|
|
}
|
|
|
|
Ok(())
|
|
};
|
|
let mut add_view = |view: &TextureView<A>, location| {
|
|
let render_extent = view.render_extent.map_err(|reason| {
|
|
RenderPassErrorInner::TextureViewIsNotRenderable { location, reason }
|
|
})?;
|
|
if let Some(ex) = extent {
|
|
if ex != render_extent {
|
|
return Err(RenderPassErrorInner::AttachmentsDimensionMismatch {
|
|
expected_location: attachment_location,
|
|
expected_extent: ex,
|
|
actual_location: location,
|
|
actual_extent: render_extent,
|
|
});
|
|
}
|
|
} else {
|
|
extent = Some(render_extent);
|
|
}
|
|
if sample_count == 0 {
|
|
sample_count = view.samples;
|
|
} else if sample_count != view.samples {
|
|
return Err(RenderPassErrorInner::AttachmentSampleCountMismatch {
|
|
expected_location: attachment_location,
|
|
expected_samples: sample_count,
|
|
actual_location: location,
|
|
actual_samples: view.samples,
|
|
});
|
|
}
|
|
attachment_location = location;
|
|
Ok(())
|
|
};
|
|
|
|
let mut colors =
|
|
ArrayVec::<Option<hal::ColorAttachment<A>>, { hal::MAX_COLOR_ATTACHMENTS }>::new();
|
|
let mut depth_stencil = None;
|
|
|
|
if let Some(at) = depth_stencil_attachment {
|
|
let view: &TextureView<A> = trackers
|
|
.views
|
|
.add_single(view_guard, at.view)
|
|
.ok_or(RenderPassErrorInner::InvalidAttachment(at.view))?;
|
|
check_multiview(view)?;
|
|
add_view(view, AttachmentErrorLocation::Depth)?;
|
|
|
|
let ds_aspects = view.desc.aspects();
|
|
if ds_aspects.contains(hal::FormatAspects::COLOR) {
|
|
return Err(RenderPassErrorInner::InvalidDepthStencilAttachmentFormat(
|
|
view.desc.format,
|
|
));
|
|
}
|
|
|
|
if !ds_aspects.contains(hal::FormatAspects::STENCIL)
|
|
|| (at.stencil.load_op == at.depth.load_op
|
|
&& at.stencil.store_op == at.depth.store_op)
|
|
{
|
|
Self::add_pass_texture_init_actions(
|
|
&at.depth,
|
|
texture_memory_actions,
|
|
view,
|
|
&mut pending_discard_init_fixups,
|
|
);
|
|
} else if !ds_aspects.contains(hal::FormatAspects::DEPTH) {
|
|
Self::add_pass_texture_init_actions(
|
|
&at.stencil,
|
|
texture_memory_actions,
|
|
view,
|
|
&mut pending_discard_init_fixups,
|
|
);
|
|
} else {
|
|
// This is the only place (anywhere in wgpu) where Stencil &
|
|
// Depth init state can diverge.
|
|
//
|
|
// To safe us the overhead of tracking init state of texture
|
|
// aspects everywhere, we're going to cheat a little bit in
|
|
// order to keep the init state of both Stencil and Depth
|
|
// aspects in sync. The expectation is that we hit this path
|
|
// extremely rarely!
|
|
//
|
|
// Diverging LoadOp, i.e. Load + Clear:
|
|
//
|
|
// Record MemoryInitKind::NeedsInitializedMemory for the entire
|
|
// surface, a bit wasteful on unit but no negative effect!
|
|
//
|
|
// Rationale: If the loaded channel is uninitialized it needs
|
|
// clearing, the cleared channel doesn't care. (If everything is
|
|
// already initialized nothing special happens)
|
|
//
|
|
// (possible minor optimization: Clear caused by
|
|
// NeedsInitializedMemory should know that it doesn't need to
|
|
// clear the aspect that was set to C)
|
|
let need_init_beforehand =
|
|
at.depth.load_op == LoadOp::Load || at.stencil.load_op == LoadOp::Load;
|
|
if need_init_beforehand {
|
|
pending_discard_init_fixups.extend(
|
|
texture_memory_actions.register_init_action(&TextureInitTrackerAction {
|
|
texture: view.parent.clone(),
|
|
range: TextureInitRange::from(view.selector.clone()),
|
|
kind: MemoryInitKind::NeedsInitializedMemory,
|
|
}),
|
|
);
|
|
}
|
|
|
|
// Diverging Store, i.e. Discard + Store:
|
|
//
|
|
// Immediately zero out channel that is set to discard after
|
|
// we're done with the render pass. This allows us to set the
|
|
// entire surface to MemoryInitKind::ImplicitlyInitialized (if
|
|
// it isn't already set to NeedsInitializedMemory).
|
|
//
|
|
// (possible optimization: Delay and potentially drop this zeroing)
|
|
if at.depth.store_op != at.stencil.store_op {
|
|
if !need_init_beforehand {
|
|
texture_memory_actions.register_implicit_init(
|
|
&view.parent,
|
|
TextureInitRange::from(view.selector.clone()),
|
|
);
|
|
}
|
|
divergent_discarded_depth_stencil_aspect = Some((
|
|
if at.depth.store_op == StoreOp::Discard {
|
|
wgt::TextureAspect::DepthOnly
|
|
} else {
|
|
wgt::TextureAspect::StencilOnly
|
|
},
|
|
view,
|
|
));
|
|
} else if at.depth.store_op == StoreOp::Discard {
|
|
// Both are discarded using the regular path.
|
|
discarded_surfaces.push(TextureSurfaceDiscard {
|
|
texture: view.parent.clone(),
|
|
mip_level: view.selector.mips.start,
|
|
layer: view.selector.layers.start,
|
|
});
|
|
}
|
|
}
|
|
|
|
(is_depth_read_only, is_stencil_read_only) = at.depth_stencil_read_only(ds_aspects)?;
|
|
|
|
let usage = if is_depth_read_only
|
|
&& is_stencil_read_only
|
|
&& device
|
|
.downlevel
|
|
.flags
|
|
.contains(wgt::DownlevelFlags::READ_ONLY_DEPTH_STENCIL)
|
|
{
|
|
hal::TextureUses::DEPTH_STENCIL_READ | hal::TextureUses::RESOURCE
|
|
} else {
|
|
hal::TextureUses::DEPTH_STENCIL_WRITE
|
|
};
|
|
render_attachments.push(view.to_render_attachment(usage));
|
|
|
|
depth_stencil = Some(hal::DepthStencilAttachment {
|
|
target: hal::Attachment {
|
|
view: view
|
|
.raw(snatch_guard)
|
|
.ok_or_else(|| RenderPassErrorInner::InvalidAttachment(view.info.id()))?,
|
|
usage,
|
|
},
|
|
depth_ops: at.depth.hal_ops(),
|
|
stencil_ops: at.stencil.hal_ops(),
|
|
clear_value: (at.depth.clear_value, at.stencil.clear_value),
|
|
});
|
|
}
|
|
|
|
for (index, attachment) in color_attachments.iter().enumerate() {
|
|
let at = if let Some(attachment) = attachment.as_ref() {
|
|
attachment
|
|
} else {
|
|
colors.push(None);
|
|
continue;
|
|
};
|
|
let color_view: &TextureView<A> = trackers
|
|
.views
|
|
.add_single(view_guard, at.view)
|
|
.ok_or(RenderPassErrorInner::InvalidAttachment(at.view))?;
|
|
check_multiview(color_view)?;
|
|
add_view(
|
|
color_view,
|
|
AttachmentErrorLocation::Color {
|
|
index,
|
|
resolve: false,
|
|
},
|
|
)?;
|
|
|
|
if !color_view
|
|
.desc
|
|
.aspects()
|
|
.contains(hal::FormatAspects::COLOR)
|
|
{
|
|
return Err(RenderPassErrorInner::ColorAttachment(
|
|
ColorAttachmentError::InvalidFormat(color_view.desc.format),
|
|
));
|
|
}
|
|
|
|
Self::add_pass_texture_init_actions(
|
|
&at.channel,
|
|
texture_memory_actions,
|
|
color_view,
|
|
&mut pending_discard_init_fixups,
|
|
);
|
|
render_attachments
|
|
.push(color_view.to_render_attachment(hal::TextureUses::COLOR_TARGET));
|
|
|
|
let mut hal_resolve_target = None;
|
|
if let Some(resolve_target) = at.resolve_target {
|
|
let resolve_view: &TextureView<A> = trackers
|
|
.views
|
|
.add_single(view_guard, resolve_target)
|
|
.ok_or(RenderPassErrorInner::InvalidAttachment(resolve_target))?;
|
|
|
|
check_multiview(resolve_view)?;
|
|
|
|
let resolve_location = AttachmentErrorLocation::Color {
|
|
index,
|
|
resolve: true,
|
|
};
|
|
|
|
let render_extent = resolve_view.render_extent.map_err(|reason| {
|
|
RenderPassErrorInner::TextureViewIsNotRenderable {
|
|
location: resolve_location,
|
|
reason,
|
|
}
|
|
})?;
|
|
if color_view.render_extent.unwrap() != render_extent {
|
|
return Err(RenderPassErrorInner::AttachmentsDimensionMismatch {
|
|
expected_location: attachment_location,
|
|
expected_extent: extent.unwrap_or_default(),
|
|
actual_location: resolve_location,
|
|
actual_extent: render_extent,
|
|
});
|
|
}
|
|
if color_view.samples == 1 || resolve_view.samples != 1 {
|
|
return Err(RenderPassErrorInner::InvalidResolveSampleCounts {
|
|
location: resolve_location,
|
|
src: color_view.samples,
|
|
dst: resolve_view.samples,
|
|
});
|
|
}
|
|
if color_view.desc.format != resolve_view.desc.format {
|
|
return Err(RenderPassErrorInner::MismatchedResolveTextureFormat {
|
|
location: resolve_location,
|
|
src: color_view.desc.format,
|
|
dst: resolve_view.desc.format,
|
|
});
|
|
}
|
|
if !resolve_view
|
|
.format_features
|
|
.flags
|
|
.contains(wgt::TextureFormatFeatureFlags::MULTISAMPLE_RESOLVE)
|
|
{
|
|
return Err(RenderPassErrorInner::UnsupportedResolveTargetFormat {
|
|
location: resolve_location,
|
|
format: resolve_view.desc.format,
|
|
});
|
|
}
|
|
|
|
texture_memory_actions.register_implicit_init(
|
|
&resolve_view.parent,
|
|
TextureInitRange::from(resolve_view.selector.clone()),
|
|
);
|
|
render_attachments
|
|
.push(resolve_view.to_render_attachment(hal::TextureUses::COLOR_TARGET));
|
|
|
|
hal_resolve_target = Some(hal::Attachment {
|
|
view: resolve_view.raw(snatch_guard).ok_or_else(|| {
|
|
RenderPassErrorInner::InvalidResolveTarget(resolve_view.info.id())
|
|
})?,
|
|
usage: hal::TextureUses::COLOR_TARGET,
|
|
});
|
|
}
|
|
|
|
colors.push(Some(hal::ColorAttachment {
|
|
target: hal::Attachment {
|
|
view: color_view.raw(snatch_guard).ok_or_else(|| {
|
|
RenderPassErrorInner::InvalidAttachment(color_view.info.id())
|
|
})?,
|
|
usage: hal::TextureUses::COLOR_TARGET,
|
|
},
|
|
resolve_target: hal_resolve_target,
|
|
ops: at.channel.hal_ops(),
|
|
clear_value: at.channel.clear_value,
|
|
}));
|
|
}
|
|
|
|
let extent = extent.ok_or(RenderPassErrorInner::MissingAttachments)?;
|
|
let multiview = detected_multiview.expect("Multiview was not detected, no attachments");
|
|
|
|
let view_data = AttachmentData {
|
|
colors: color_attachments
|
|
.iter()
|
|
.map(|at| at.as_ref().map(|at| view_guard.get(at.view).unwrap()))
|
|
.collect(),
|
|
resolves: color_attachments
|
|
.iter()
|
|
.filter_map(|at| match *at {
|
|
Some(RenderPassColorAttachment {
|
|
resolve_target: Some(resolve),
|
|
..
|
|
}) => Some(view_guard.get(resolve).unwrap()),
|
|
_ => None,
|
|
})
|
|
.collect(),
|
|
depth_stencil: depth_stencil_attachment.map(|at| view_guard.get(at.view).unwrap()),
|
|
};
|
|
|
|
let context = RenderPassContext {
|
|
attachments: view_data.map(|view| view.desc.format),
|
|
sample_count,
|
|
multiview,
|
|
};
|
|
|
|
let timestamp_writes = if let Some(tw) = timestamp_writes {
|
|
let query_set = trackers
|
|
.query_sets
|
|
.add_single(query_set_guard, tw.query_set)
|
|
.ok_or(RenderPassErrorInner::InvalidQuerySet(tw.query_set))?;
|
|
|
|
if let Some(index) = tw.beginning_of_pass_write_index {
|
|
pending_query_resets.use_query_set(tw.query_set, query_set, index);
|
|
}
|
|
if let Some(index) = tw.end_of_pass_write_index {
|
|
pending_query_resets.use_query_set(tw.query_set, query_set, index);
|
|
}
|
|
|
|
Some(hal::RenderPassTimestampWrites {
|
|
query_set: query_set.raw.as_ref().unwrap(),
|
|
beginning_of_pass_write_index: tw.beginning_of_pass_write_index,
|
|
end_of_pass_write_index: tw.end_of_pass_write_index,
|
|
})
|
|
} else {
|
|
None
|
|
};
|
|
|
|
let occlusion_query_set = if let Some(occlusion_query_set) = occlusion_query_set {
|
|
let query_set = trackers
|
|
.query_sets
|
|
.add_single(query_set_guard, occlusion_query_set)
|
|
.ok_or(RenderPassErrorInner::InvalidQuerySet(occlusion_query_set))?;
|
|
|
|
Some(query_set.raw.as_ref().unwrap())
|
|
} else {
|
|
None
|
|
};
|
|
|
|
let hal_desc = hal::RenderPassDescriptor {
|
|
label: hal_label(label, device.instance_flags),
|
|
extent,
|
|
sample_count,
|
|
color_attachments: &colors,
|
|
depth_stencil_attachment: depth_stencil,
|
|
multiview,
|
|
timestamp_writes,
|
|
occlusion_query_set,
|
|
};
|
|
unsafe {
|
|
encoder.raw.begin_render_pass(&hal_desc);
|
|
};
|
|
|
|
Ok(Self {
|
|
context,
|
|
usage_scope: UsageScope::new(buffer_guard, texture_guard),
|
|
render_attachments,
|
|
is_depth_read_only,
|
|
is_stencil_read_only,
|
|
extent,
|
|
_phantom: PhantomData,
|
|
pending_discard_init_fixups,
|
|
divergent_discarded_depth_stencil_aspect,
|
|
multiview,
|
|
})
|
|
}
|
|
|
|
fn finish(
|
|
mut self,
|
|
raw: &mut A::CommandEncoder,
|
|
snatch_guard: &SnatchGuard,
|
|
) -> Result<(UsageScope<A>, SurfacesInDiscardState<A>), RenderPassErrorInner> {
|
|
profiling::scope!("RenderPassInfo::finish");
|
|
unsafe {
|
|
raw.end_render_pass();
|
|
}
|
|
|
|
for ra in self.render_attachments {
|
|
let texture = &ra.texture;
|
|
check_texture_usage(texture.desc.usage, TextureUsages::RENDER_ATTACHMENT)?;
|
|
|
|
// the tracker set of the pass is always in "extend" mode
|
|
unsafe {
|
|
self.usage_scope
|
|
.textures
|
|
.merge_single(texture, Some(ra.selector.clone()), ra.usage)
|
|
.map_err(UsageConflict::from)?
|
|
};
|
|
}
|
|
|
|
// If either only stencil or depth was discarded, we put in a special
|
|
// clear pass to keep the init status of the aspects in sync. We do this
|
|
// so we don't need to track init state for depth/stencil aspects
|
|
// individually.
|
|
//
|
|
// Note that we don't go the usual route of "brute force" initializing
|
|
// the texture when need arises here, since this path is actually
|
|
// something a user may genuinely want (where as the other cases are
|
|
// more seen along the lines as gracefully handling a user error).
|
|
if let Some((aspect, view)) = self.divergent_discarded_depth_stencil_aspect {
|
|
let (depth_ops, stencil_ops) = if aspect == wgt::TextureAspect::DepthOnly {
|
|
(
|
|
hal::AttachmentOps::STORE, // clear depth
|
|
hal::AttachmentOps::LOAD | hal::AttachmentOps::STORE, // unchanged stencil
|
|
)
|
|
} else {
|
|
(
|
|
hal::AttachmentOps::LOAD | hal::AttachmentOps::STORE, // unchanged stencil
|
|
hal::AttachmentOps::STORE, // clear depth
|
|
)
|
|
};
|
|
let desc = hal::RenderPassDescriptor {
|
|
label: Some("(wgpu internal) Zero init discarded depth/stencil aspect"),
|
|
extent: view.render_extent.unwrap(),
|
|
sample_count: view.samples,
|
|
color_attachments: &[],
|
|
depth_stencil_attachment: Some(hal::DepthStencilAttachment {
|
|
target: hal::Attachment {
|
|
view: view.raw(snatch_guard).ok_or_else(|| {
|
|
RenderPassErrorInner::InvalidAttachment(view.info.id())
|
|
})?,
|
|
usage: hal::TextureUses::DEPTH_STENCIL_WRITE,
|
|
},
|
|
depth_ops,
|
|
stencil_ops,
|
|
clear_value: (0.0, 0),
|
|
}),
|
|
multiview: self.multiview,
|
|
timestamp_writes: None,
|
|
occlusion_query_set: None,
|
|
};
|
|
unsafe {
|
|
raw.begin_render_pass(&desc);
|
|
raw.end_render_pass();
|
|
}
|
|
}
|
|
|
|
Ok((self.usage_scope, self.pending_discard_init_fixups))
|
|
}
|
|
}
|
|
|
|
// Common routines between render/compute
|
|
|
|
impl<G: GlobalIdentityHandlerFactory> Global<G> {
|
|
pub fn command_encoder_run_render_pass<A: HalApi>(
|
|
&self,
|
|
encoder_id: id::CommandEncoderId,
|
|
pass: &RenderPass,
|
|
) -> Result<(), RenderPassError> {
|
|
self.command_encoder_run_render_pass_impl::<A>(
|
|
encoder_id,
|
|
pass.base.as_ref(),
|
|
&pass.color_targets,
|
|
pass.depth_stencil_target.as_ref(),
|
|
pass.timestamp_writes.as_ref(),
|
|
pass.occlusion_query_set_id,
|
|
)
|
|
}
|
|
|
|
#[doc(hidden)]
|
|
pub fn command_encoder_run_render_pass_impl<A: HalApi>(
|
|
&self,
|
|
encoder_id: id::CommandEncoderId,
|
|
base: BasePassRef<RenderCommand>,
|
|
color_attachments: &[Option<RenderPassColorAttachment>],
|
|
depth_stencil_attachment: Option<&RenderPassDepthStencilAttachment>,
|
|
timestamp_writes: Option<&RenderPassTimestampWrites>,
|
|
occlusion_query_set_id: Option<id::QuerySetId>,
|
|
) -> Result<(), RenderPassError> {
|
|
profiling::scope!(
|
|
"CommandEncoder::run_render_pass {}",
|
|
base.label.unwrap_or("")
|
|
);
|
|
|
|
let discard_hal_labels = self
|
|
.instance
|
|
.flags
|
|
.contains(wgt::InstanceFlags::DISCARD_HAL_LABELS);
|
|
let label = hal_label(base.label, self.instance.flags);
|
|
|
|
let pass_scope = PassErrorScope::Pass(encoder_id);
|
|
|
|
let hub = A::hub(self);
|
|
|
|
let cmd_buf = CommandBuffer::get_encoder(hub, encoder_id).map_pass_err(pass_scope)?;
|
|
let device = &cmd_buf.device;
|
|
let snatch_guard = device.snatchable_lock.read();
|
|
|
|
let (scope, pending_discard_init_fixups) = {
|
|
let mut cmd_buf_data = cmd_buf.data.lock();
|
|
let cmd_buf_data = cmd_buf_data.as_mut().unwrap();
|
|
|
|
#[cfg(feature = "trace")]
|
|
if let Some(ref mut list) = cmd_buf_data.commands {
|
|
list.push(crate::device::trace::Command::RunRenderPass {
|
|
base: BasePass::from_ref(base),
|
|
target_colors: color_attachments.to_vec(),
|
|
target_depth_stencil: depth_stencil_attachment.cloned(),
|
|
timestamp_writes: timestamp_writes.cloned(),
|
|
occlusion_query_set_id,
|
|
});
|
|
}
|
|
|
|
if !device.is_valid() {
|
|
return Err(DeviceError::Lost).map_pass_err(pass_scope);
|
|
}
|
|
|
|
let encoder = &mut cmd_buf_data.encoder;
|
|
let status = &mut cmd_buf_data.status;
|
|
let tracker = &mut cmd_buf_data.trackers;
|
|
let buffer_memory_init_actions = &mut cmd_buf_data.buffer_memory_init_actions;
|
|
let texture_memory_actions = &mut cmd_buf_data.texture_memory_actions;
|
|
let pending_query_resets = &mut cmd_buf_data.pending_query_resets;
|
|
|
|
// We automatically keep extending command buffers over time, and because
|
|
// we want to insert a command buffer _before_ what we're about to record,
|
|
// we need to make sure to close the previous one.
|
|
encoder.close().map_pass_err(pass_scope)?;
|
|
// We will reset this to `Recording` if we succeed, acts as a fail-safe.
|
|
*status = CommandEncoderStatus::Error;
|
|
encoder.open_pass(label).map_pass_err(pass_scope)?;
|
|
|
|
let bundle_guard = hub.render_bundles.read();
|
|
let bind_group_guard = hub.bind_groups.read();
|
|
let render_pipeline_guard = hub.render_pipelines.read();
|
|
let query_set_guard = hub.query_sets.read();
|
|
let buffer_guard = hub.buffers.read();
|
|
let texture_guard = hub.textures.read();
|
|
let view_guard = hub.texture_views.read();
|
|
|
|
log::trace!(
|
|
"Encoding render pass begin in command buffer {:?}",
|
|
encoder_id
|
|
);
|
|
|
|
let mut info = RenderPassInfo::start(
|
|
device,
|
|
label,
|
|
color_attachments,
|
|
depth_stencil_attachment,
|
|
timestamp_writes,
|
|
occlusion_query_set_id,
|
|
encoder,
|
|
tracker,
|
|
texture_memory_actions,
|
|
pending_query_resets,
|
|
&*view_guard,
|
|
&*buffer_guard,
|
|
&*texture_guard,
|
|
&*query_set_guard,
|
|
&snatch_guard,
|
|
)
|
|
.map_pass_err(pass_scope)?;
|
|
|
|
tracker.set_size(
|
|
Some(&*buffer_guard),
|
|
Some(&*texture_guard),
|
|
Some(&*view_guard),
|
|
None,
|
|
Some(&*bind_group_guard),
|
|
None,
|
|
Some(&*render_pipeline_guard),
|
|
Some(&*bundle_guard),
|
|
Some(&*query_set_guard),
|
|
);
|
|
|
|
let raw = &mut encoder.raw;
|
|
|
|
let mut state = State {
|
|
pipeline_flags: PipelineFlags::empty(),
|
|
binder: Binder::new(),
|
|
blend_constant: OptionalState::Unused,
|
|
stencil_reference: 0,
|
|
pipeline: None,
|
|
index: IndexState::default(),
|
|
vertex: VertexState::default(),
|
|
debug_scope_depth: 0,
|
|
};
|
|
let mut temp_offsets = Vec::new();
|
|
let mut dynamic_offset_count = 0;
|
|
let mut string_offset = 0;
|
|
let mut active_query = None;
|
|
|
|
for command in base.commands {
|
|
match *command {
|
|
RenderCommand::SetBindGroup {
|
|
index,
|
|
num_dynamic_offsets,
|
|
bind_group_id,
|
|
} => {
|
|
api_log!("RenderPass::set_bind_group {index} {bind_group_id:?}");
|
|
|
|
let scope = PassErrorScope::SetBindGroup(bind_group_id);
|
|
let max_bind_groups = device.limits.max_bind_groups;
|
|
if index >= max_bind_groups {
|
|
return Err(RenderCommandError::BindGroupIndexOutOfRange {
|
|
index,
|
|
max: max_bind_groups,
|
|
})
|
|
.map_pass_err(scope);
|
|
}
|
|
|
|
temp_offsets.clear();
|
|
temp_offsets.extend_from_slice(
|
|
&base.dynamic_offsets
|
|
[dynamic_offset_count..dynamic_offset_count + num_dynamic_offsets],
|
|
);
|
|
dynamic_offset_count += num_dynamic_offsets;
|
|
|
|
let bind_group = tracker
|
|
.bind_groups
|
|
.add_single(&*bind_group_guard, bind_group_id)
|
|
.ok_or(RenderCommandError::InvalidBindGroup(bind_group_id))
|
|
.map_pass_err(scope)?;
|
|
|
|
if bind_group.device.as_info().id() != device.as_info().id() {
|
|
return Err(DeviceError::WrongDevice).map_pass_err(scope);
|
|
}
|
|
|
|
bind_group
|
|
.validate_dynamic_bindings(index, &temp_offsets, &cmd_buf.limits)
|
|
.map_pass_err(scope)?;
|
|
|
|
// merge the resource tracker in
|
|
unsafe {
|
|
info.usage_scope
|
|
.merge_bind_group(&bind_group.used)
|
|
.map_pass_err(scope)?;
|
|
}
|
|
//Note: stateless trackers are not merged: the lifetime reference
|
|
// is held to the bind group itself.
|
|
|
|
buffer_memory_init_actions.extend(
|
|
bind_group.used_buffer_ranges.iter().filter_map(|action| {
|
|
action
|
|
.buffer
|
|
.initialization_status
|
|
.read()
|
|
.check_action(action)
|
|
}),
|
|
);
|
|
for action in bind_group.used_texture_ranges.iter() {
|
|
info.pending_discard_init_fixups
|
|
.extend(texture_memory_actions.register_init_action(action));
|
|
}
|
|
|
|
let pipeline_layout = state.binder.pipeline_layout.clone();
|
|
let entries =
|
|
state
|
|
.binder
|
|
.assign_group(index as usize, bind_group, &temp_offsets);
|
|
if !entries.is_empty() && pipeline_layout.is_some() {
|
|
let pipeline_layout = pipeline_layout.as_ref().unwrap().raw();
|
|
for (i, e) in entries.iter().enumerate() {
|
|
if let Some(group) = e.group.as_ref() {
|
|
let raw_bg = group
|
|
.raw(&snatch_guard)
|
|
.ok_or(RenderPassErrorInner::InvalidBindGroup(i))
|
|
.map_pass_err(scope)?;
|
|
unsafe {
|
|
raw.set_bind_group(
|
|
pipeline_layout,
|
|
index + i as u32,
|
|
raw_bg,
|
|
&e.dynamic_offsets,
|
|
);
|
|
}
|
|
}
|
|
}
|
|
}
|
|
}
|
|
RenderCommand::SetPipeline(pipeline_id) => {
|
|
api_log!("RenderPass::set_pipeline {pipeline_id:?}");
|
|
|
|
let scope = PassErrorScope::SetPipelineRender(pipeline_id);
|
|
state.pipeline = Some(pipeline_id);
|
|
|
|
let pipeline: &pipeline::RenderPipeline<A> = tracker
|
|
.render_pipelines
|
|
.add_single(&*render_pipeline_guard, pipeline_id)
|
|
.ok_or(RenderCommandError::InvalidPipeline(pipeline_id))
|
|
.map_pass_err(scope)?;
|
|
|
|
if pipeline.device.as_info().id() != device.as_info().id() {
|
|
return Err(DeviceError::WrongDevice).map_pass_err(scope);
|
|
}
|
|
|
|
info.context
|
|
.check_compatible(
|
|
&pipeline.pass_context,
|
|
RenderPassCompatibilityCheckType::RenderPipeline,
|
|
)
|
|
.map_err(RenderCommandError::IncompatiblePipelineTargets)
|
|
.map_pass_err(scope)?;
|
|
|
|
state.pipeline_flags = pipeline.flags;
|
|
|
|
if (pipeline.flags.contains(PipelineFlags::WRITES_DEPTH)
|
|
&& info.is_depth_read_only)
|
|
|| (pipeline.flags.contains(PipelineFlags::WRITES_STENCIL)
|
|
&& info.is_stencil_read_only)
|
|
{
|
|
return Err(RenderCommandError::IncompatiblePipelineRods)
|
|
.map_pass_err(scope);
|
|
}
|
|
|
|
state
|
|
.blend_constant
|
|
.require(pipeline.flags.contains(PipelineFlags::BLEND_CONSTANT));
|
|
|
|
unsafe {
|
|
raw.set_render_pipeline(pipeline.raw());
|
|
}
|
|
|
|
if pipeline.flags.contains(PipelineFlags::STENCIL_REFERENCE) {
|
|
unsafe {
|
|
raw.set_stencil_reference(state.stencil_reference);
|
|
}
|
|
}
|
|
|
|
// Rebind resource
|
|
if state.binder.pipeline_layout.is_none()
|
|
|| !state
|
|
.binder
|
|
.pipeline_layout
|
|
.as_ref()
|
|
.unwrap()
|
|
.is_equal(&pipeline.layout)
|
|
{
|
|
let (start_index, entries) = state.binder.change_pipeline_layout(
|
|
&pipeline.layout,
|
|
&pipeline.late_sized_buffer_groups,
|
|
);
|
|
if !entries.is_empty() {
|
|
for (i, e) in entries.iter().enumerate() {
|
|
if let Some(group) = e.group.as_ref() {
|
|
let raw_bg = group
|
|
.raw(&snatch_guard)
|
|
.ok_or(RenderPassErrorInner::InvalidBindGroup(i))
|
|
.map_pass_err(scope)?;
|
|
unsafe {
|
|
raw.set_bind_group(
|
|
pipeline.layout.raw(),
|
|
start_index as u32 + i as u32,
|
|
raw_bg,
|
|
&e.dynamic_offsets,
|
|
);
|
|
}
|
|
}
|
|
}
|
|
}
|
|
|
|
// Clear push constant ranges
|
|
let non_overlapping = super::bind::compute_nonoverlapping_ranges(
|
|
&pipeline.layout.push_constant_ranges,
|
|
);
|
|
for range in non_overlapping {
|
|
let offset = range.range.start;
|
|
let size_bytes = range.range.end - offset;
|
|
super::push_constant_clear(
|
|
offset,
|
|
size_bytes,
|
|
|clear_offset, clear_data| unsafe {
|
|
raw.set_push_constants(
|
|
pipeline.layout.raw(),
|
|
range.stages,
|
|
clear_offset,
|
|
clear_data,
|
|
);
|
|
},
|
|
);
|
|
}
|
|
}
|
|
|
|
state.index.pipeline_format = pipeline.strip_index_format;
|
|
|
|
let vertex_steps_len = pipeline.vertex_steps.len();
|
|
state.vertex.buffers_required = vertex_steps_len as u32;
|
|
|
|
// Initialize each `vertex.inputs[i].step` from
|
|
// `pipeline.vertex_steps[i]`. Enlarge `vertex.inputs`
|
|
// as necessary to accomodate all slots in the
|
|
// pipeline. If `vertex.inputs` is longer, fill the
|
|
// extra entries with default `VertexStep`s.
|
|
while state.vertex.inputs.len() < vertex_steps_len {
|
|
state.vertex.inputs.push(VertexBufferState::EMPTY);
|
|
}
|
|
|
|
// This is worse as a `zip`, but it's close.
|
|
let mut steps = pipeline.vertex_steps.iter();
|
|
for input in state.vertex.inputs.iter_mut() {
|
|
input.step = steps.next().cloned().unwrap_or_default();
|
|
}
|
|
|
|
// Update vertex buffer limits.
|
|
state.vertex.update_limits();
|
|
}
|
|
RenderCommand::SetIndexBuffer {
|
|
buffer_id,
|
|
index_format,
|
|
offset,
|
|
size,
|
|
} => {
|
|
api_log!("RenderPass::set_index_buffer {buffer_id:?}");
|
|
|
|
let scope = PassErrorScope::SetIndexBuffer(buffer_id);
|
|
let buffer = info
|
|
.usage_scope
|
|
.buffers
|
|
.merge_single(&*buffer_guard, buffer_id, hal::BufferUses::INDEX)
|
|
.map_pass_err(scope)?;
|
|
|
|
if buffer.device.as_info().id() != device.as_info().id() {
|
|
return Err(DeviceError::WrongDevice).map_pass_err(scope);
|
|
}
|
|
|
|
check_buffer_usage(buffer.usage, BufferUsages::INDEX)
|
|
.map_pass_err(scope)?;
|
|
let buf_raw = buffer
|
|
.raw
|
|
.get(&snatch_guard)
|
|
.ok_or(RenderCommandError::DestroyedBuffer(buffer_id))
|
|
.map_pass_err(scope)?;
|
|
|
|
let end = match size {
|
|
Some(s) => offset + s.get(),
|
|
None => buffer.size,
|
|
};
|
|
state.index.bound_buffer_view = Some((buffer_id, offset..end));
|
|
|
|
state.index.format = Some(index_format);
|
|
state.index.update_limit();
|
|
|
|
buffer_memory_init_actions.extend(
|
|
buffer.initialization_status.read().create_action(
|
|
buffer,
|
|
offset..end,
|
|
MemoryInitKind::NeedsInitializedMemory,
|
|
),
|
|
);
|
|
|
|
let bb = hal::BufferBinding {
|
|
buffer: buf_raw,
|
|
offset,
|
|
size,
|
|
};
|
|
unsafe {
|
|
raw.set_index_buffer(bb, index_format);
|
|
}
|
|
}
|
|
RenderCommand::SetVertexBuffer {
|
|
slot,
|
|
buffer_id,
|
|
offset,
|
|
size,
|
|
} => {
|
|
api_log!("RenderPass::set_vertex_buffer {slot} {buffer_id:?}");
|
|
|
|
let scope = PassErrorScope::SetVertexBuffer(buffer_id);
|
|
let buffer = info
|
|
.usage_scope
|
|
.buffers
|
|
.merge_single(&*buffer_guard, buffer_id, hal::BufferUses::VERTEX)
|
|
.map_pass_err(scope)?;
|
|
|
|
if buffer.device.as_info().id() != device.as_info().id() {
|
|
return Err(DeviceError::WrongDevice).map_pass_err(scope);
|
|
}
|
|
|
|
let max_vertex_buffers = device.limits.max_vertex_buffers;
|
|
if slot >= max_vertex_buffers {
|
|
return Err(RenderCommandError::VertexBufferIndexOutOfRange {
|
|
index: slot,
|
|
max: max_vertex_buffers,
|
|
})
|
|
.map_pass_err(scope);
|
|
}
|
|
|
|
check_buffer_usage(buffer.usage, BufferUsages::VERTEX)
|
|
.map_pass_err(scope)?;
|
|
let buf_raw = buffer
|
|
.raw
|
|
.get(&snatch_guard)
|
|
.ok_or(RenderCommandError::DestroyedBuffer(buffer_id))
|
|
.map_pass_err(scope)?;
|
|
|
|
let empty_slots =
|
|
(1 + slot as usize).saturating_sub(state.vertex.inputs.len());
|
|
state
|
|
.vertex
|
|
.inputs
|
|
.extend(iter::repeat(VertexBufferState::EMPTY).take(empty_slots));
|
|
let vertex_state = &mut state.vertex.inputs[slot as usize];
|
|
//TODO: where are we checking that the offset is in bound?
|
|
vertex_state.total_size = match size {
|
|
Some(s) => s.get(),
|
|
None => buffer.size - offset,
|
|
};
|
|
vertex_state.bound = true;
|
|
|
|
buffer_memory_init_actions.extend(
|
|
buffer.initialization_status.read().create_action(
|
|
buffer,
|
|
offset..(offset + vertex_state.total_size),
|
|
MemoryInitKind::NeedsInitializedMemory,
|
|
),
|
|
);
|
|
|
|
let bb = hal::BufferBinding {
|
|
buffer: buf_raw,
|
|
offset,
|
|
size,
|
|
};
|
|
unsafe {
|
|
raw.set_vertex_buffer(slot, bb);
|
|
}
|
|
state.vertex.update_limits();
|
|
}
|
|
RenderCommand::SetBlendConstant(ref color) => {
|
|
api_log!("RenderPass::set_blend_constant");
|
|
|
|
state.blend_constant = OptionalState::Set;
|
|
let array = [
|
|
color.r as f32,
|
|
color.g as f32,
|
|
color.b as f32,
|
|
color.a as f32,
|
|
];
|
|
unsafe {
|
|
raw.set_blend_constants(&array);
|
|
}
|
|
}
|
|
RenderCommand::SetStencilReference(value) => {
|
|
api_log!("RenderPass::set_stencil_reference {value}");
|
|
|
|
state.stencil_reference = value;
|
|
if state
|
|
.pipeline_flags
|
|
.contains(PipelineFlags::STENCIL_REFERENCE)
|
|
{
|
|
unsafe {
|
|
raw.set_stencil_reference(value);
|
|
}
|
|
}
|
|
}
|
|
RenderCommand::SetViewport {
|
|
ref rect,
|
|
depth_min,
|
|
depth_max,
|
|
} => {
|
|
api_log!("RenderPass::set_viewport {rect:?}");
|
|
|
|
let scope = PassErrorScope::SetViewport;
|
|
if rect.x < 0.0
|
|
|| rect.y < 0.0
|
|
|| rect.w <= 0.0
|
|
|| rect.h <= 0.0
|
|
|| rect.x + rect.w > info.extent.width as f32
|
|
|| rect.y + rect.h > info.extent.height as f32
|
|
{
|
|
return Err(RenderCommandError::InvalidViewportRect(
|
|
*rect,
|
|
info.extent,
|
|
))
|
|
.map_pass_err(scope);
|
|
}
|
|
if !(0.0..=1.0).contains(&depth_min) || !(0.0..=1.0).contains(&depth_max) {
|
|
return Err(RenderCommandError::InvalidViewportDepth(
|
|
depth_min, depth_max,
|
|
))
|
|
.map_pass_err(scope);
|
|
}
|
|
let r = hal::Rect {
|
|
x: rect.x,
|
|
y: rect.y,
|
|
w: rect.w,
|
|
h: rect.h,
|
|
};
|
|
unsafe {
|
|
raw.set_viewport(&r, depth_min..depth_max);
|
|
}
|
|
}
|
|
RenderCommand::SetPushConstant {
|
|
stages,
|
|
offset,
|
|
size_bytes,
|
|
values_offset,
|
|
} => {
|
|
api_log!("RenderPass::set_push_constants");
|
|
|
|
let scope = PassErrorScope::SetPushConstant;
|
|
let values_offset = values_offset
|
|
.ok_or(RenderPassErrorInner::InvalidValuesOffset)
|
|
.map_pass_err(scope)?;
|
|
|
|
let end_offset_bytes = offset + size_bytes;
|
|
let values_end_offset =
|
|
(values_offset + size_bytes / wgt::PUSH_CONSTANT_ALIGNMENT) as usize;
|
|
let data_slice =
|
|
&base.push_constant_data[(values_offset as usize)..values_end_offset];
|
|
|
|
let pipeline_layout = state
|
|
.binder
|
|
.pipeline_layout
|
|
.as_ref()
|
|
.ok_or(DrawError::MissingPipeline)
|
|
.map_pass_err(scope)?;
|
|
|
|
pipeline_layout
|
|
.validate_push_constant_ranges(stages, offset, end_offset_bytes)
|
|
.map_err(RenderCommandError::from)
|
|
.map_pass_err(scope)?;
|
|
|
|
unsafe {
|
|
raw.set_push_constants(
|
|
pipeline_layout.raw(),
|
|
stages,
|
|
offset,
|
|
data_slice,
|
|
)
|
|
}
|
|
}
|
|
RenderCommand::SetScissor(ref rect) => {
|
|
api_log!("RenderPass::set_scissor_rect {rect:?}");
|
|
|
|
let scope = PassErrorScope::SetScissorRect;
|
|
if rect.x + rect.w > info.extent.width
|
|
|| rect.y + rect.h > info.extent.height
|
|
{
|
|
return Err(RenderCommandError::InvalidScissorRect(*rect, info.extent))
|
|
.map_pass_err(scope);
|
|
}
|
|
let r = hal::Rect {
|
|
x: rect.x,
|
|
y: rect.y,
|
|
w: rect.w,
|
|
h: rect.h,
|
|
};
|
|
unsafe {
|
|
raw.set_scissor_rect(&r);
|
|
}
|
|
}
|
|
RenderCommand::Draw {
|
|
vertex_count,
|
|
instance_count,
|
|
first_vertex,
|
|
first_instance,
|
|
} => {
|
|
api_log!(
|
|
"RenderPass::draw {vertex_count} {instance_count} {first_vertex} {first_instance}"
|
|
);
|
|
|
|
let indexed = false;
|
|
let scope = PassErrorScope::Draw {
|
|
indexed,
|
|
indirect: false,
|
|
pipeline: state.pipeline,
|
|
};
|
|
state.is_ready(indexed).map_pass_err(scope)?;
|
|
|
|
let last_vertex = first_vertex as u64 + vertex_count as u64;
|
|
let vertex_limit = state.vertex.vertex_limit;
|
|
if last_vertex > vertex_limit {
|
|
return Err(DrawError::VertexBeyondLimit {
|
|
last_vertex,
|
|
vertex_limit,
|
|
slot: state.vertex.vertex_limit_slot,
|
|
})
|
|
.map_pass_err(scope);
|
|
}
|
|
let last_instance = first_instance as u64 + instance_count as u64;
|
|
let instance_limit = state.vertex.instance_limit;
|
|
if last_instance > instance_limit {
|
|
return Err(DrawError::InstanceBeyondLimit {
|
|
last_instance,
|
|
instance_limit,
|
|
slot: state.vertex.instance_limit_slot,
|
|
})
|
|
.map_pass_err(scope);
|
|
}
|
|
|
|
unsafe {
|
|
if instance_count > 0 && vertex_count > 0 {
|
|
raw.draw(
|
|
first_vertex,
|
|
vertex_count,
|
|
first_instance,
|
|
instance_count,
|
|
);
|
|
}
|
|
}
|
|
}
|
|
RenderCommand::DrawIndexed {
|
|
index_count,
|
|
instance_count,
|
|
first_index,
|
|
base_vertex,
|
|
first_instance,
|
|
} => {
|
|
api_log!("RenderPass::draw_indexed {index_count} {instance_count} {first_index} {base_vertex} {first_instance}");
|
|
|
|
let indexed = true;
|
|
let scope = PassErrorScope::Draw {
|
|
indexed,
|
|
indirect: false,
|
|
pipeline: state.pipeline,
|
|
};
|
|
state.is_ready(indexed).map_pass_err(scope)?;
|
|
|
|
let last_index = first_index as u64 + index_count as u64;
|
|
let index_limit = state.index.limit;
|
|
if last_index > index_limit {
|
|
return Err(DrawError::IndexBeyondLimit {
|
|
last_index,
|
|
index_limit,
|
|
})
|
|
.map_pass_err(scope);
|
|
}
|
|
let last_instance = first_instance as u64 + instance_count as u64;
|
|
let instance_limit = state.vertex.instance_limit;
|
|
if last_instance > instance_limit {
|
|
return Err(DrawError::InstanceBeyondLimit {
|
|
last_instance,
|
|
instance_limit,
|
|
slot: state.vertex.instance_limit_slot,
|
|
})
|
|
.map_pass_err(scope);
|
|
}
|
|
|
|
unsafe {
|
|
if instance_count > 0 && index_count > 0 {
|
|
raw.draw_indexed(
|
|
first_index,
|
|
index_count,
|
|
base_vertex,
|
|
first_instance,
|
|
instance_count,
|
|
);
|
|
}
|
|
}
|
|
}
|
|
RenderCommand::MultiDrawIndirect {
|
|
buffer_id,
|
|
offset,
|
|
count,
|
|
indexed,
|
|
} => {
|
|
api_log!("RenderPass::draw_indirect (indexed:{indexed}) {buffer_id:?} {offset} {count:?}");
|
|
|
|
let scope = PassErrorScope::Draw {
|
|
indexed,
|
|
indirect: true,
|
|
pipeline: state.pipeline,
|
|
};
|
|
state.is_ready(indexed).map_pass_err(scope)?;
|
|
|
|
let stride = match indexed {
|
|
false => mem::size_of::<wgt::DrawIndirectArgs>(),
|
|
true => mem::size_of::<wgt::DrawIndexedIndirectArgs>(),
|
|
};
|
|
|
|
if count.is_some() {
|
|
device
|
|
.require_features(wgt::Features::MULTI_DRAW_INDIRECT)
|
|
.map_pass_err(scope)?;
|
|
}
|
|
device
|
|
.require_downlevel_flags(wgt::DownlevelFlags::INDIRECT_EXECUTION)
|
|
.map_pass_err(scope)?;
|
|
|
|
let indirect_buffer = info
|
|
.usage_scope
|
|
.buffers
|
|
.merge_single(&*buffer_guard, buffer_id, hal::BufferUses::INDIRECT)
|
|
.map_pass_err(scope)?;
|
|
check_buffer_usage(indirect_buffer.usage, BufferUsages::INDIRECT)
|
|
.map_pass_err(scope)?;
|
|
let indirect_raw = indirect_buffer
|
|
.raw
|
|
.get(&snatch_guard)
|
|
.ok_or(RenderCommandError::DestroyedBuffer(buffer_id))
|
|
.map_pass_err(scope)?;
|
|
|
|
let actual_count = count.map_or(1, |c| c.get());
|
|
|
|
let end_offset = offset + stride as u64 * actual_count as u64;
|
|
if end_offset > indirect_buffer.size {
|
|
return Err(RenderPassErrorInner::IndirectBufferOverrun {
|
|
count,
|
|
offset,
|
|
end_offset,
|
|
buffer_size: indirect_buffer.size,
|
|
})
|
|
.map_pass_err(scope);
|
|
}
|
|
|
|
buffer_memory_init_actions.extend(
|
|
indirect_buffer.initialization_status.read().create_action(
|
|
indirect_buffer,
|
|
offset..end_offset,
|
|
MemoryInitKind::NeedsInitializedMemory,
|
|
),
|
|
);
|
|
|
|
match indexed {
|
|
false => unsafe {
|
|
raw.draw_indirect(indirect_raw, offset, actual_count);
|
|
},
|
|
true => unsafe {
|
|
raw.draw_indexed_indirect(indirect_raw, offset, actual_count);
|
|
},
|
|
}
|
|
}
|
|
RenderCommand::MultiDrawIndirectCount {
|
|
buffer_id,
|
|
offset,
|
|
count_buffer_id,
|
|
count_buffer_offset,
|
|
max_count,
|
|
indexed,
|
|
} => {
|
|
api_log!("RenderPass::multi_draw_indirect_count (indexed:{indexed}) {buffer_id:?} {offset} {count_buffer_id:?} {count_buffer_offset:?} {max_count:?}");
|
|
|
|
let scope = PassErrorScope::Draw {
|
|
indexed,
|
|
indirect: true,
|
|
pipeline: state.pipeline,
|
|
};
|
|
state.is_ready(indexed).map_pass_err(scope)?;
|
|
|
|
let stride = match indexed {
|
|
false => mem::size_of::<wgt::DrawIndirectArgs>(),
|
|
true => mem::size_of::<wgt::DrawIndexedIndirectArgs>(),
|
|
} as u64;
|
|
|
|
device
|
|
.require_features(wgt::Features::MULTI_DRAW_INDIRECT_COUNT)
|
|
.map_pass_err(scope)?;
|
|
device
|
|
.require_downlevel_flags(wgt::DownlevelFlags::INDIRECT_EXECUTION)
|
|
.map_pass_err(scope)?;
|
|
|
|
let indirect_buffer = info
|
|
.usage_scope
|
|
.buffers
|
|
.merge_single(&*buffer_guard, buffer_id, hal::BufferUses::INDIRECT)
|
|
.map_pass_err(scope)?;
|
|
check_buffer_usage(indirect_buffer.usage, BufferUsages::INDIRECT)
|
|
.map_pass_err(scope)?;
|
|
let indirect_raw = indirect_buffer
|
|
.raw
|
|
.get(&snatch_guard)
|
|
.ok_or(RenderCommandError::DestroyedBuffer(buffer_id))
|
|
.map_pass_err(scope)?;
|
|
|
|
let count_buffer = info
|
|
.usage_scope
|
|
.buffers
|
|
.merge_single(
|
|
&*buffer_guard,
|
|
count_buffer_id,
|
|
hal::BufferUses::INDIRECT,
|
|
)
|
|
.map_pass_err(scope)?;
|
|
check_buffer_usage(count_buffer.usage, BufferUsages::INDIRECT)
|
|
.map_pass_err(scope)?;
|
|
let count_raw = count_buffer
|
|
.raw
|
|
.get(&snatch_guard)
|
|
.ok_or(RenderCommandError::DestroyedBuffer(count_buffer_id))
|
|
.map_pass_err(scope)?;
|
|
|
|
let end_offset = offset + stride * max_count as u64;
|
|
if end_offset > indirect_buffer.size {
|
|
return Err(RenderPassErrorInner::IndirectBufferOverrun {
|
|
count: None,
|
|
offset,
|
|
end_offset,
|
|
buffer_size: indirect_buffer.size,
|
|
})
|
|
.map_pass_err(scope);
|
|
}
|
|
buffer_memory_init_actions.extend(
|
|
indirect_buffer.initialization_status.read().create_action(
|
|
indirect_buffer,
|
|
offset..end_offset,
|
|
MemoryInitKind::NeedsInitializedMemory,
|
|
),
|
|
);
|
|
|
|
let begin_count_offset = count_buffer_offset;
|
|
let end_count_offset = count_buffer_offset + 4;
|
|
if end_count_offset > count_buffer.size {
|
|
return Err(RenderPassErrorInner::IndirectCountBufferOverrun {
|
|
begin_count_offset,
|
|
end_count_offset,
|
|
count_buffer_size: count_buffer.size,
|
|
})
|
|
.map_pass_err(scope);
|
|
}
|
|
buffer_memory_init_actions.extend(
|
|
count_buffer.initialization_status.read().create_action(
|
|
count_buffer,
|
|
count_buffer_offset..end_count_offset,
|
|
MemoryInitKind::NeedsInitializedMemory,
|
|
),
|
|
);
|
|
|
|
match indexed {
|
|
false => unsafe {
|
|
raw.draw_indirect_count(
|
|
indirect_raw,
|
|
offset,
|
|
count_raw,
|
|
count_buffer_offset,
|
|
max_count,
|
|
);
|
|
},
|
|
true => unsafe {
|
|
raw.draw_indexed_indirect_count(
|
|
indirect_raw,
|
|
offset,
|
|
count_raw,
|
|
count_buffer_offset,
|
|
max_count,
|
|
);
|
|
},
|
|
}
|
|
}
|
|
RenderCommand::PushDebugGroup { color: _, len } => {
|
|
state.debug_scope_depth += 1;
|
|
if !discard_hal_labels {
|
|
let label = str::from_utf8(
|
|
&base.string_data[string_offset..string_offset + len],
|
|
)
|
|
.unwrap();
|
|
|
|
api_log!("RenderPass::push_debug_group {label:?}");
|
|
unsafe {
|
|
raw.begin_debug_marker(label);
|
|
}
|
|
}
|
|
string_offset += len;
|
|
}
|
|
RenderCommand::PopDebugGroup => {
|
|
api_log!("RenderPass::pop_debug_group");
|
|
|
|
let scope = PassErrorScope::PopDebugGroup;
|
|
if state.debug_scope_depth == 0 {
|
|
return Err(RenderPassErrorInner::InvalidPopDebugGroup)
|
|
.map_pass_err(scope);
|
|
}
|
|
state.debug_scope_depth -= 1;
|
|
if !discard_hal_labels {
|
|
unsafe {
|
|
raw.end_debug_marker();
|
|
}
|
|
}
|
|
}
|
|
RenderCommand::InsertDebugMarker { color: _, len } => {
|
|
if !discard_hal_labels {
|
|
let label = str::from_utf8(
|
|
&base.string_data[string_offset..string_offset + len],
|
|
)
|
|
.unwrap();
|
|
api_log!("RenderPass::insert_debug_marker {label:?}");
|
|
unsafe {
|
|
raw.insert_debug_marker(label);
|
|
}
|
|
}
|
|
string_offset += len;
|
|
}
|
|
RenderCommand::WriteTimestamp {
|
|
query_set_id,
|
|
query_index,
|
|
} => {
|
|
api_log!("RenderPass::write_timestamps {query_set_id:?} {query_index}");
|
|
let scope = PassErrorScope::WriteTimestamp;
|
|
|
|
device
|
|
.require_features(wgt::Features::TIMESTAMP_QUERY_INSIDE_PASSES)
|
|
.map_pass_err(scope)?;
|
|
|
|
let query_set = tracker
|
|
.query_sets
|
|
.add_single(&*query_set_guard, query_set_id)
|
|
.ok_or(RenderCommandError::InvalidQuerySet(query_set_id))
|
|
.map_pass_err(scope)?;
|
|
|
|
query_set
|
|
.validate_and_write_timestamp(
|
|
raw,
|
|
query_set_id,
|
|
query_index,
|
|
Some(&mut cmd_buf_data.pending_query_resets),
|
|
)
|
|
.map_pass_err(scope)?;
|
|
}
|
|
RenderCommand::BeginOcclusionQuery { query_index } => {
|
|
api_log!("RenderPass::begin_occlusion_query {query_index}");
|
|
let scope = PassErrorScope::BeginOcclusionQuery;
|
|
|
|
let query_set_id = occlusion_query_set_id
|
|
.ok_or(RenderPassErrorInner::MissingOcclusionQuerySet)
|
|
.map_pass_err(scope)?;
|
|
|
|
let query_set = tracker
|
|
.query_sets
|
|
.add_single(&*query_set_guard, query_set_id)
|
|
.ok_or(RenderCommandError::InvalidQuerySet(query_set_id))
|
|
.map_pass_err(scope)?;
|
|
|
|
query_set
|
|
.validate_and_begin_occlusion_query(
|
|
raw,
|
|
query_set_id,
|
|
query_index,
|
|
Some(&mut cmd_buf_data.pending_query_resets),
|
|
&mut active_query,
|
|
)
|
|
.map_pass_err(scope)?;
|
|
}
|
|
RenderCommand::EndOcclusionQuery => {
|
|
api_log!("RenderPass::end_occlusion_query");
|
|
let scope = PassErrorScope::EndOcclusionQuery;
|
|
|
|
end_occlusion_query(raw, &*query_set_guard, &mut active_query)
|
|
.map_pass_err(scope)?;
|
|
}
|
|
RenderCommand::BeginPipelineStatisticsQuery {
|
|
query_set_id,
|
|
query_index,
|
|
} => {
|
|
api_log!("RenderPass::begin_pipeline_statistics_query {query_set_id:?} {query_index}");
|
|
let scope = PassErrorScope::BeginPipelineStatisticsQuery;
|
|
|
|
let query_set = tracker
|
|
.query_sets
|
|
.add_single(&*query_set_guard, query_set_id)
|
|
.ok_or(RenderCommandError::InvalidQuerySet(query_set_id))
|
|
.map_pass_err(scope)?;
|
|
|
|
query_set
|
|
.validate_and_begin_pipeline_statistics_query(
|
|
raw,
|
|
query_set_id,
|
|
query_index,
|
|
Some(&mut cmd_buf_data.pending_query_resets),
|
|
&mut active_query,
|
|
)
|
|
.map_pass_err(scope)?;
|
|
}
|
|
RenderCommand::EndPipelineStatisticsQuery => {
|
|
api_log!("RenderPass::end_pipeline_statistics_query");
|
|
let scope = PassErrorScope::EndPipelineStatisticsQuery;
|
|
|
|
end_pipeline_statistics_query(raw, &*query_set_guard, &mut active_query)
|
|
.map_pass_err(scope)?;
|
|
}
|
|
RenderCommand::ExecuteBundle(bundle_id) => {
|
|
api_log!("RenderPass::execute_bundle {bundle_id:?}");
|
|
let scope = PassErrorScope::ExecuteBundle;
|
|
let bundle: &command::RenderBundle<A> = tracker
|
|
.bundles
|
|
.add_single(&*bundle_guard, bundle_id)
|
|
.ok_or(RenderCommandError::InvalidRenderBundle(bundle_id))
|
|
.map_pass_err(scope)?;
|
|
|
|
if bundle.device.as_info().id() != device.as_info().id() {
|
|
return Err(DeviceError::WrongDevice).map_pass_err(scope);
|
|
}
|
|
|
|
info.context
|
|
.check_compatible(
|
|
&bundle.context,
|
|
RenderPassCompatibilityCheckType::RenderBundle,
|
|
)
|
|
.map_err(RenderPassErrorInner::IncompatibleBundleTargets)
|
|
.map_pass_err(scope)?;
|
|
|
|
if (info.is_depth_read_only && !bundle.is_depth_read_only)
|
|
|| (info.is_stencil_read_only && !bundle.is_stencil_read_only)
|
|
{
|
|
return Err(
|
|
RenderPassErrorInner::IncompatibleBundleReadOnlyDepthStencil {
|
|
pass_depth: info.is_depth_read_only,
|
|
pass_stencil: info.is_stencil_read_only,
|
|
bundle_depth: bundle.is_depth_read_only,
|
|
bundle_stencil: bundle.is_stencil_read_only,
|
|
},
|
|
)
|
|
.map_pass_err(scope);
|
|
}
|
|
|
|
buffer_memory_init_actions.extend(
|
|
bundle
|
|
.buffer_memory_init_actions
|
|
.iter()
|
|
.filter_map(|action| {
|
|
action
|
|
.buffer
|
|
.initialization_status
|
|
.read()
|
|
.check_action(action)
|
|
}),
|
|
);
|
|
for action in bundle.texture_memory_init_actions.iter() {
|
|
info.pending_discard_init_fixups
|
|
.extend(texture_memory_actions.register_init_action(action));
|
|
}
|
|
|
|
unsafe { bundle.execute(raw) }
|
|
.map_err(|e| match e {
|
|
ExecutionError::DestroyedBuffer(id) => {
|
|
RenderCommandError::DestroyedBuffer(id)
|
|
}
|
|
ExecutionError::InvalidBindGroup(id) => {
|
|
RenderCommandError::InvalidBindGroup(id)
|
|
}
|
|
ExecutionError::Unimplemented(what) => {
|
|
RenderCommandError::Unimplemented(what)
|
|
}
|
|
})
|
|
.map_pass_err(scope)?;
|
|
|
|
unsafe {
|
|
info.usage_scope
|
|
.merge_render_bundle(&bundle.used)
|
|
.map_pass_err(scope)?;
|
|
tracker
|
|
.add_from_render_bundle(&bundle.used)
|
|
.map_pass_err(scope)?;
|
|
};
|
|
state.reset_bundle();
|
|
}
|
|
}
|
|
}
|
|
|
|
log::trace!("Merging renderpass into cmd_buf {:?}", encoder_id);
|
|
let (trackers, pending_discard_init_fixups) =
|
|
info.finish(raw, &snatch_guard).map_pass_err(pass_scope)?;
|
|
|
|
encoder.close().map_pass_err(pass_scope)?;
|
|
(trackers, pending_discard_init_fixups)
|
|
};
|
|
|
|
let cmd_buf = hub.command_buffers.get(encoder_id.transmute()).unwrap();
|
|
let mut cmd_buf_data = cmd_buf.data.lock();
|
|
let cmd_buf_data = cmd_buf_data.as_mut().unwrap();
|
|
|
|
let query_set_guard = hub.query_sets.read();
|
|
|
|
let encoder = &mut cmd_buf_data.encoder;
|
|
let status = &mut cmd_buf_data.status;
|
|
let tracker = &mut cmd_buf_data.trackers;
|
|
|
|
{
|
|
let transit = encoder.open().map_pass_err(pass_scope)?;
|
|
|
|
fixup_discarded_surfaces(
|
|
pending_discard_init_fixups.into_iter(),
|
|
transit,
|
|
&mut tracker.textures,
|
|
&cmd_buf.device,
|
|
);
|
|
|
|
cmd_buf_data
|
|
.pending_query_resets
|
|
.reset_queries(
|
|
transit,
|
|
&query_set_guard,
|
|
cmd_buf.device.info.id().backend(),
|
|
)
|
|
.map_err(RenderCommandError::InvalidQuerySet)
|
|
.map_pass_err(PassErrorScope::QueryReset)?;
|
|
|
|
super::CommandBuffer::insert_barriers_from_scope(
|
|
transit,
|
|
tracker,
|
|
&scope,
|
|
&snatch_guard,
|
|
);
|
|
}
|
|
|
|
*status = CommandEncoderStatus::Recording;
|
|
encoder.close_and_swap().map_pass_err(pass_scope)?;
|
|
|
|
Ok(())
|
|
}
|
|
}
|
|
|
|
pub mod render_ffi {
|
|
use super::{
|
|
super::{Rect, RenderCommand},
|
|
RenderPass,
|
|
};
|
|
use crate::{id, RawString};
|
|
use std::{convert::TryInto, ffi, num::NonZeroU32, slice};
|
|
use wgt::{BufferAddress, BufferSize, Color, DynamicOffset, IndexFormat};
|
|
|
|
/// # Safety
|
|
///
|
|
/// This function is unsafe as there is no guarantee that the given pointer is
|
|
/// valid for `offset_length` elements.
|
|
#[no_mangle]
|
|
pub unsafe extern "C" fn wgpu_render_pass_set_bind_group(
|
|
pass: &mut RenderPass,
|
|
index: u32,
|
|
bind_group_id: id::BindGroupId,
|
|
offsets: *const DynamicOffset,
|
|
offset_length: usize,
|
|
) {
|
|
let redundant = unsafe {
|
|
pass.current_bind_groups.set_and_check_redundant(
|
|
bind_group_id,
|
|
index,
|
|
&mut pass.base.dynamic_offsets,
|
|
offsets,
|
|
offset_length,
|
|
)
|
|
};
|
|
|
|
if redundant {
|
|
return;
|
|
}
|
|
|
|
pass.base.commands.push(RenderCommand::SetBindGroup {
|
|
index,
|
|
num_dynamic_offsets: offset_length,
|
|
bind_group_id,
|
|
});
|
|
}
|
|
|
|
#[no_mangle]
|
|
pub extern "C" fn wgpu_render_pass_set_pipeline(
|
|
pass: &mut RenderPass,
|
|
pipeline_id: id::RenderPipelineId,
|
|
) {
|
|
if pass.current_pipeline.set_and_check_redundant(pipeline_id) {
|
|
return;
|
|
}
|
|
|
|
pass.base
|
|
.commands
|
|
.push(RenderCommand::SetPipeline(pipeline_id));
|
|
}
|
|
|
|
#[no_mangle]
|
|
pub extern "C" fn wgpu_render_pass_set_vertex_buffer(
|
|
pass: &mut RenderPass,
|
|
slot: u32,
|
|
buffer_id: id::BufferId,
|
|
offset: BufferAddress,
|
|
size: Option<BufferSize>,
|
|
) {
|
|
pass.base.commands.push(RenderCommand::SetVertexBuffer {
|
|
slot,
|
|
buffer_id,
|
|
offset,
|
|
size,
|
|
});
|
|
}
|
|
|
|
#[no_mangle]
|
|
pub extern "C" fn wgpu_render_pass_set_index_buffer(
|
|
pass: &mut RenderPass,
|
|
buffer: id::BufferId,
|
|
index_format: IndexFormat,
|
|
offset: BufferAddress,
|
|
size: Option<BufferSize>,
|
|
) {
|
|
pass.set_index_buffer(buffer, index_format, offset, size);
|
|
}
|
|
|
|
#[no_mangle]
|
|
pub extern "C" fn wgpu_render_pass_set_blend_constant(pass: &mut RenderPass, color: &Color) {
|
|
pass.base
|
|
.commands
|
|
.push(RenderCommand::SetBlendConstant(*color));
|
|
}
|
|
|
|
#[no_mangle]
|
|
pub extern "C" fn wgpu_render_pass_set_stencil_reference(pass: &mut RenderPass, value: u32) {
|
|
pass.base
|
|
.commands
|
|
.push(RenderCommand::SetStencilReference(value));
|
|
}
|
|
|
|
#[no_mangle]
|
|
pub extern "C" fn wgpu_render_pass_set_viewport(
|
|
pass: &mut RenderPass,
|
|
x: f32,
|
|
y: f32,
|
|
w: f32,
|
|
h: f32,
|
|
depth_min: f32,
|
|
depth_max: f32,
|
|
) {
|
|
pass.base.commands.push(RenderCommand::SetViewport {
|
|
rect: Rect { x, y, w, h },
|
|
depth_min,
|
|
depth_max,
|
|
});
|
|
}
|
|
|
|
#[no_mangle]
|
|
pub extern "C" fn wgpu_render_pass_set_scissor_rect(
|
|
pass: &mut RenderPass,
|
|
x: u32,
|
|
y: u32,
|
|
w: u32,
|
|
h: u32,
|
|
) {
|
|
pass.base
|
|
.commands
|
|
.push(RenderCommand::SetScissor(Rect { x, y, w, h }));
|
|
}
|
|
|
|
/// # Safety
|
|
///
|
|
/// This function is unsafe as there is no guarantee that the given pointer is
|
|
/// valid for `size_bytes` bytes.
|
|
#[no_mangle]
|
|
pub unsafe extern "C" fn wgpu_render_pass_set_push_constants(
|
|
pass: &mut RenderPass,
|
|
stages: wgt::ShaderStages,
|
|
offset: u32,
|
|
size_bytes: u32,
|
|
data: *const u8,
|
|
) {
|
|
assert_eq!(
|
|
offset & (wgt::PUSH_CONSTANT_ALIGNMENT - 1),
|
|
0,
|
|
"Push constant offset must be aligned to 4 bytes."
|
|
);
|
|
assert_eq!(
|
|
size_bytes & (wgt::PUSH_CONSTANT_ALIGNMENT - 1),
|
|
0,
|
|
"Push constant size must be aligned to 4 bytes."
|
|
);
|
|
let data_slice = unsafe { slice::from_raw_parts(data, size_bytes as usize) };
|
|
let value_offset = pass.base.push_constant_data.len().try_into().expect(
|
|
"Ran out of push constant space. Don't set 4gb of push constants per RenderPass.",
|
|
);
|
|
|
|
pass.base.push_constant_data.extend(
|
|
data_slice
|
|
.chunks_exact(wgt::PUSH_CONSTANT_ALIGNMENT as usize)
|
|
.map(|arr| u32::from_ne_bytes([arr[0], arr[1], arr[2], arr[3]])),
|
|
);
|
|
|
|
pass.base.commands.push(RenderCommand::SetPushConstant {
|
|
stages,
|
|
offset,
|
|
size_bytes,
|
|
values_offset: Some(value_offset),
|
|
});
|
|
}
|
|
|
|
#[no_mangle]
|
|
pub extern "C" fn wgpu_render_pass_draw(
|
|
pass: &mut RenderPass,
|
|
vertex_count: u32,
|
|
instance_count: u32,
|
|
first_vertex: u32,
|
|
first_instance: u32,
|
|
) {
|
|
pass.base.commands.push(RenderCommand::Draw {
|
|
vertex_count,
|
|
instance_count,
|
|
first_vertex,
|
|
first_instance,
|
|
});
|
|
}
|
|
|
|
#[no_mangle]
|
|
pub extern "C" fn wgpu_render_pass_draw_indexed(
|
|
pass: &mut RenderPass,
|
|
index_count: u32,
|
|
instance_count: u32,
|
|
first_index: u32,
|
|
base_vertex: i32,
|
|
first_instance: u32,
|
|
) {
|
|
pass.base.commands.push(RenderCommand::DrawIndexed {
|
|
index_count,
|
|
instance_count,
|
|
first_index,
|
|
base_vertex,
|
|
first_instance,
|
|
});
|
|
}
|
|
|
|
#[no_mangle]
|
|
pub extern "C" fn wgpu_render_pass_draw_indirect(
|
|
pass: &mut RenderPass,
|
|
buffer_id: id::BufferId,
|
|
offset: BufferAddress,
|
|
) {
|
|
pass.base.commands.push(RenderCommand::MultiDrawIndirect {
|
|
buffer_id,
|
|
offset,
|
|
count: None,
|
|
indexed: false,
|
|
});
|
|
}
|
|
|
|
#[no_mangle]
|
|
pub extern "C" fn wgpu_render_pass_draw_indexed_indirect(
|
|
pass: &mut RenderPass,
|
|
buffer_id: id::BufferId,
|
|
offset: BufferAddress,
|
|
) {
|
|
pass.base.commands.push(RenderCommand::MultiDrawIndirect {
|
|
buffer_id,
|
|
offset,
|
|
count: None,
|
|
indexed: true,
|
|
});
|
|
}
|
|
|
|
#[no_mangle]
|
|
pub extern "C" fn wgpu_render_pass_multi_draw_indirect(
|
|
pass: &mut RenderPass,
|
|
buffer_id: id::BufferId,
|
|
offset: BufferAddress,
|
|
count: u32,
|
|
) {
|
|
pass.base.commands.push(RenderCommand::MultiDrawIndirect {
|
|
buffer_id,
|
|
offset,
|
|
count: NonZeroU32::new(count),
|
|
indexed: false,
|
|
});
|
|
}
|
|
|
|
#[no_mangle]
|
|
pub extern "C" fn wgpu_render_pass_multi_draw_indexed_indirect(
|
|
pass: &mut RenderPass,
|
|
buffer_id: id::BufferId,
|
|
offset: BufferAddress,
|
|
count: u32,
|
|
) {
|
|
pass.base.commands.push(RenderCommand::MultiDrawIndirect {
|
|
buffer_id,
|
|
offset,
|
|
count: NonZeroU32::new(count),
|
|
indexed: true,
|
|
});
|
|
}
|
|
|
|
#[no_mangle]
|
|
pub extern "C" fn wgpu_render_pass_multi_draw_indirect_count(
|
|
pass: &mut RenderPass,
|
|
buffer_id: id::BufferId,
|
|
offset: BufferAddress,
|
|
count_buffer_id: id::BufferId,
|
|
count_buffer_offset: BufferAddress,
|
|
max_count: u32,
|
|
) {
|
|
pass.base
|
|
.commands
|
|
.push(RenderCommand::MultiDrawIndirectCount {
|
|
buffer_id,
|
|
offset,
|
|
count_buffer_id,
|
|
count_buffer_offset,
|
|
max_count,
|
|
indexed: false,
|
|
});
|
|
}
|
|
|
|
#[no_mangle]
|
|
pub extern "C" fn wgpu_render_pass_multi_draw_indexed_indirect_count(
|
|
pass: &mut RenderPass,
|
|
buffer_id: id::BufferId,
|
|
offset: BufferAddress,
|
|
count_buffer_id: id::BufferId,
|
|
count_buffer_offset: BufferAddress,
|
|
max_count: u32,
|
|
) {
|
|
pass.base
|
|
.commands
|
|
.push(RenderCommand::MultiDrawIndirectCount {
|
|
buffer_id,
|
|
offset,
|
|
count_buffer_id,
|
|
count_buffer_offset,
|
|
max_count,
|
|
indexed: true,
|
|
});
|
|
}
|
|
|
|
/// # Safety
|
|
///
|
|
/// This function is unsafe as there is no guarantee that the given `label`
|
|
/// is a valid null-terminated string.
|
|
#[no_mangle]
|
|
pub unsafe extern "C" fn wgpu_render_pass_push_debug_group(
|
|
pass: &mut RenderPass,
|
|
label: RawString,
|
|
color: u32,
|
|
) {
|
|
let bytes = unsafe { ffi::CStr::from_ptr(label) }.to_bytes();
|
|
pass.base.string_data.extend_from_slice(bytes);
|
|
|
|
pass.base.commands.push(RenderCommand::PushDebugGroup {
|
|
color,
|
|
len: bytes.len(),
|
|
});
|
|
}
|
|
|
|
#[no_mangle]
|
|
pub extern "C" fn wgpu_render_pass_pop_debug_group(pass: &mut RenderPass) {
|
|
pass.base.commands.push(RenderCommand::PopDebugGroup);
|
|
}
|
|
|
|
/// # Safety
|
|
///
|
|
/// This function is unsafe as there is no guarantee that the given `label`
|
|
/// is a valid null-terminated string.
|
|
#[no_mangle]
|
|
pub unsafe extern "C" fn wgpu_render_pass_insert_debug_marker(
|
|
pass: &mut RenderPass,
|
|
label: RawString,
|
|
color: u32,
|
|
) {
|
|
let bytes = unsafe { ffi::CStr::from_ptr(label) }.to_bytes();
|
|
pass.base.string_data.extend_from_slice(bytes);
|
|
|
|
pass.base.commands.push(RenderCommand::InsertDebugMarker {
|
|
color,
|
|
len: bytes.len(),
|
|
});
|
|
}
|
|
|
|
#[no_mangle]
|
|
pub extern "C" fn wgpu_render_pass_write_timestamp(
|
|
pass: &mut RenderPass,
|
|
query_set_id: id::QuerySetId,
|
|
query_index: u32,
|
|
) {
|
|
pass.base.commands.push(RenderCommand::WriteTimestamp {
|
|
query_set_id,
|
|
query_index,
|
|
});
|
|
}
|
|
|
|
#[no_mangle]
|
|
pub extern "C" fn wgpu_render_pass_begin_occlusion_query(
|
|
pass: &mut RenderPass,
|
|
query_index: u32,
|
|
) {
|
|
pass.base
|
|
.commands
|
|
.push(RenderCommand::BeginOcclusionQuery { query_index });
|
|
}
|
|
|
|
#[no_mangle]
|
|
pub extern "C" fn wgpu_render_pass_end_occlusion_query(pass: &mut RenderPass) {
|
|
pass.base.commands.push(RenderCommand::EndOcclusionQuery);
|
|
}
|
|
|
|
#[no_mangle]
|
|
pub extern "C" fn wgpu_render_pass_begin_pipeline_statistics_query(
|
|
pass: &mut RenderPass,
|
|
query_set_id: id::QuerySetId,
|
|
query_index: u32,
|
|
) {
|
|
pass.base
|
|
.commands
|
|
.push(RenderCommand::BeginPipelineStatisticsQuery {
|
|
query_set_id,
|
|
query_index,
|
|
});
|
|
}
|
|
|
|
#[no_mangle]
|
|
pub extern "C" fn wgpu_render_pass_end_pipeline_statistics_query(pass: &mut RenderPass) {
|
|
pass.base
|
|
.commands
|
|
.push(RenderCommand::EndPipelineStatisticsQuery);
|
|
}
|
|
|
|
/// # Safety
|
|
///
|
|
/// This function is unsafe as there is no guarantee that the given pointer is
|
|
/// valid for `render_bundle_ids_length` elements.
|
|
#[no_mangle]
|
|
pub unsafe extern "C" fn wgpu_render_pass_execute_bundles(
|
|
pass: &mut RenderPass,
|
|
render_bundle_ids: *const id::RenderBundleId,
|
|
render_bundle_ids_length: usize,
|
|
) {
|
|
for &bundle_id in
|
|
unsafe { slice::from_raw_parts(render_bundle_ids, render_bundle_ids_length) }
|
|
{
|
|
pass.base
|
|
.commands
|
|
.push(RenderCommand::ExecuteBundle(bundle_id));
|
|
}
|
|
pass.current_pipeline.reset();
|
|
pass.current_bind_groups.reset();
|
|
}
|
|
}
|