VisionOS 2 beta 5 ,unity textmesh shader errors
Reality Composer Pro
RSS for tagPrototype and produce content for AR experiences using Reality Composer Pro.
Post
Replies
Boosts
Views
Activity
I have a plane with a texture that was made in Blender & then exported using the Reality Converter from a USDC to a USDZ.
The translucency looks 100% translucent in RCP but it looks a bit like glass with some reflection in a Reality Kit scene in visionOS.
Is there a material setting that I need to change?
I’m encountering an issue with recording in my Unity game through Reality Composer Pro. When I attempt to record video or take screenshots, it results in a black screen once my game launches. Screenshots and videos outside my game record fine, but within the game, the recordings are just black.
Additionally, when using my headset, the display is distorted and only my right eye shows anything, while the left eye remains black.
Here are some specifics:
My game is developed in Unity.
I’m using all the betas: Xcode 16 beta, the new macOS beta, and VisionOS 2 beta.
In the attached screenshot, you can see an Apple UI overlay with a black screen behind it. However, when I’m in the headset, I actually see my game along with that UI overlay, so it seems like the game itself isn’t getting recorded.
Also, I noticed on the Apple webpage that they recommend using the Developer Capture feature in Reality Composer Pro for high-quality screenshots and app previews. However, I find that using Control Center for recording works pretty well despite the lower quality and foveated resolution. If I can’t get Reality Composer Pro to capture in 4K, is it still acceptable to use screenshots and record videos from the Control Center?
Has anyone encountered similar issues or have any insights on what might be causing this? And regarding the secondary question, I’d appreciate any guidance from Apple on the acceptability of using Control Center recordings as a fallback. Here's a video preview I made with Control Center recordings. Is this quality acceptable?
https://youtu.be/z4VIO7obNNg?si=2irqHEfeGjkNBUvb
I have a custom material using Shader Graph in Reality Composer Pro, and I am trying to rig up sliders to values to control the shader. I am able to read the values from the Shader Graph without a problem, and I can even update them when setting them from the LLDB command line and then getting the values back. But the changes are not reflected in the graphics. Is there some sort of update() method or something that is required to read the changed parameter values?
On a related note, I am trying to understand what the MaterialParameters.Handle property is and why one would access a MaterialParameter via the handle vs just the name.
I tested all variations. The checkboxes in Reality Composer Pro 2 (beta 4) in the Physics Body component:
are absolute and not parent-relative. Also, regardless of what I set the center of mass to:
it always rotates around the center of mass despite the local rotation being correctly at the center of origin (imported from Blender). Thus I can get the door to turn but never to swing because it always rotates around its center of mass.
Tell me if this is expected behaviour or if there is a simple way to make this work.
I have developed a code that initiates the Timeline in the Reality Composer Pro scene every 12.93 seconds.
RealityView { … }
.onAppear {
startTimer()
}
.onDisappear {
stopTimer()
}
func startTimer() {
timer = Timer.scheduledTimer(withTimeInterval: 12.93, repeats: true) { _ in
action()
}
}
func stopTimer() {
timer?.invalidate()
}
func action() {
print(“SunUpDown”)
NotificationCenter.default.post(
name: NSNotification.Name(“RealityKit.NotificationTrigger”),
object: nil,
userInfo: [
“RealityKit.NotificationTrigger.Scene”: scene as Any,
“RealityKit.NotificationTrigger.Identifier”: “SunUpDown”
]
)
}
Upon receiving the “SunUpDown” command, Timeline will be executed.
However, everything was functioning normally when I was running the scene, and I could continue looping until I attempted to zoom in on the window and discovered that it ceased looping. Could you please provide an explanation for this behavior?
Note: The window type is volumetric, and the parameter of the defaultWorldScaling modifier is dynamic.
Hi all,
I am fairly new to Swift development so go easy on me!
I am working through a few examples of using Reality Kit content within my projects and whilst trying to work on adding gestures to RealityKit entities, I have come across a weird issue.
Downloading and running the example here
This works fine for me.
When adding the same things to my own code - in this case a class called EntityGestureState to my GestureComponent file (within the reality kit project) I constantly get this error:
"Static property 'shared' is not concurrency-safe because it is non-isolated global shared mutable state"
Even just troubleshooting with something as simple as:
public class EntityGestureState {
// The entity currently being dragged if a gesture is in progress.
// Singleton shared instance
static let shared: EntityGestureState = EntityGestureState()
}
I immediately get the error and from a bunch of trial and error and reading different sources I can't seem to get around this.
Could anyone help here? I am running on Xcode 16 beta 3 so am wondering if it's a bug but also more than likely user-error.
In a scenario involving one of the entities in a Reality Composer Pro environment, I intend for this entity to display a blue material when viewed by the user. To achieve this, I have added the following Shader Graphs to the materials associated with this entity:
Additionally, I have included the HoverEffectComponent component to the Reality View in the code:
RealityView { content in
if let model = try? await Entity(named: “WorldScene”, in: realityKitContentBundle) {
let hoverEffect = HoverEffectComponent(.shader(.default))
model.components.set(hoverEffect)
content.add(model)
}
}
However, hover this entity, I am unable to observe any visual reaction. Could you please provide guidance on how to resolve this issue?
Ok, I am loading an object from a Reality Composer Pro scene that has two entities inside its hierarchy that both have a Physics Body and a Collision component like this
Root
Outer Box Mesh
Hinge + physics(static/kinematic) + collision
Door. + physics(dynamic) + collision
I tried to keep the physics/collision components only to the hinge and the door while I move the root or the outer box via code around. The behaviour I see is that it either
moves the hinge and the door around relative to the top level (despite me checking the movement locking) OR
starts rotating! the root or outer box even though I only set its position.
What is the correct setup in this case? What I want is that I can move the whole object around and settle it somewhere and still have the door pinned at a fixed relative position and have one degree freedom on the hinge axis.
I know how to do it in code but I really want to use the build in Reality Composer Pro settings/components. I am using the latest beta 4.
Hello all,
I'm developing an application for visionOS and I'm trying to implement 2 different animations:
First animation
Initially, I have a map that should not be visible. I would like to create an animation effect where it appears as if a drop of water falls in the center of the map and the expanding waves gradually reveal the entire map.
Is there a way to do it directly on SwiftUI or I need an animation on my USDZ?
Second animation
I want an animation effect similar to a cinema screen opening from the center, gradually revealing a video that was initially hidden.
Is there a way to do it directly on SwiftUI?
Can someone help me with this topic?
Thanks ;)
How to solve the problem of using Model3D to load a local model file in Unity project, clicking on NavigationLink multiple times to load the local model file, and receiving a prompt "assertion failure: 'stagingBuffer.buffer.isValid()' (createMetalBuffer:line 2971) Failed to create staging buffer for texture upload"?
How should I set the window of WindowGrop to resemble a curved screen style?
Baffled by the new ExtractBits shader graph node only supporting String input. Is this a bug? Trying to extract an integer from a float value, but have no idea how to pass it into Extract Bits. Convert nodes don't support number to string.
Hello. I am a designer developing a Vision Pro app. I have Two Problem in my App Develop Process.
I am trying to import free 3D national heritage content from Korea into Reality Composer Pro and place it in the app's internal space. However, there is an issue where the textures are not being imported correctly.
in Reality Composer Pro
in Simulator
In Reality Composer Pro, the textures are displayed correctly, but when I run the app on the Simulator in Xcode, the textures appear white and are not displayed properly. The content I imported is an .obj file, and I applied all the textures in jpg format using Reality Converter and exported it as a .usdz file, but the same issue persists.
I checked to see if the problem only occurs on the Simulator, but the same issue occurs on the Vision Pro device as well. How can I resolve this problem?
The following error code appears in Xcode, and the simulator does not run. I think it might be due to the size of the object added to the scene, so I tried compressing it with Reality Converter, but the issue still persists. Is there any other way to resolve this?
[MTLDebugDevice newBufferWithBytesNoCopy:length:options:deallocator:]:700: failed assertion Buffer Validation
newBufferWith*:length 0x280cc000 must not exceed 256 MB.
Hello! I’ve got a USDZ export from Maya pipeline working with animation, and they load up nicely in the Vision Pro.
I’ve been checking out the animated sample files in the Augmented Reality/Quick Loop sample page, specifically, the first three at the top of the page.
I would like to know how they are created. I’m a 3d modeler and animator, not a programmer, so dipping my toe in RCP and Xcode/SwiftUI, but could used some informative tutorials for proper workflow. For example, in the Lunar Rover sample, there are lines emanating from the model, then text windows appear. Would I need to create all these extras inside Reality Composer Pro? I’d like to start creating immersive, narrative experiences (both in a volume, and fully immersive) but for prototyping, I want to learn the proper way to add this type of functionality. I think I remember seeing something to do with “schemas” involved. I’m assuming there might be some coding to setup in RCP for when items are selected, then an associated animation is triggered. Can anyone point me towards the relevant documentation to help me get started? Remember, I don’t code. ;)
Here are my recent Vision Pro experimentations.
https://youtube.com/playlist?list=PLCH753rZ9r6eqXxpIemaSlcyYxjFgR210&si=P_7AY2aL97Upm61i
I’m also proficient with Unreal Engine, but getting content packaged and over to AVP is still not ready for prime time, so i’m exploring the native approach.
Thanks for helping point me in the right direction!
When I run my visionOS App, RealityKitContent Report an error:
Tool terminated by signal 'Segmentation fault: 11'
And it points to a USDZ model I imported, but in the scene, my model can be displayed normally and there is no damage. Why does an error occur? How can I check and repair it?
Hello everyone
I am looking to build a simple app for displaying a spatial video using the quick look preview API. I have been following this video which is useful:
https://developer.apple.com/videos/play/wwdc2024/10166/#:~:text=QuickLook%20is%20the%20system%20standard,just%20like%20the%20Photos%20app.
I am new to building apps in Xcode, and I could do with some advice on how to build the rest of the project mentioned in the above video. I was wondering if there is source code or a project example available anywhere for an app the uses the Quick Look preview API?
I'm using Reality Composer Pro Version 2.0 Version 2.0 (448.0.10.0.2) avaliable in Xcode_16_beta_4
When adding a animation from the Animation Library component on my armature to a timeline - the animation does not 'freeze' on the last frame.
Is there a way to 'freeze' the first or last frames when adding animations to the timeline? And how should I expect the first and last keys on my animations to behave with the default 'rest pose' on the imported usd file?
Hey there,
I'm wonder if any one knows how to make the shader graph which is shown in wwdc24 video. I tired couple things but couldn't get same result I could make it in Unity and Blender but not in RCP.
thank you.
https://developer.apple.com/wwdc24/10106
On visionOS 2 beta 3, Reality Composer Pro will open a cached copy of a scene (for example an usdc file I just changed) on the first try. Closing it and re-opening it will open the correct version.
Am I doing something wrong?