Reality Composer Pro

RSS for tag

Prototype and produce content for AR experiences using Reality Composer Pro.

Learn More

Post

Replies

Boosts

Views

Activity

Recording Issues in Unity Game with Reality Composer Pro - Black Screen
I’m encountering an issue with recording in my Unity game through Reality Composer Pro. When I attempt to record video or take screenshots, it results in a black screen once my game launches. Screenshots and videos outside my game record fine, but within the game, the recordings are just black. Additionally, when using my headset, the display is distorted and only my right eye shows anything, while the left eye remains black. Here are some specifics: My game is developed in Unity. I’m using all the betas: Xcode 16 beta, the new macOS beta, and VisionOS 2 beta. In the attached screenshot, you can see an Apple UI overlay with a black screen behind it. However, when I’m in the headset, I actually see my game along with that UI overlay, so it seems like the game itself isn’t getting recorded. Also, I noticed on the Apple webpage that they recommend using the Developer Capture feature in Reality Composer Pro for high-quality screenshots and app previews. However, I find that using Control Center for recording works pretty well despite the lower quality and foveated resolution. If I can’t get Reality Composer Pro to capture in 4K, is it still acceptable to use screenshots and record videos from the Control Center? Has anyone encountered similar issues or have any insights on what might be causing this? And regarding the secondary question, I’d appreciate any guidance from Apple on the acceptability of using Control Center recordings as a fallback. Here's a video preview I made with Control Center recordings. Is this quality acceptable? https://youtu.be/z4VIO7obNNg?si=2irqHEfeGjkNBUvb
0
0
352
Aug ’24
RealityKit ShaderGraphMaterial parameters in Reality Composer Pro
I have a custom material using Shader Graph in Reality Composer Pro, and I am trying to rig up sliders to values to control the shader. I am able to read the values from the Shader Graph without a problem, and I can even update them when setting them from the LLDB command line and then getting the values back. But the changes are not reflected in the graphics. Is there some sort of update() method or something that is required to read the changed parameter values? On a related note, I am trying to understand what the MaterialParameters.Handle property is and why one would access a MaterialParameter via the handle vs just the name.
1
0
442
Aug ’24
Is it true that without code, Reality Composer Pro cannot make a vertical door that swings with gravity?
I tested all variations. The checkboxes in Reality Composer Pro 2 (beta 4) in the Physics Body component: are absolute and not parent-relative. Also, regardless of what I set the center of mass to: it always rotates around the center of mass despite the local rotation being correctly at the center of origin (imported from Blender). Thus I can get the door to turn but never to swing because it always rotates around its center of mass. Tell me if this is expected behaviour or if there is a simple way to make this work.
8
0
638
Aug ’24
It cannot be looped after enlarging the window.
I have developed a code that initiates the Timeline in the Reality Composer Pro scene every 12.93 seconds. RealityView { … } .onAppear { startTimer() } .onDisappear { stopTimer() } func startTimer() { timer = Timer.scheduledTimer(withTimeInterval: 12.93, repeats: true) { _ in action() } } func stopTimer() { timer?.invalidate() } func action() { print(“SunUpDown”) NotificationCenter.default.post( name: NSNotification.Name(“RealityKit.NotificationTrigger”), object: nil, userInfo: [ “RealityKit.NotificationTrigger.Scene”: scene as Any, “RealityKit.NotificationTrigger.Identifier”: “SunUpDown” ] ) } Upon receiving the “SunUpDown” command, Timeline will be executed. However, everything was functioning normally when I was running the scene, and I could continue looping until I attempted to zoom in on the window and discovered that it ceased looping. Could you please provide an explanation for this behavior? Note: The window type is volumetric, and the parameter of the defaultWorldScaling modifier is dynamic.
1
0
386
Aug ’24
Static property 'shared' is not concurrency-safe because it is non-isolated global shared mutable state
Hi all, I am fairly new to Swift development so go easy on me! I am working through a few examples of using Reality Kit content within my projects and whilst trying to work on adding gestures to RealityKit entities, I have come across a weird issue. Downloading and running the example here This works fine for me. When adding the same things to my own code - in this case a class called EntityGestureState to my GestureComponent file (within the reality kit project) I constantly get this error: "Static property 'shared' is not concurrency-safe because it is non-isolated global shared mutable state" Even just troubleshooting with something as simple as: public class EntityGestureState { // The entity currently being dragged if a gesture is in progress. // Singleton shared instance static let shared: EntityGestureState = EntityGestureState() } I immediately get the error and from a bunch of trial and error and reading different sources I can't seem to get around this. Could anyone help here? I am running on Xcode 16 beta 3 so am wondering if it's a bug but also more than likely user-error.
2
1
1.1k
Aug ’24
HoverState in RealityView
In a scenario involving one of the entities in a Reality Composer Pro environment, I intend for this entity to display a blue material when viewed by the user. To achieve this, I have added the following Shader Graphs to the materials associated with this entity: Additionally, I have included the HoverEffectComponent component to the Reality View in the code: RealityView { content in if let model = try? await Entity(named: “WorldScene”, in: realityKitContentBundle) { let hoverEffect = HoverEffectComponent(.shader(.default)) model.components.set(hoverEffect) content.add(model) } } However, hover this entity, I am unable to observe any visual reaction. Could you please provide guidance on how to resolve this issue?
1
0
424
Aug ’24
Physics Body Components within a hierarchy behave weird
Ok, I am loading an object from a Reality Composer Pro scene that has two entities inside its hierarchy that both have a Physics Body and a Collision component like this Root Outer Box Mesh Hinge + physics(static/kinematic) + collision Door. + physics(dynamic) + collision I tried to keep the physics/collision components only to the hinge and the door while I move the root or the outer box via code around. The behaviour I see is that it either moves the hinge and the door around relative to the top level (despite me checking the movement locking) OR starts rotating! the root or outer box even though I only set its position. What is the correct setup in this case? What I want is that I can move the whole object around and settle it somewhere and still have the door pinned at a fixed relative position and have one degree freedom on the hinge axis. I know how to do it in code but I really want to use the build in Reality Composer Pro settings/components. I am using the latest beta 4.
0
0
270
Jul ’24
VisionOS animation on USDZ
Hello all, I'm developing an application for visionOS and I'm trying to implement 2 different animations: First animation Initially, I have a map that should not be visible. I would like to create an animation effect where it appears as if a drop of water falls in the center of the map and the expanding waves gradually reveal the entire map. Is there a way to do it directly on SwiftUI or I need an animation on my USDZ? Second animation I want an animation effect similar to a cinema screen opening from the center, gradually revealing a video that was initially hidden. Is there a way to do it directly on SwiftUI? Can someone help me with this topic? Thanks ;)
2
1
347
Jul ’24
The textures load correctly in Reality Composer Pro, but they do not appear in the Simulator or on the Vision Pro device
Hello. I am a designer developing a Vision Pro app. I have Two Problem in my App Develop Process. I am trying to import free 3D national heritage content from Korea into Reality Composer Pro and place it in the app's internal space. However, there is an issue where the textures are not being imported correctly. in Reality Composer Pro in Simulator In Reality Composer Pro, the textures are displayed correctly, but when I run the app on the Simulator in Xcode, the textures appear white and are not displayed properly. The content I imported is an .obj file, and I applied all the textures in jpg format using Reality Converter and exported it as a .usdz file, but the same issue persists. I checked to see if the problem only occurs on the Simulator, but the same issue occurs on the Vision Pro device as well. How can I resolve this problem? The following error code appears in Xcode, and the simulator does not run. I think it might be due to the size of the object added to the scene, so I tried compressing it with Reality Converter, but the issue still persists. Is there any other way to resolve this? [MTLDebugDevice newBufferWithBytesNoCopy:length:options:deallocator:]:700: failed assertion Buffer Validation newBufferWith*:length 0x280cc000 must not exceed 256 MB.
0
0
316
Jul ’24
How to create Lunar Rover USDZ Animated Sample File
Hello! I’ve got a USDZ export from Maya pipeline working with animation, and they load up nicely in the Vision Pro. I’ve been checking out the animated sample files in the Augmented Reality/Quick Loop sample page, specifically, the first three at the top of the page. I would like to know how they are created. I’m a 3d modeler and animator, not a programmer, so dipping my toe in RCP and Xcode/SwiftUI, but could used some informative tutorials for proper workflow. For example, in the Lunar Rover sample, there are lines emanating from the model, then text windows appear. Would I need to create all these extras inside Reality Composer Pro? I’d like to start creating immersive, narrative experiences (both in a volume, and fully immersive) but for prototyping, I want to learn the proper way to add this type of functionality. I think I remember seeing something to do with “schemas” involved. I’m assuming there might be some coding to setup in RCP for when items are selected, then an associated animation is triggered. Can anyone point me towards the relevant documentation to help me get started? Remember, I don’t code. ;) Here are my recent Vision Pro experimentations. https://youtube.com/playlist?list=PLCH753rZ9r6eqXxpIemaSlcyYxjFgR210&si=P_7AY2aL97Upm61i I’m also proficient with Unreal Engine, but getting content packaged and over to AVP is still not ready for prime time, so i’m exploring the native approach. Thanks for helping point me in the right direction!
0
0
282
Jul ’24
Help Building spatial video app using Quick Look preview
Hello everyone I am looking to build a simple app for displaying a spatial video using the quick look preview API. I have been following this video which is useful: https://developer.apple.com/videos/play/wwdc2024/10166/#:~:text=QuickLook%20is%20the%20system%20standard,just%20like%20the%20Photos%20app. I am new to building apps in Xcode, and I could do with some advice on how to build the rest of the project mentioned in the above video. I was wondering if there is source code or a project example available anywhere for an app the uses the Quick Look preview API?
0
0
336
Jul ’24
Timeline Animation in Reality Composer Pro
I'm using Reality Composer Pro Version 2.0 Version 2.0 (448.0.10.0.2) avaliable in Xcode_16_beta_4 When adding a animation from the Animation Library component on my armature to a timeline - the animation does not 'freeze' on the last frame. Is there a way to 'freeze' the first or last frames when adding animations to the timeline? And how should I expect the first and last keys on my animations to behave with the default 'rest pose' on the imported usd file?
1
0
566
Jul ’24