Post

Replies

Boosts

Views

Activity

Some question about visionOS
I have some question about visionOS: Does Apple open the eye tracking API to developers? If I want to know how to achieve it, when the eyes are stared at a specific View, a Boolean value can be changed to true, and when the eyes are removed from this View, it will become false. In ImmersiveSpace, when immersionStyle is .full or .progressive, a black background will appear by default. How can I turn this background into a panorama of my own? In ImmersiveSpace, how to make a View always follow the user?
2
0
703
Nov ’23
visionOS's "offset3D"
In visionOS, have many code with 3D attributes (SwiftUI) is adapted from the code in iOS, like: .padding(_:) to .padding3D(_:). In iOS have .offset(x:_, y:_), it only have X and Y, but in visionOS, view is in a 3D scene, so I want offset have Z, but offset can't use Z, so I try offset3D: import SwiftUI //Some View .offset3D(x: Number, y: Number, z: Number) Xcode report an error: Error: No member 'offset3D' So do you now how to use like offset's Modifiers, and can use Z in visionOS.
1
0
680
Nov ’23
Fixed background in visonOS
I successfully changed a picture to the background in ImmersiveSpace in a full state with the following code. import RealityKit struct MainBackground: View { var body: some View { RealityView { content in guard let resource = try? await TextureResource(named: "Image_Name") else { fatalError("Error.") } var material = UnlitMaterial() material.color = .init(texture: .init(resource)) let entity = Entity() entity.components.set(ModelComponent( mesh: .generateSphere(radius: 1000), materials: [material] )) entity.scale *= .init(x: -1, y: 1, z: 1) content.add(entity) } } } However, when running, I found that when the user moves, the background is not fixed, but follows the user's movement, which I feel unrealistic. How to fix the background in the place where it first appears, and give the user a kind of movement that is really like walking in the real world, instead of letting the background follow the user.
1
0
734
Dec ’23
Model3D and Reality Composer Pro
I created a RealityKitContent in the Packages folder of the visionOS app project. At first, I tried to add a USDA model directly to its rkassets. I used Model3D(named: "ModelName", bundle: realityKitContentBundle) can The model is displayed normally, but then when I add a folder in rkassets and then put the USDA model in that folder, use Model3D(named: "ModelName", bundle: realityKit ContentBundle) cannot display the model normally. What should I do? If you know how to solve the above problems, please let me know and check if you know how to solve the following problems. If you know, please also tell me the answer. Thank you! The USDA model I mentioned in the question just now contains an animation, but when I used Model3D(named: "ModelName", bundle: realityKitContentBundle) , I found that the animation was not played by default, but needed additional code. Is there any documentation, video or listing code in this regard?
0
0
706
Dec ’23
wwdc2023-10113 Video's manipulationState variable
Question 1: The manipulationState variable has been used many times in the 3D gestures part of wwdc2023-10113 Video. What do I need to use Expected Type for this variable? I tried to use the following code: @State private var manipulationState: GestureState Xcode error Reference to generic type 'GestureState' requires arguments in <... > Insert '<Any>' But I don't know what parameter to refer to in <Any> Question 2: If you have all the code snippets (or similar) of the 3D gestures part in this tutorial, please send it to me. Thank you.
0
0
388
Dec ’23
Spatial Audio
How does visionOS play an MP4 audio to Spatial Audio through SwiftUI or RealityKit? Note: Since I can only test the App through Simulator, in order to ensure that my Spatial Audio is played correctly in the space, please tell me how to display the location of Spatial Audio in the space. Ew and how to delete this View after the test, thank you!
0
0
651
Dec ’23
Play Spatial Audio
I want to play RealityKitContent USDA model's Spatial Audio, I use this code: RealityView{ content in do { let entity = try await Entity(named: "isWateringBasin", in: RealityKitContent.realityKitContentBundle) let audio = entity.spatialAudio entity.playAudio(audio) content.add(entity) } catch { print("Entity encountered an error while loading the model.") return } } entity.playAudio(audio) this code need add a 'AudioResource' back of audio, Excuse me, what should AudioResource be?
0
0
570
Dec ’23
Join the Apple developer program with the same identity
I now have a developer account to join the Apple developer program. Now I want to add a new developer account to the Apple developer program (with the same identity) without my original account being affected. What should I do?
Replies
0
Boosts
0
Views
439
Activity
Sep ’23
Join the Apple developer program with the same identity
I now have a developer account to join the Apple developer program. Now I want to add a new developer account to the Apple developer program (with the same identity) without my original account being affected. What should I do?
Replies
0
Boosts
0
Views
434
Activity
Sep ’23
Reputation of Apple Developer Program account
May I ask how to find the credibility of the Apple Developer Program account?
Replies
1
Boosts
0
Views
560
Activity
Oct ’23
TipKit with visionOS
Is TipKit compatible with visionOS?
Replies
1
Boosts
0
Views
720
Activity
Nov ’23
visionOS with GroupActivities
In visionOS, in order to fully implement Group Activities, do we need to write additional code?
Replies
1
Boosts
0
Views
718
Activity
Nov ’23
Some question about visionOS
I have some question about visionOS: Does Apple open the eye tracking API to developers? If I want to know how to achieve it, when the eyes are stared at a specific View, a Boolean value can be changed to true, and when the eyes are removed from this View, it will become false. In ImmersiveSpace, when immersionStyle is .full or .progressive, a black background will appear by default. How can I turn this background into a panorama of my own? In ImmersiveSpace, how to make a View always follow the user?
Replies
2
Boosts
0
Views
703
Activity
Nov ’23
visionOS's "offset3D"
In visionOS, have many code with 3D attributes (SwiftUI) is adapted from the code in iOS, like: .padding(_:) to .padding3D(_:). In iOS have .offset(x:_, y:_), it only have X and Y, but in visionOS, view is in a 3D scene, so I want offset have Z, but offset can't use Z, so I try offset3D: import SwiftUI //Some View .offset3D(x: Number, y: Number, z: Number) Xcode report an error: Error: No member 'offset3D' So do you now how to use like offset's Modifiers, and can use Z in visionOS.
Replies
1
Boosts
0
Views
680
Activity
Nov ’23
Fixed background in visonOS
I successfully changed a picture to the background in ImmersiveSpace in a full state with the following code. import RealityKit struct MainBackground: View { var body: some View { RealityView { content in guard let resource = try? await TextureResource(named: "Image_Name") else { fatalError("Error.") } var material = UnlitMaterial() material.color = .init(texture: .init(resource)) let entity = Entity() entity.components.set(ModelComponent( mesh: .generateSphere(radius: 1000), materials: [material] )) entity.scale *= .init(x: -1, y: 1, z: 1) content.add(entity) } } } However, when running, I found that when the user moves, the background is not fixed, but follows the user's movement, which I feel unrealistic. How to fix the background in the place where it first appears, and give the user a kind of movement that is really like walking in the real world, instead of letting the background follow the user.
Replies
1
Boosts
0
Views
734
Activity
Dec ’23
Model3D and Reality Composer Pro
I created a RealityKitContent in the Packages folder of the visionOS app project. At first, I tried to add a USDA model directly to its rkassets. I used Model3D(named: "ModelName", bundle: realityKitContentBundle) can The model is displayed normally, but then when I add a folder in rkassets and then put the USDA model in that folder, use Model3D(named: "ModelName", bundle: realityKit ContentBundle) cannot display the model normally. What should I do? If you know how to solve the above problems, please let me know and check if you know how to solve the following problems. If you know, please also tell me the answer. Thank you! The USDA model I mentioned in the question just now contains an animation, but when I used Model3D(named: "ModelName", bundle: realityKitContentBundle) , I found that the animation was not played by default, but needed additional code. Is there any documentation, video or listing code in this regard?
Replies
0
Boosts
0
Views
706
Activity
Dec ’23
Beyond the visionOS virtual machine
Some functions cannot be tested in the normal state of the visionOS virtual machine, such as hand tracking, and I need to test these functions. Excuse me, in addition to applying for Dev Kit and participating in the laboratory, how can I test these special functions?
Replies
1
Boosts
0
Views
783
Activity
Dec ’23
wwdc2023-10113 Video's manipulationState variable
Question 1: The manipulationState variable has been used many times in the 3D gestures part of wwdc2023-10113 Video. What do I need to use Expected Type for this variable? I tried to use the following code: @State private var manipulationState: GestureState Xcode error Reference to generic type 'GestureState' requires arguments in <... > Insert '<Any>' But I don't know what parameter to refer to in <Any> Question 2: If you have all the code snippets (or similar) of the 3D gestures part in this tutorial, please send it to me. Thank you.
Replies
0
Boosts
0
Views
388
Activity
Dec ’23
Spatial Audio
How does visionOS play an MP4 audio to Spatial Audio through SwiftUI or RealityKit? Note: Since I can only test the App through Simulator, in order to ensure that my Spatial Audio is played correctly in the space, please tell me how to display the location of Spatial Audio in the space. Ew and how to delete this View after the test, thank you!
Replies
0
Boosts
0
Views
651
Activity
Dec ’23
Add light source
How to add a light source to the View of the visionOS APP.
Replies
1
Boosts
0
Views
816
Activity
Jan ’24
Immersive Audio
visionOS App how to play Spatial Audio, Ambient Audio and Channel Audio ?
Replies
0
Boosts
0
Views
613
Activity
Dec ’23
Play Spatial Audio
I want to play RealityKitContent USDA model's Spatial Audio, I use this code: RealityView{ content in do { let entity = try await Entity(named: "isWateringBasin", in: RealityKitContent.realityKitContentBundle) let audio = entity.spatialAudio entity.playAudio(audio) content.add(entity) } catch { print("Entity encountered an error while loading the model.") return } } entity.playAudio(audio) this code need add a 'AudioResource' back of audio, Excuse me, what should AudioResource be?
Replies
0
Boosts
0
Views
570
Activity
Dec ’23