Post

Replies

Boosts

Views

Activity

Will Apple Reject Apps That Read MacBook Lid-Angle Sensor via Private APIs?
Hey I’m working on a macOS app that wants to detect the MacBook lid / hinge angle (i.e. how far the screen is open) by directly reading the internal sensor via HID / IOKit (a private / undocumented API). I came across this project: LidAngleSensor — GitHub: https://github.com/samhenrigold/LidAngleSensor?tab=readme-ov-file Before investing too much effort, I’d like to ask the community: Has anyone succeeded in getting such an app accepted on the Mac App Store when it includes sensor-level, private API access like this? What were the reviewer feedback or rejection reasons (if any)? Are there documented cases (positive or negative) where Apple approved or rejected apps for accessing non-public hardware sensors? What’s the risk of getting banned or permanently rejected for integrating this kind of functionality? If you have direct experience (whether it passed or failed), I’d love to hear your stories, strategies, or pointers. Thanks in advance!
1
0
158
Sep ’25
UIKit in visionOS
The default framework of visionOS is SwiftUI, but it can also be combined with UIKit in SwiftUI, and Apple has also released video documents and other content for this purpose. But what I want to know is that SwiftUI is much simpler than UIKit, but why does Apple recommend developers to combine with UIKit in SwiftUI? I also want to know what are the advantages of SwiftUI and UIKit in visionOS? I hope you can reply to my two questions. Thank you!
1
0
865
Sep ’23
Download Xcode15.1 Beta3 from the new
Just now, I accidentally deleted Xcode15.1 Beta3. Now I want to download it again, but I found that Xcode 15.1 Release Candidate replaced Xcode15.1 Beta3 in https://developer.apple.com/download/applications/, but there is no visionOS SDK in Xcode 15.1 Release Candidate, so I really need to download Xcode15.1 Beta3 now. Who has any way for me to download it again?
1
0
1k
Dec ’23
Play the USDZ animation in RealityView
I have a USDZ model called 'GooseNModel' in the visionOS App project. I'm sure that this model contains an animation, so I wrote the following code to display the model with animation: import SwiftUI import RealityKit RealityView{ content in if let GooseNModel = try? await Entity(named: "GooseNModel"), let animation = GooseNModel.availableAnimations.first { GooseNModel.playAnimation(animation) content.add(GooseNModel) } } But when I ran it, I found that it did not contain animation as I imagined, but a static model. How to solve it?
3
0
2.4k
Dec ’23
Model3D with deep stretching
I used Model3D to display a model: Model3D(named: "Model", bundle: realityKitContentBundle) { phase in switch phase { case .empty: ProgressView() case .failure(let error): Text("Error \(error.localizedDescription)") case .success(let model): model.resizable() } } However, when I ran, I found that the width and length were not stretched, but when I looked at the depth from the side, they were seriously stretched. What should I do? Note: For some reason, I can't use the Frame modifier. Image: width and length error depth
0
0
496
Dec ’23
Strange visionOS Simulator
I found that my visionOS Simulator is very strange. Many functions and features are missing. For example, I learned from the Internet that the immersive scenes of Environments in their visionOS Simulator can be opened, but I click There was no response after the attack. There are not only these, but also many system features. I saw on the Internet that other developers have them, and I am missing. I'm worried that this will have an impact on me when testing my app. May I ask why? Some information: My updated Xcode version is the latest Xcode15.1Beta. Device: iMac (2021) Simulator system number: 21N305
1
0
777
Aug ’24
Follow the body
In visionOS project, How does the model in RealityView track a certain part of the body, such as left foot, right leg, left arm, etc., and store the position information of the model moving in the following part in a variable (in order to realize the transmission of body part information in the call). Please let me know if Apple publishes a document about the Persona virtual camera.
1
0
853
Jan ’24
Developer's Vision Pro
Apple launched the Developer Strap on the developer's official website, which allows developers to connect Vision Pro and Mac with USB-C, but I can't get it in China, so I want to ask developers who already have Vision Pro or People who understand these questions: Can Vision Pro without Developer Strap be debugged wirelessly? If so, How is the experience of unlimited debugging? and how does it compare with wired connection? Thank you!
2
0
1.7k
Feb ’24