For the subscription API, we’re using the Get All Subscription Statuses API to replace the deprecated verifyReceipt method. To determine if a user has canceled their subscription, we’re using the expirationIntent key from JWSRenewalInfo data. However, we’ve noticed that we sometimes receive this key and other times not.
We’re also facing an issue with the offertype key. We use this key to determine if a user is currently in the introductory offer, the promotion offer, or neither. To obtain this key, we’re using JWSTransaction, but we occasionally receive it and other times not.
Note: These issues are being tested in the sandbox environment.
Thank you.
This is a dedicated space for developers to connect, share ideas, collaborate, and ask questions. Introduce yourself, network with other developers, and join us in fostering a supportive community.
Selecting any option will automatically load the page
Post
Replies
Boosts
Views
Activity
Issue with Heart Rate Graph from Apple Watch in iOS App
Hello Community!
I hope everyone is doing well. I'm developing an iOS app that includes a feature to display a graph of the heart rate recorded during a route using the Apple Watch.
Current Implementation
I have successfully implemented data collection and transfer from the Apple Watch to the iPhone.
The app's main UI correctly displays the user's real-time BPM.
When the user stops the route, an analysis view is generated, showing speed and altitude (both work perfectly).
Expected Behavior
The user starts recording a route while monitoring heart rate.
After tapping "Stop Route," an analysis view should display a heart rate graph along with speed and altitude data.
Issue Encountered
Despite multiple attempts, I have not been able to properly generate the heart rate graph.
The only result I have achieved so far is a flat line with points representing the recorded BPM over time, instead of a proper graph.
Request for Help
If anyone has experience implementing a similar feature, I would appreciate any guidance on correcting my implementation.
I can share:
The file responsible for graph visualization.
The file handling data processing after the user stops the route.
Any help or suggestions would be greatly appreciated!
Summary
I'm trying to display a heart rate graph from Apple Watch data in my iOS app. While data collection and transfer work fine, the graph only shows a flat line instead of a proper visualization. Speed and altitude graphs work correctly. If anyone has experience with this, I’d love some guidance!
Topic:
Community
SubTopic:
Apple Developers
I have an old application. It worked at one time. Due to changes in the user environment, it is now possible to download applications from the AppStore. I built the app and sent it for review. And I received a report that my app crashed at startup. I tried to reproduce the fall for myself, but it didn't work out. In the crash log, I couldn't figure out what was wrong with the app. Please help me and tell me what I should look for in the log or in the application.
The crash location (which I found in the log):
_UIPathLazyImageAsset imageWithConfiguration
cl-1.crash
cl-1-n.crash
Topic:
Community
SubTopic:
Apple Developers
Hi everyone,
Our Apple Developer account (ID/Email: [*****]) was suddenly disabled, showing "Terminated" status in the backend, but we received no violation notice via email. Current issues:
Context:
Last operation before termination: "App submission"
No warning emails or policy violation details received (checked spam folder)
Attempted Solutions:
Submitted appeal via Developer Support Page, no response yet
Reviewed all apps against 《App Store Review Guidelines》, no obvious violations found
Questions:
Are there alternative ways to contact Apple support directly?
What are common causes for such terminations (e.g., payment issues, account association risks)?
Should we create a new account, or is reinstatement possible?
Note: Sensitive details redacted. Will provide more info if needed. Thanks for any advice!
Topic:
Community
SubTopic:
Apple Developers
Another developer tried to start a screen-sharing session with me on my M1 MacBook Air this morning (Sequoia 15.3). He reported however that another session was already under way on my machine. I looked in System Settings and found that the "Screen Sharing" setting was grayed out, with the notation that "This service is currently being controlled by the Remote Management service". Access to the service is disabled however, and I have no way of turning it off.
It's possible that the person who tried to help me got this session started, and that it then got aborted in some irregular way, but he see no indication of that and says there is no connection between us as seen from his end. This makes me nervous security-wise. Is there some way I can turn this off and ensure that I am not being screen-shared against my will, as it were?
TIA,
Howard
Topic:
Community
SubTopic:
Apple Developers
I have the following questions that need official answers:
1: Whether the Bluetooth scanning cycle can be controlled through the software level. If it can be controlled, please tell me how to control it.
2: Now the IOS mobile phone scanning strategy, how long is the scanning cycle, and what is the scanning duty cycle in this scanning cycle
Topic:
Community
SubTopic:
Apple Developers
I hope questions about third party APIs are allowed. If not I guess a delete is in order. That said....
I am looking to do an app for dog owners, trainers and enterprises such as shelters & boarding, and a couple of features require access to one or more cameras in real time. So I have been looking for hardware that provides an API and coming up empty. So far as I can tell none of the major pet cam companies, like Furbo, Petcube or Eufy. Even security cameras like the TP-Link Tapo doesn't seem to.
Am I just failing epically in my search formulation, or are all these companies playing the printer game and selling the camera cheap and making all their money in the monthly subscription to actually be able to use the hardware, and keeping that to themselves?
And, can anyone recommend decent to good third party hardware that does provide a Swift API? Or am I better off just targeting the use of iPhone & iPad and Mac cameras? It's not ideal, but it's better than nothing.
With CarPlay communication plugin R18.1, I followed these steps to integrate Enhance siri, the music sound was output from carplay and there is no option for output to Car.
============================================
Enhanced Siri
Declare supported audioFormats for the AuxIn and AuxOut streams
Since the AuxIn and AuxOut streams for Siri do not have to be both active at the same time, the accessory must claim audio formats support
for AuxIn Audio and AuxOut audio independently. The audio formats for each stream can differ from each other (48KHz for AuxOut and 16KHz
for AuxIn). The new audio types represent these new streams - AuxIn/speechRecognition & AuxOut/speechRecognition.
Check if connected device supports the feature
AirPlayReceiverSessionHasFeatureEnhancedSiri()
Claim support in the Setup Response message if device supports the feature
Add kAirPlayKeyAccessoryEnabledFeature_EnhancedSiri key through the AirPlayReceiverServer delegate AirPlayReceiverServerCopyProperty_f
function for the kAirPlayKey_AccessoryEnabledFeatures key.
Helper function: CFArrayAppendValue()
Add Enhanced Siri parameters dictionary in the INFO message
Add dictionary through the AirPlayReceiverServer delegate AirPlayReceiverServerCopyProperty_f
function for the kAirPlayKey_EnhancedSiriInfo key.
kAirPlayKey_EnhancedSiriInfo dictionary parameters:
Voice activation of Siri - kAirPlayKey_EnhancedSiriVoice
Current language of voice model - kAirPlayKey_VoiceModelCurrentLanguage
Supported languages of voice model - kAirPlayKey_VoiceModelSupportedLanguages
Enhanced Button activation of Siri - kAirPlayKey_EnhancedSiriButton
Supported zone(s) if any - kAirPlayKey_SupportedSiriTriggerZones
Update AudioStream
Get state of the AuxIn state by providing an implementation of AudioStreamUpdateState() - Off, local buffering, or streaming to device.
Decouple input streams from output streams. AuxIn is an independent input stream only. The property kAudioStreamProperty_Direction will
provide the necessary information if the stream is input, output or input & output.
Provide a handler for the AirPlayReceiverSessionDelegate setEnhancedSiriParams_f
This will provide additional information:
Activation type
Setting the language of the voice model
Invoke the Communication Plugin to start buffering
Once the activation type has been specified, the accessory can request the plugin to start buffering using
AirPlayReceiverSessionAuxInStart().
Use the new APIs to trigger Siri:
AirPlayReceiverSessionRequestSiriActionWithLatency()
AirPlayReceiverSessionRequestSiriVoiceActivationWithLatency()
AirPlayReceiverSessionRequestSiriVoiceActivationWithSample()
Button presses and voice activations should use this new APIs which adds a timestamp of the activation. These APIs allow
a choice of a latency or a sample for button and voice activations.
If there is a delay between the user pressing the button to notifying the device on the button press, this latency value
should represent this time.
If the accessory can determine which zone activated, it can provide the zone with the request.
Invoke the Communication Plugin to stop buffering
You may need to invoke the plugin to stop buffering (AirPlayReceiverSessionAuxInStop()) if exclusive access to the microphone is necessary.
Such instances may include but not limited to:
Native voice recognition session
Telephony
Another stream function which uses the microphone starts
modesChanged notification can be used to determine if a resource is being used
Note, if the session ends, the plugin will automatically stop buffering
Hello,
I am currently developing an application for Apple TV using TVML, and I am trying to create a user registration or login form with multiple input fields (e.g., "Username" and "Password"). However, I am facing issues with displaying multiple textField components in the formTemplate.
Here are the approaches I have tried:
Attempt 1:
<document>
<formTemplate>
<banner>
<title>Login</title>
</banner>
<banner>
<textField>UserName</textField>
</banner>
<banner>
<textField>Password</textField>
</banner>
<footer>
<button id="button1">
<text>Button 1</text>
</button>
</footer>
</formTemplate>
</document>
Attemp 2:
<document>
<formTemplate>
<banner>
<title>Login</title>
</banner>
<textField>UserName</textField>
<textField>Password</textField>
<footer>
<button id="button1">
<text>Button 1</text>
</button>
</footer>
</formTemplate>
</document>
In both cases, the layout does not render the fields as expected. Either the textField components do not display at all, or the structure seems incorrect.
Could someone please guide me on the proper way to display multiple textField components in a formTemplate? Is there a limitation or specific requirement for structuring these elements?
Thank you in advance for your assistance!
Hi all, if anyone from apple Dev team is seeing this, please do consider supporting device mirroring without the need to login to icloud and with cable connection in addition to wi-fi.
I am working in a corporate setting which does not allow us to login to icloud, thus I am unable to use device mirroring when my work really needs is, I am stuck with Quick time preview.
I am quite sure, that this does not only apply to me.
I’m experiencing something strange with my phone’s storage, and I’m not sure what’s going on. Last year, I had nearly maxed out my storage at 255/256 GB, but suddenly, it dropped to around 100 GB, and now it’s fluctuating again—it decreased to 99 GB today. This is concerning, and I’m wondering if something might be wrong with my phone. Could someone help me understand what’s happening?
Here’s some context: my phone was being used for filming a school project, and I often got “insufficient storage” notifications. To make room, I deleted some apps, but I was still hovering around 250-253/256 GB of used storage.
Later, I heard that getting iCloud storage would free up space by storing photos in the cloud, so I subscribed to 200 GB of iCloud storage. I used it for about a day but then canceled the subscription after hearing it wasn’t effective at freeing up much space. Unfortunately, I didn’t check my storage before canceling the subscription, which I now realize was a mistake.
In December, I took a lot of photos and videos during trips, but I didn’t check my storage, assuming I’d hit full capacity soon. However, just two days ago, I checked my storage and was shocked to see that it had suddenly dropped to around 105 GB. This morning, it dropped again to about 100.96 GB. I read in an article that this could be a bug and that updating the software might help, so I updated my phone to iOS 18.
After the update, I checked my storage again, and it had dropped further to 94.16 GB. Then, in the same minute, it fluctuated to 94.99 GB, and four minutes later, it went down to 94.65 GB. Now, as I’m writing this, it’s back up to 95.58 GB.
Could anyone explain what’s going on?
Hello,
I'm trying to publish my app, but I'm constantly getting rejected by Apple. They're telling me I'm having issues with tracking user data.
This item has been rejected for the following reasons:
5.1.2 Legal: Privacy - Data Use and Sharing
I've indicated that I don't use this data for ads, that it's only used for personalization and to understand who saves items.
I added the NSUserTrackingUsageDescription property to the info.plist.
I run AppTrackingTransparency.requestTrackingAuthorization() when the user logs into the app, displaying a warning message.
I'd say I meet all the requirements they've set for me, but they still haven't approved my app. What do you recommend? How can I speak to a physical person who can help me?
Thank you very much and best regards.
when i send a http request with NSURLSession, app is crashed.
_quic_frame_write
_quic_frame_write
__quic_packet_builder_assemble
_quic_conn_assemble_and_encrypt
_quic_conn_send_frames_for_key_state
_quic_conn_send_internal
_quic_recovery_pto
_quic_recovery_timer_fired
_quic_timer_run
___nw_protocol_instance_schedule_wakeup_block_invoke
___nw_context_reset_timer_block_with_time_block_invoke
__dispatch_client_callout
__dispatch_continuation_pop
__dispatch_source_latch_and_call
__dispatch_source_invoke
__dispatch_workloop_invoke
__dispatch_root_queue_drain_deferred_wlh
__dispatch_workloop_worker_thread
__pthread_wqthread
Topic:
Community
SubTopic:
Apple Developers
Scenario: hardware keyboard is attached to iPhone
Before iOS 18.2, if I setup only two keyboards (e.g., english and emoji) in the system, click Globe key can switch to iOS emoji keyboard without problem. Since iOS 18.2, it stop working, and emoji keyboard is never displayed. I need to add one more keyboard, then I can switch to emoji keyboard (via the keyboard selector).
Topic:
Community
SubTopic:
Apple Developers
Hi All,
Im writing a flutter app and have had no issue thus far. My latest review says the iPad air (5th) is hanging on the splash (launch) screen. I have no idea why this is. The main method In flutter is more or less empty, as functions have been offset elsewhere.
I'm not quite sure what to do next. I resorted to rebuilding the IOS dir and that make a difference for the iPhone issue not the iPad.
I run it in the simulator and notice one issue which I imagine is the problem. it states:
Errors found! Invalidating cache...
fopen failed for data file: errno = 2 (No such file or directory)
Errors found! Invalidating cache...
I'm using plane old flutter and don't believe I access any assets that aren't specified. Any pointer would be a great help!
Topic:
Community
SubTopic:
Apple Developers
Our app provides configuration and maintenance features for Siemens electrical devices. The app supports multiple devices and the devices have their own specific features, which also need to be supported by our App.
The devices can be released to the market only when our app supports those features.
Our app release timeline is once in every 4 months. We face a challenge when some of these devices are not able to deliver their features by our release time.
So, we would like to make these devices plugins, which can be independently released, without releasing the App every time. The app would, of course, be released once in every 4 months, depending on the common features.
I'm not able to find how to release only the plugins within the app, without releasing the entire app.
Some one has suggested app extensions, but that does not fit our solution as in some cases, we need continuous polling of data from one or mode devices for minutes or hours, as the app can be used by the user to monitor some of the device values.
Topic:
Community
SubTopic:
Apple Developers
Hi,
I have a Retina 13 inch 2020 macbook air and an iPad 7.
My macbook is updated to macOS sonoma and my iPad is updated to iPadOS 18.3.
I want to use the sidecar feature on my mac and project the macbook display onto my iPad. I remember it working around a year ago when I used this feature before, but now it's telling me I have issues connecting due to my iPad's WiFi being 'unavailable.' However, my iPad isn't connecting to hot-spot, my mac isn't sharing the internet connection, they're linked to the same apple ID account, and they're on the same WiFi network.
I've tried connecting both wirelessly and wired (and trusted the computer) but no matter how much I troubleshoot it, it won't connect.
I'd be really grateful if somebody could help me fix this.
Thanks :D
Topic:
Community
SubTopic:
Apple Developers
I am trying to install macOS Big Sur on my VM. It gets stuck on “About 14 minutes remaining ”Does anyone know how to get past it?
Hi there, with the latest update, the ability to click in the file name field at the top of a document (eg Pages, Numbers) to add or change the name of the file, then select filing location, no longer works. Previously, a user could do this and then hit 'return' and it would save to that location with that new file name. It no longer does this. This is a very useful functionality to have. Can it be returned in a future update?
Topic:
Community
SubTopic:
Apple Developers
Is there a way to check whether third-party code can be inserted into parts of the ARM CPU architecture from outside? I have the paranoid impression that my MacBook Pro, M2 processor is being manipulated from outside. For example: Safari, when opened, displays pages scaled down without my conscious intervention, when clicked on the original size it scales back to normal for a short time and then implodes again; I use the navigation bar on the left excessively, it is my most important tool in Safari, today it was impossible to insert a link using drag & drop, the bar scrolls up or down very quickly; a behavior that I only know from Intel processors. For example, notes, I encrypted an important note, no problem even though photos and links are inserted, yesterday I could no longer open the note, neither on the MacBook Pro nor on the iPhone. When opening the notes app in the terminal on the MacBook Pro, an EXEC_CRASH report was generated. I have a terrible suspicion that the Apple branch in Munich is tampering with the devices and software of users in Germany. Sorry, I'm a big fan of Apple and its products, but this is very spooky. I just wanted to share my distrust with you!
Topic:
Community
SubTopic:
Apple Developers