When I drag and drop a file with flag "shouldAttemptToOpenInPlace: true", I was able to access the original file name in macOS 15.
After upgrading to macOS 26, I can't access the original file name anymore. Instead, I got some useless file name such as ".com.apple.Foundation.NSItemProvider.gKZ91u.tmp".
The app no longer works with these tmp filenames because it needs the orignal file name to do the file transfer. (Btw, this is a WinSCP like app on Mac platform)
Could you please check and fix this issue? Thank you.
FileRepresentation(contentType: .item, shouldAttemptToOpenInPlace: true)
Selecting any option will automatically load the page
Post
Replies
Boosts
Views
Activity
Regarding the offer codes in https://developer.apple.com/help/app-store-connect/manage-subscriptions/set-up-offer-codes, I set up some incorrect offer codes by mistake.
Currently, I can only deactivate it but I can't delete it. Is it possible to let me delete the incorrect offer codes due to the below reasons? Thank you.
I can't recreate the offer code due to error message This reference name is or has been used by another offer in this subscription group.
When I use the offer code redemption view in https://developer.apple.com/documentation/SwiftUI/View/offerCodeRedemption(isPresented:onCompletion:), it shows all the deactivated offer codes. I don't want them to be displayed in the page.
I have tested the MagnifyGesture code below on multiple devices:
Vision Pro - working
iPhone - working
iPad - working
macOS - not working
In Reality Composer Pro, I have also added the below components to the test model entity:
Input Target
Collision
For macOS, I tried the touchpad pinch gesture and mouse scroll wheel, but neither approach works. How to resolve this issue? Thank you.
import SwiftUI
import RealityKit
import RealityKitContent
struct ContentView: View {
var body: some View {
RealityView { content in
// Add the initial RealityKit content
if let immersiveContentEntity = try? await Entity(named: "Immersive", in: realityKitContentBundle) {
content.add(immersiveContentEntity)
}
}
.gesture(MagnifyGesture()
.targetedToAnyEntity()
.onChanged(onMagnifyChanged)
.onEnded(onMagnifyEnded))
}
func onMagnifyChanged(_ value: EntityTargetValue<MagnifyGesture.Value>) {
print("onMagnifyChanged")
}
func onMagnifyEnded(_ value: EntityTargetValue<MagnifyGesture.Value>) {
print("onMagnifyEnded")
}
}
Topic:
Spatial Computing
SubTopic:
General
How to create fire effect in Reality Composer Pro? Should I use particle emitter?
I have no clue at all. Any suggestions are welcome. Thank you.
I tried to use the application icon from sample project https://developer.apple.com/documentation/visionos/diorama, but the 3 layers of the app icon are not separated when I hover on the icon in the Vision Pro simulator. Could you please advise how to fix the problem? I am using the latest Xcode Version 15.4 (15F31d). Thank you.
I have followed the tutorial below to build a Vision Pro app.
https://developer.apple.com/tutorials/develop-in-swift/create-3d-models-in-the-shared-space
When I run the app in Vision Pro Simulator, the models are much darker than the sample in the last picture of the above link.
How to fix the problem? Do I need to create lights to illuminate the model? It was not mentioned in the tutorial.
device.supportsRaytracing can be used to check if the device supports Raytracing. It seems most of the devices support Raytracing. However, only M3 and iPhone 15 Pro supports hardware Raytracing. How do I check programmatically if a device supports hardware Raytracing? Thank you.