Overview

Post

Replies

Boosts

Views

Activity

UIScreen.isCaptured and sceneCaptureState report inactive while system screen recording continues when Live Activity is expanded
Hi, I’m trying to reliably detect when system screen recording finishes, and I’m observing behavior that I don’t fully understand when a Live Activity is expanded via Dynamic Island. Environment Devices: iPhone 16 Pro iOS: 26.2 Xcode: 26.2 Using UIScreen.isCaptured and UIWindowScene.sceneCaptureState Implementation: I observe capture state like this: private var observation: NSKeyValueObservation? func startObserving() { observation = UIScreen.main.observe(\.isCaptured, options: [.new]) { _, change in print("isCaptured:", change.newValue ?? false) } } I also check: window.traitCollection.sceneCaptureState Steps to Reproduce Start system screen recording from Control Center. Confirm UIScreen.main.isCaptured == true. Expand a Live Activity via the Dynamic Island (e.g. timer or call). Observe capture state values while the Live Activity UI is expanded. Observed Behavior While screen recording is still active: UIScreen.main.isCaptured becomes false sceneCaptureState becomes .inactive This state persists while the recording Live Activity is expanded The system recording indicator remains visible The device continues recording Expected Behavior (or Clarification Needed) My understanding was that UIScreen.isCaptured indicates whether the device screen is currently being captured (e.g. screen recording or mirroring). However, this behavior suggests that both isCaptured and sceneCaptureState reflect whether the current scene is part of the capture surface, rather than whether device-level recording is active. Is this the intended behavior when system-owned surfaces (such as expanded Live Activities) are promoted above the app’s scene? If so: Is there any supported way to reliably detect device-level screen recording state (as opposed to scene-level capture participation), in order to trigger logic when recording finishes? Thank you for any clarification.
Topic: UI Frameworks SubTopic: UIKit
0
0
68
4w
Delay in app review : more than 13 business days
Hello, I'm desperately seeking your help here. I submitted an app for review, on Feb 16th, it's still sitting at Waiting for review status. I raised a follow-up ticket on Feb 20th, inquiring about the status of the review. Auto response was a feedback will be given by email in 2 business days. It's been now more than 10 business days, no feedback on the ticket. I raised a new one today. I really cannot understand why this is taking so long, while I'm told by other friends around me that this usually doesn't take more than a week max. Please help!! I'm losing all credibility with my potential customers and users at the moment because of this delay.
1
0
53
3w
Xcode 26.3 MCP server output missing `structredContent`
I am using official MCP SDK. According to official guide, Servers MUST provide structured results that conform to this schema. https://modelcontextprotocol.io/specification/draft/server/tools#output-schema I could see output schema defined, but result have no structured content. Current output schema: { name: "XcodeListWindows", title: "List Windows", description: "Lists the current Xcode windows and their workspace information", inputSchema: { type: "object", properties: { }, required: [ ], }, outputSchema: { type: "object", properties: { message: { description: "Description of all open Xcode windows", type: "string", }, }, required: [ "message", ], }, } Current response: { "result": { "content": [ { type: "text", text: "{\"message\":\"* tabIdentifier: windowtab1, workspacePath: \\xxx\\n* tabIdentifier: windowtab2, workspacePath: \\xxx\\n\"}", }, ] } } Expected: { "result": { "content": [ { type: "text", text: "{\"message\":\"* tabIdentifier: windowtab1, workspacePath: \\xxx\\n* tabIdentifier: windowtab2, workspacePath: \\xxx\\n\"}", }, ], "structuredContent": { "message": "* tabIdentifier: windowtab1, workspacePath: \\xxx\\n* tabIdentifier: windowtab2, workspacePath: \\xxx\\n", } } }
1
0
139
4w
Xcode 26 – Organizer does not show “App Store Connect” for watchOS standalone app (only Release Testing available)
Hello, I am trying to upload a watchOS standalone app to App Store Connect using Xcode 26.3, but Organizer does not show the “App Store Connect” upload option. It only displays: • Release Testing (Ad Hoc) • Enterprise • Debugging Project setup • Target type: watchOS App (standalone, not companion) • PRODUCT_TYPE = com.apple.product-type.application • PRODUCT_BUNDLE_PACKAGE_TYPE = APPL • WRAPPER_EXTENSION = app • SUPPORTED_PLATFORMS = watchos watchsimulator • Archive configuration: Release • Base SDK: watchOS The app does not depend on iOS. Bundle ID Originally: com.example.app.watchkitapp Recreated as: com.example.app HealthKit capability enabled in: • Developer Portal • Xcode (Signing & Capabilities) Signing • Apple Distribution certificate active • App Store provisioning profile created • Manual signing configured for Release (Apple Distribution + App Store profile) • Also tested Automatic signing • Deleted old Ad Hoc profiles • Cleared Derived Data and regenerated archive App Store Connect • App created successfully using bundle ID com.example.app • Account role: Account Holder • No pending Agreements, Tax, or Banking items Despite a valid Release archive signed with Apple Distribution, Organizer only allows Ad Hoc distribution and never shows “App Store Connect”. Is this a known issue in Xcode 26 affecting independent watchOS apps, or is there an additional configuration step required? Thank you.
1
0
64
4w
How can I change the output dimensions of a CoreML model in Xcode when the outputs come from a NonMaximumSuppression layer?
After exerting a custom model with nms=True. In Xcode, the outputs show as: confidence: MultiArray (0 × 5) coordinates: MultiArray (0 × 4) I want to set fixed shapes (e.g., 100 × 5, 100 × 4), but Xcode does not allow editing—the shape fields are locked. The model graph shows both outputs come directly from a NonMaximumSuppression layer. Is it possible to set fixed output dimensions for NMS outputs in CoreML?
2
0
225
4w
Waiting for Review for 21 Days, Need Assistance
One of my apps has been stuck in “Waiting for Review” for the past 21 days. App ID: 6758734829. I have contacted other support channels over the past few weeks but have not received any response. I also have three additional apps, App IDs 6758755469, 6759047311, and 6759095185, that have been in review for the last 14 days. All of these apps have already been live on the Google Play Store for months. I am trying to launch them on the App Store so I can move forward with my marketing plans. I kindly request Apple to review these submissions and help resolve whatever is causing the delay.
2
0
103
3w
Which tools and concepts are essential to learn first in iOS development?
I’m beginning my journey in iOS app development with a strong Java background and an app idea in mind, and I want to learn the best practices, tools, and workflows to build it the right way from the start. I’d appreciate guidance on useful AI tools for coding and learning Swift or SwiftUI, how developers approach app design, and recommended sources for UI inspiration, graphics, and overall development tips.
2
0
62
4w
Report inactivity
Hey has someone figured out how or when does a report request become inactive? Is there like a fixed time period for how long an ONGOING accessType stays active? And does the same rule apply for ONE_TIME_SNAPSHOT? Has someone managed to create multiple active reports for both accessTypes per one app? And can you have multiple inactive ones?
0
0
75
3w
Unity iOS (Metal) → WebRTC (Unity WebRTC) video stream to remote Unity client: PeerConnection connects but receiver renders black frames
Hello! My name is Mason Prather. I'm a graduate student at Kennesaw State University and a Research Engineer working in XR environments through my Graduate Research Assistant role. I’m currently building a research prototype that connects a mobile companion application to a VR headset. The mobile application is built in Unity and deployed on iOS, and it streams video frames to a remote Unity client using WebRTC. Environment Device: iPhone 15 OS: iOS 26.3 (tested on physical device, not Simulator) Engine: Unity 2022.3.57f1 Graphics API: Metal Streaming Technology: WebRTC (Unity WebRTC package) Architecture: Mobile Unity app streaming video frames to a remote Unity client Receiver Device: Meta Quest Pro headset (Unity application) Networking: LAN (UDP discovery + TCP signaling) Video Source: Unity RenderTexture Goal The goal of the system is to allow a VR user to view media stored on their phone inside a VR environment. The iOS app: renders or captures media content converts frames into a WebRTC video track streams the video to the headset Current Status Connection setup works correctly. Observed behavior: Signaling connection successful ICE candidate exchange successful PeerConnection state becomes Connected Video track created successfully However, the receiving application displays black frames. iOS App Details The video source originates from a Unity RenderTexture. Inside the phone application: RenderTexture displays correctly Frames appear correct locally But the receiving peer does not display the frames. Relevant Components Unity WebRTC package iOS Metal rendering pipeline Custom TCP signaling LAN discovery via UDP Expected Behavior Rendered frames should transmit via WebRTC and appear on the remote device. Actual Behavior The remote video track is active, but the rendered frames appear black on the receiving client. Questions Are there known issues involving Unity WebRTC + iOS Metal texture capture? Are there specific pixel format requirements when streaming textures from Unity on iOS? Could the issue relate to texture readback limitations or GPU synchronization? I am more than happy to provide screenshots and console logs upon request. If anyone has experience streaming Unity video frames via WebRTC on iOS, I would greatly appreciate any guidance.
0
0
210
3w
Virtual Camera Shows Jittering Frames and Solid Accent Color on macOS
Hello Apple Developer Support, I’m developing a virtual camera using the CMIOExtensionDevice / CMIOExtensionStreamSource APIs on macOS. While the virtual camera appears in system settings and apps like Zoom and Google Meet, the video output exhibits the following issues: Jittering frames: The first frame sometimes appears correctly, but subsequent frames flicker or jitter. Solid color fill: Eventually, the camera feed fills entirely with a solid accent color (e.g., blue), rather than the intended video content. Console logs: Repeated messages appear in Console.app: Invalid display 0x00000000 Setup details: The virtual camera is created using CMIOExtensionDevice and CMIOExtensionStream. Video frames are rendered from NSImage/CGImage using CGContext and copied into CVPixelBuffers. Frame delivery is controlled by a DispatchSourceTimer at 60 FPS. macOS version: [Your macOS version here] Xcode version: [Your Xcode version here] Observations: The Invalid display 0x00000000 logs suggest that CGContext drawing or NSImage operations are failing in headless mode (i.e., there is no real display attached to the virtual camera). Using CIContext with .useSoftwareRenderer = true appears to mitigate some flicker, but not entirely. Questions / Requests: Is it expected that CoreMediaIO virtual cameras cannot reliably render CGImage / NSImage frames offscreen? Are there recommended APIs or approaches to render virtual camera frames fully headless to avoid display-dependent jitter? Is there any documentation or sample code from Apple showing stable video output from a virtual camera extension that does not rely on a physical display? Any guidance or examples would be greatly appreciated. This issue prevents the virtual camera from being used reliably in standard video apps. Thank you, Savvy
0
0
58
4w
Standard Memory Debugging Tools
In many cases I’ll be talking to folks with a memory management problem and I’ll say “You should investigate this with the standard memory debugging tools.” They then turn around and ask me “What are those tools?” Well, this is what I mean: Zombies — This lets you quickly detect when an object is used after it’s been deallocated. Learn more about it in the Finding zombies section of the Instruments Help and in Investigating crashes for zombie objects. There was also an excellent WWDC video about this, namely, WWDC 2010 Session 311 Advanced Memory Analysis with Instruments. This is no longer available in the video archive but if you can find a copy it’s well worth a watch. Address Sanitizer — This is a lower-level tool that finds a variety of common memory management issues, including use after free and buffer overruns. Learn more about this in Diagnosing memory, thread, and crash issues early and the various articles it links to. There’s also a good discussion of this tool, and other Xcode runtime diagnostic tools, in WWDC 2017 Session 406 Finding Bugs Using Xcode Runtime Tools (also no longer available from Apple). Hardware memory tagging — If you have access to a device with hardware memory tagging — starting with the A19 or M5 processors — consider enabling Memory Integrity Enforcement (MIE). For the details, see Enabling enhanced security for your app or watch Secure your app with Memory Integrity Enforcement. Even if you don’t want to do that for your production app, enable it just for development by checking Hardware Memory Tagging in the Options tab of Xcode’s scheme editor. Older tools — There are a variety of older tools that might be useful in some specific circumstances. See the Enabling the Malloc Debugging Features section of the Memory Usage Performance Guidelines for more information about these. Of specific interest is libgmalloc, which is documented in a UNIX man page. For some practical examples of how to identity a memory management crash report and then investigate that crash with these tools, take a look at WWDC 2018 Session 414 Understanding Crashes and Crash Logs. Share and Enjoy — Quinn “The Eskimo!” @ Developer Technical Support @ Apple let myEmail = "eskimo" + "1" + "@" + "apple.com" Revision History: 2026-03-02 Added a discussion of hardware memory tagging. 2023-05-09 Added a link to Investigating crashes for zombie objects. 2023-03-22 Removed another WWDC session video link. Made minor editorial changes. 2020-10-23 Fixed some formatting errors. 2019-10-30 Removed the link to WWDC 2010 Session 311 Advanced Memory Analysis with Instruments because it’s not long available in the archive. Refreshed all the other links. 2019-01-22 Fixed the link to libgmalloc. 2018-11-02 Updated to include a reference to WWDC 2018 Session 414 Understanding Crashes and Crash Logs. 2017-11-16 First posted.
0
0
18k
4w
macOS Tahoe 26.3 - System Is Playing NSBeep At Inappropriate Times When Text Editing Ends Via -cancelOperation: (field editor)
When I end editing pressing the escape key, the system sometimes plays NSBeep(). I noticed this with NSBrowser. Every time I press escape to end editing the system beeps. At first I thought it was somewhere in my app but I set a symbolic breakpoint and discovered it was not coming from my code. I filed FB22127038. Since then I discovered that NSBeep playing at inappropriate times is not exclusive to NSBrowser. It appears if there is a NSTableView in the window and you just press the escape key (even if you aren't editing anything) AppKit beeps. It can be traced to: #0 0x0000000199e3184c in NSBeep () #1 0x000000019aa03fac in -[NSWindow doCommandBySelector:] () #2 0x000000019ac1d01c in -[NSTableView(NSTableViewViewBased) cancelOperation:] ()
3
0
97
3w
How do I resolve the "Automatic signing cannot update bundle identifier..." error?
When I create an archive file and attempt to upload the app using the "Distribute App" button, the upload fails with the error "Automatic signing cannot update bundle identifier...". (The detailed message is below.) When creating an archive file in Xcode, I unchecked "Automatically Manage Signing" and proceeded with the archive. The message says "Font Enumeration," but other apps with the same option enabled upload successfully. Therefore, I believe the "Font Enumeration" option is not the issue. I tried creating a new provisioning file, but it still doesn't work. I deleted all DerivedData files from my Mac storage, restarted Xcode, and tried again, but it still doesn't work. This keeps happening only for certain targets (specific apps) in Xcode. Does anyone know how to fix this? Xcode is the latest version. Message: Automatic signing cannot update bundle identifier "com.xxxxxx.xxxxxx". Automatic signing cannot update your registered bundle identifier to enable Font Enumeration. Update your bundle identifier on https://developer.apple.com/account and then try again.
1
0
124
4w
LiveCallerId OHTTP Relay: Works in TestFlight, failing in Production (Bundle ID: no.opplysningen.bedrift.LiveCallerId)
We’ve been implementing LiveCallerId using OHTTP and have hit a wall with the production environment. The setup works perfectly in TestFlight, but the release version of the app is consistently being rejected by the Apple OHTTP Relay when trying to tunnel traffic to our gateway. Timeline & Status: Applied via the form in September 2025. Received confirmation in November 2025 that our /.well-known/ohttp-keys endpoint was correctly configured. Since then, we've struggled to get a dialogue with Apple to confirm the final production whitelisting. Technical Observations: Our ohttp-keys endpoint is being polled frequently (every few minutes). Based on the traffic, this is clearly the Apple Relay infrastructure fetching/refreshing the keys, not the devices themselves. This suggests the Relay "sees" our configuration, yet it still refuses to tunnel traffic to our gateway in the production environment. Since everything is functional in TestFlight, our implementation seems correct. It feels like there is a configuration mismatch or a missing "production flip" on the Relay side for our Bundle ID. If anyone from the Apple engineering team could verify the status for this Bundle ID, it would be a huge help. We've been stuck in this "TestFlight-only" state for quite a while now.
0
0
167
3w
Trapped in a "Black Box": Almost 3 weeks of App Review silence and ignored support tickets. Is this normal?
Hi everyone, I am writing this out of pure desperation because I honestly don't know what else to do or where to turn. My app has been stuck in the review process since February 9th (almost 3 weeks now), and I feel like I am completely trapped in a black box. I have tried every official channel available to get at least some information: I submitted requests for an expedited review. I opened 3 separate tickets with Developer Support, but received absolutely zero response. Just complete, deafening silence. I finally called the developer hotline. The representative was polite but completely unhelpful. I was basically told that "everything is fine, this timeline is normal, we can't do anything to help you, just keep waiting." Just to put this wait time into perspective, while sitting in this Apple "black box" waiting for a single message, I have managed to do the following: Registered a Google Play Developer account and fully verified my identity. Almost completely rewrote my entire app from scratch, translating it from Swift to Kotlin. Uploaded the Android version, got it approved, and I am already on Day 8 of closed testing in Google Play! Meanwhile, in App Store Connect, absolutely nothing has changed. No status updates, no rejections, no questions from the reviewers, no replies to my tickets. Just an absolute void. Has anyone else experienced this kind of extreme delay and complete lack of communication recently? For context, here are my details: Submission ID: db1ffb72-8f0f-49bc-b215-179e1e179b7d Apple ID: 6756239914 Any advice or shared experiences would be deeply appreciated. Thank you.
2
0
201
4w
False Positive "Shared Device" Flag blocking Enrolment (Case 102822291300)
Dear All, I am writing to report what appears to be a bug in the Apple Developer App’s enrollment verification system that is incorrectly flagging personal hardware. Problem Statement: I am unable to complete enrollment for the Apple Developer Program. On the "Personal Information" or "Identity Verification" stage, the app displays: "Contact Us to Continue. There may be an issue with your account that needs to be resolved..." Hardware History: Device: iPhone 14 (iOS latest) The device was previously owned by a family member who never enrolled in the Developer Program. The device was factory reset/wiped before my account was added. Despite this, Support (Case 102822291300) states the device is flagged as a "shared device" and refuses to proceed with identity verification. The Impasse: Support has indicated they cannot provide further assistance. However, as an individual developer, this is my primary device. If the system is flagging a reset device as "shared" when there is no history of developer use, it suggests a flaw in the automated verification logic. Request: Can a moderator or support representative look into Case 102822291300 to verify why this specific hardware is being restricted? Is there a manual path for identity verification (Passport/National ID) that doesn't rely on the Apple Developer App's automated device check? Allow me to do manual verification and provide the resolution to the issue. I am eager to bring my app to the platform and would appreciate a path to resolution. Kind Regards, Pradeep Bangalore
0
0
38
3w
Final Cut Pro not loading my plugin
I have exhausted standard debugging approaches and need guidance on Final Cut Pro's AU plugin loading behavior. PLUGIN OVERVIEW My plugin is an AUv2 audio plugin built with JUCE 8. The plugin loads and functions correctly in: Pro Tools (AAX) Media Composer (AAX) Reaper (AU + VST3) Logic Pro (AU) GarageBand (AU) Audacity (AU) DaVinci Resolve / Fairlight (AU) Harrison Mixbus (AU) Ableton Live (AU) Cubase (VST3) Nuendo (VST3) It does NOT load in Final Cut Pro (tested on 10.8.x, macOS 14.6 and 15.2). DIAGNOSTICS COMPLETED auval passes cleanly: auval -v aufx Hwhy Manu → AU VALIDATION SUCCEEDED Plugin is notarized and stapled: xcrun stapler validate → The validate action worked spctl --assess --type exec → Note: returns 'not an app' which we understand is expected for .component bundles. Hardened Runtime is enabled on the bundle. We identified that our Info.plist contained a 'resourceUsage' dictionary in the AudioComponents entry. We found via system logs that this was setting au_componentFlags = 2 (kAudioComponentFlag_SandboxSafe), causing FCP to attempt loading the plugin in-process inside its sandbox, where our UDP networking is denied. We removed the resourceUsage dict, confirmed au_componentFlags = 0 in the CAReportingClient log, and FCP now loads the plugin out-of-process via AUHostingServiceXPC_arrow. Despite au_componentFlags = 0 and out-of-process loading confirmed, the plugin still does not appear in FCP's effects browser. We also identified and fixed a channel layout issue — our isBusesLayoutSupported was not returning true for 5-channel layouts (which FCP uses internally). This is now fixed. AU cache has been fully cleared: ~/Library/Caches/com.apple.audio.AudioComponentRegistrar /Library/Caches/com.apple.audio.AudioComponentRegistrar coreaudiod and AudioComponentRegistrar killed to force rescan auval -a run to force re-registration
0
0
215
4w
UIScreen.isCaptured and sceneCaptureState report inactive while system screen recording continues when Live Activity is expanded
Hi, I’m trying to reliably detect when system screen recording finishes, and I’m observing behavior that I don’t fully understand when a Live Activity is expanded via Dynamic Island. Environment Devices: iPhone 16 Pro iOS: 26.2 Xcode: 26.2 Using UIScreen.isCaptured and UIWindowScene.sceneCaptureState Implementation: I observe capture state like this: private var observation: NSKeyValueObservation? func startObserving() { observation = UIScreen.main.observe(\.isCaptured, options: [.new]) { _, change in print("isCaptured:", change.newValue ?? false) } } I also check: window.traitCollection.sceneCaptureState Steps to Reproduce Start system screen recording from Control Center. Confirm UIScreen.main.isCaptured == true. Expand a Live Activity via the Dynamic Island (e.g. timer or call). Observe capture state values while the Live Activity UI is expanded. Observed Behavior While screen recording is still active: UIScreen.main.isCaptured becomes false sceneCaptureState becomes .inactive This state persists while the recording Live Activity is expanded The system recording indicator remains visible The device continues recording Expected Behavior (or Clarification Needed) My understanding was that UIScreen.isCaptured indicates whether the device screen is currently being captured (e.g. screen recording or mirroring). However, this behavior suggests that both isCaptured and sceneCaptureState reflect whether the current scene is part of the capture surface, rather than whether device-level recording is active. Is this the intended behavior when system-owned surfaces (such as expanded Live Activities) are promoted above the app’s scene? If so: Is there any supported way to reliably detect device-level screen recording state (as opposed to scene-level capture participation), in order to trigger logic when recording finishes? Thank you for any clarification.
Topic: UI Frameworks SubTopic: UIKit
Replies
0
Boosts
0
Views
68
Activity
4w
Delay in app review : more than 13 business days
Hello, I'm desperately seeking your help here. I submitted an app for review, on Feb 16th, it's still sitting at Waiting for review status. I raised a follow-up ticket on Feb 20th, inquiring about the status of the review. Auto response was a feedback will be given by email in 2 business days. It's been now more than 10 business days, no feedback on the ticket. I raised a new one today. I really cannot understand why this is taking so long, while I'm told by other friends around me that this usually doesn't take more than a week max. Please help!! I'm losing all credibility with my potential customers and users at the moment because of this delay.
Replies
1
Boosts
0
Views
53
Activity
3w
Xcode 26.3 MCP server output missing `structredContent`
I am using official MCP SDK. According to official guide, Servers MUST provide structured results that conform to this schema. https://modelcontextprotocol.io/specification/draft/server/tools#output-schema I could see output schema defined, but result have no structured content. Current output schema: { name: "XcodeListWindows", title: "List Windows", description: "Lists the current Xcode windows and their workspace information", inputSchema: { type: "object", properties: { }, required: [ ], }, outputSchema: { type: "object", properties: { message: { description: "Description of all open Xcode windows", type: "string", }, }, required: [ "message", ], }, } Current response: { "result": { "content": [ { type: "text", text: "{\"message\":\"* tabIdentifier: windowtab1, workspacePath: \\xxx\\n* tabIdentifier: windowtab2, workspacePath: \\xxx\\n\"}", }, ] } } Expected: { "result": { "content": [ { type: "text", text: "{\"message\":\"* tabIdentifier: windowtab1, workspacePath: \\xxx\\n* tabIdentifier: windowtab2, workspacePath: \\xxx\\n\"}", }, ], "structuredContent": { "message": "* tabIdentifier: windowtab1, workspacePath: \\xxx\\n* tabIdentifier: windowtab2, workspacePath: \\xxx\\n", } } }
Replies
1
Boosts
0
Views
139
Activity
4w
Xcode 26 – Organizer does not show “App Store Connect” for watchOS standalone app (only Release Testing available)
Hello, I am trying to upload a watchOS standalone app to App Store Connect using Xcode 26.3, but Organizer does not show the “App Store Connect” upload option. It only displays: • Release Testing (Ad Hoc) • Enterprise • Debugging Project setup • Target type: watchOS App (standalone, not companion) • PRODUCT_TYPE = com.apple.product-type.application • PRODUCT_BUNDLE_PACKAGE_TYPE = APPL • WRAPPER_EXTENSION = app • SUPPORTED_PLATFORMS = watchos watchsimulator • Archive configuration: Release • Base SDK: watchOS The app does not depend on iOS. Bundle ID Originally: com.example.app.watchkitapp Recreated as: com.example.app HealthKit capability enabled in: • Developer Portal • Xcode (Signing & Capabilities) Signing • Apple Distribution certificate active • App Store provisioning profile created • Manual signing configured for Release (Apple Distribution + App Store profile) • Also tested Automatic signing • Deleted old Ad Hoc profiles • Cleared Derived Data and regenerated archive App Store Connect • App created successfully using bundle ID com.example.app • Account role: Account Holder • No pending Agreements, Tax, or Banking items Despite a valid Release archive signed with Apple Distribution, Organizer only allows Ad Hoc distribution and never shows “App Store Connect”. Is this a known issue in Xcode 26 affecting independent watchOS apps, or is there an additional configuration step required? Thank you.
Replies
1
Boosts
0
Views
64
Activity
4w
How can I change the output dimensions of a CoreML model in Xcode when the outputs come from a NonMaximumSuppression layer?
After exerting a custom model with nms=True. In Xcode, the outputs show as: confidence: MultiArray (0 × 5) coordinates: MultiArray (0 × 4) I want to set fixed shapes (e.g., 100 × 5, 100 × 4), but Xcode does not allow editing—the shape fields are locked. The model graph shows both outputs come directly from a NonMaximumSuppression layer. Is it possible to set fixed output dimensions for NMS outputs in CoreML?
Replies
2
Boosts
0
Views
225
Activity
4w
Waiting for Review for 21 Days, Need Assistance
One of my apps has been stuck in “Waiting for Review” for the past 21 days. App ID: 6758734829. I have contacted other support channels over the past few weeks but have not received any response. I also have three additional apps, App IDs 6758755469, 6759047311, and 6759095185, that have been in review for the last 14 days. All of these apps have already been live on the Google Play Store for months. I am trying to launch them on the App Store so I can move forward with my marketing plans. I kindly request Apple to review these submissions and help resolve whatever is causing the delay.
Replies
2
Boosts
0
Views
103
Activity
3w
Which tools and concepts are essential to learn first in iOS development?
I’m beginning my journey in iOS app development with a strong Java background and an app idea in mind, and I want to learn the best practices, tools, and workflows to build it the right way from the start. I’d appreciate guidance on useful AI tools for coding and learning Swift or SwiftUI, how developers approach app design, and recommended sources for UI inspiration, graphics, and overall development tips.
Replies
2
Boosts
0
Views
62
Activity
4w
The Digital Services Act (DSA) Spinning forever
Hey guys I been trying to complete the compliance reqs to distribute my apps in the EU, but after I insert my information it's just loading indefinitely, I have tried with safari, chrome, incognito. nothing seems to help. anyone had this problem before? Thanks have a great day! :)
Replies
0
Boosts
0
Views
84
Activity
3w
Report inactivity
Hey has someone figured out how or when does a report request become inactive? Is there like a fixed time period for how long an ONGOING accessType stays active? And does the same rule apply for ONE_TIME_SNAPSHOT? Has someone managed to create multiple active reports for both accessTypes per one app? And can you have multiple inactive ones?
Replies
0
Boosts
0
Views
75
Activity
3w
Unity iOS (Metal) → WebRTC (Unity WebRTC) video stream to remote Unity client: PeerConnection connects but receiver renders black frames
Hello! My name is Mason Prather. I'm a graduate student at Kennesaw State University and a Research Engineer working in XR environments through my Graduate Research Assistant role. I’m currently building a research prototype that connects a mobile companion application to a VR headset. The mobile application is built in Unity and deployed on iOS, and it streams video frames to a remote Unity client using WebRTC. Environment Device: iPhone 15 OS: iOS 26.3 (tested on physical device, not Simulator) Engine: Unity 2022.3.57f1 Graphics API: Metal Streaming Technology: WebRTC (Unity WebRTC package) Architecture: Mobile Unity app streaming video frames to a remote Unity client Receiver Device: Meta Quest Pro headset (Unity application) Networking: LAN (UDP discovery + TCP signaling) Video Source: Unity RenderTexture Goal The goal of the system is to allow a VR user to view media stored on their phone inside a VR environment. The iOS app: renders or captures media content converts frames into a WebRTC video track streams the video to the headset Current Status Connection setup works correctly. Observed behavior: Signaling connection successful ICE candidate exchange successful PeerConnection state becomes Connected Video track created successfully However, the receiving application displays black frames. iOS App Details The video source originates from a Unity RenderTexture. Inside the phone application: RenderTexture displays correctly Frames appear correct locally But the receiving peer does not display the frames. Relevant Components Unity WebRTC package iOS Metal rendering pipeline Custom TCP signaling LAN discovery via UDP Expected Behavior Rendered frames should transmit via WebRTC and appear on the remote device. Actual Behavior The remote video track is active, but the rendered frames appear black on the receiving client. Questions Are there known issues involving Unity WebRTC + iOS Metal texture capture? Are there specific pixel format requirements when streaming textures from Unity on iOS? Could the issue relate to texture readback limitations or GPU synchronization? I am more than happy to provide screenshots and console logs upon request. If anyone has experience streaming Unity video frames via WebRTC on iOS, I would greatly appreciate any guidance.
Replies
0
Boosts
0
Views
210
Activity
3w
Virtual Camera Shows Jittering Frames and Solid Accent Color on macOS
Hello Apple Developer Support, I’m developing a virtual camera using the CMIOExtensionDevice / CMIOExtensionStreamSource APIs on macOS. While the virtual camera appears in system settings and apps like Zoom and Google Meet, the video output exhibits the following issues: Jittering frames: The first frame sometimes appears correctly, but subsequent frames flicker or jitter. Solid color fill: Eventually, the camera feed fills entirely with a solid accent color (e.g., blue), rather than the intended video content. Console logs: Repeated messages appear in Console.app: Invalid display 0x00000000 Setup details: The virtual camera is created using CMIOExtensionDevice and CMIOExtensionStream. Video frames are rendered from NSImage/CGImage using CGContext and copied into CVPixelBuffers. Frame delivery is controlled by a DispatchSourceTimer at 60 FPS. macOS version: [Your macOS version here] Xcode version: [Your Xcode version here] Observations: The Invalid display 0x00000000 logs suggest that CGContext drawing or NSImage operations are failing in headless mode (i.e., there is no real display attached to the virtual camera). Using CIContext with .useSoftwareRenderer = true appears to mitigate some flicker, but not entirely. Questions / Requests: Is it expected that CoreMediaIO virtual cameras cannot reliably render CGImage / NSImage frames offscreen? Are there recommended APIs or approaches to render virtual camera frames fully headless to avoid display-dependent jitter? Is there any documentation or sample code from Apple showing stable video output from a virtual camera extension that does not rely on a physical display? Any guidance or examples would be greatly appreciated. This issue prevents the virtual camera from being used reliably in standard video apps. Thank you, Savvy
Replies
0
Boosts
0
Views
58
Activity
4w
Standard Memory Debugging Tools
In many cases I’ll be talking to folks with a memory management problem and I’ll say “You should investigate this with the standard memory debugging tools.” They then turn around and ask me “What are those tools?” Well, this is what I mean: Zombies — This lets you quickly detect when an object is used after it’s been deallocated. Learn more about it in the Finding zombies section of the Instruments Help and in Investigating crashes for zombie objects. There was also an excellent WWDC video about this, namely, WWDC 2010 Session 311 Advanced Memory Analysis with Instruments. This is no longer available in the video archive but if you can find a copy it’s well worth a watch. Address Sanitizer — This is a lower-level tool that finds a variety of common memory management issues, including use after free and buffer overruns. Learn more about this in Diagnosing memory, thread, and crash issues early and the various articles it links to. There’s also a good discussion of this tool, and other Xcode runtime diagnostic tools, in WWDC 2017 Session 406 Finding Bugs Using Xcode Runtime Tools (also no longer available from Apple). Hardware memory tagging — If you have access to a device with hardware memory tagging — starting with the A19 or M5 processors — consider enabling Memory Integrity Enforcement (MIE). For the details, see Enabling enhanced security for your app or watch Secure your app with Memory Integrity Enforcement. Even if you don’t want to do that for your production app, enable it just for development by checking Hardware Memory Tagging in the Options tab of Xcode’s scheme editor. Older tools — There are a variety of older tools that might be useful in some specific circumstances. See the Enabling the Malloc Debugging Features section of the Memory Usage Performance Guidelines for more information about these. Of specific interest is libgmalloc, which is documented in a UNIX man page. For some practical examples of how to identity a memory management crash report and then investigate that crash with these tools, take a look at WWDC 2018 Session 414 Understanding Crashes and Crash Logs. Share and Enjoy — Quinn “The Eskimo!” @ Developer Technical Support @ Apple let myEmail = "eskimo" + "1" + "@" + "apple.com" Revision History: 2026-03-02 Added a discussion of hardware memory tagging. 2023-05-09 Added a link to Investigating crashes for zombie objects. 2023-03-22 Removed another WWDC session video link. Made minor editorial changes. 2020-10-23 Fixed some formatting errors. 2019-10-30 Removed the link to WWDC 2010 Session 311 Advanced Memory Analysis with Instruments because it’s not long available in the archive. Refreshed all the other links. 2019-01-22 Fixed the link to libgmalloc. 2018-11-02 Updated to include a reference to WWDC 2018 Session 414 Understanding Crashes and Crash Logs. 2017-11-16 First posted.
Replies
0
Boosts
0
Views
18k
Activity
4w
macOS Tahoe 26.3 - System Is Playing NSBeep At Inappropriate Times When Text Editing Ends Via -cancelOperation: (field editor)
When I end editing pressing the escape key, the system sometimes plays NSBeep(). I noticed this with NSBrowser. Every time I press escape to end editing the system beeps. At first I thought it was somewhere in my app but I set a symbolic breakpoint and discovered it was not coming from my code. I filed FB22127038. Since then I discovered that NSBeep playing at inappropriate times is not exclusive to NSBrowser. It appears if there is a NSTableView in the window and you just press the escape key (even if you aren't editing anything) AppKit beeps. It can be traced to: #0 0x0000000199e3184c in NSBeep () #1 0x000000019aa03fac in -[NSWindow doCommandBySelector:] () #2 0x000000019ac1d01c in -[NSTableView(NSTableViewViewBased) cancelOperation:] ()
Replies
3
Boosts
0
Views
97
Activity
3w
How do I resolve the "Automatic signing cannot update bundle identifier..." error?
When I create an archive file and attempt to upload the app using the "Distribute App" button, the upload fails with the error "Automatic signing cannot update bundle identifier...". (The detailed message is below.) When creating an archive file in Xcode, I unchecked "Automatically Manage Signing" and proceeded with the archive. The message says "Font Enumeration," but other apps with the same option enabled upload successfully. Therefore, I believe the "Font Enumeration" option is not the issue. I tried creating a new provisioning file, but it still doesn't work. I deleted all DerivedData files from my Mac storage, restarted Xcode, and tried again, but it still doesn't work. This keeps happening only for certain targets (specific apps) in Xcode. Does anyone know how to fix this? Xcode is the latest version. Message: Automatic signing cannot update bundle identifier "com.xxxxxx.xxxxxx". Automatic signing cannot update your registered bundle identifier to enable Font Enumeration. Update your bundle identifier on https://developer.apple.com/account and then try again.
Replies
1
Boosts
0
Views
124
Activity
4w
LiveCallerId OHTTP Relay: Works in TestFlight, failing in Production (Bundle ID: no.opplysningen.bedrift.LiveCallerId)
We’ve been implementing LiveCallerId using OHTTP and have hit a wall with the production environment. The setup works perfectly in TestFlight, but the release version of the app is consistently being rejected by the Apple OHTTP Relay when trying to tunnel traffic to our gateway. Timeline & Status: Applied via the form in September 2025. Received confirmation in November 2025 that our /.well-known/ohttp-keys endpoint was correctly configured. Since then, we've struggled to get a dialogue with Apple to confirm the final production whitelisting. Technical Observations: Our ohttp-keys endpoint is being polled frequently (every few minutes). Based on the traffic, this is clearly the Apple Relay infrastructure fetching/refreshing the keys, not the devices themselves. This suggests the Relay "sees" our configuration, yet it still refuses to tunnel traffic to our gateway in the production environment. Since everything is functional in TestFlight, our implementation seems correct. It feels like there is a configuration mismatch or a missing "production flip" on the Relay side for our Bundle ID. If anyone from the Apple engineering team could verify the status for this Bundle ID, it would be a huge help. We've been stuck in this "TestFlight-only" state for quite a while now.
Replies
0
Boosts
0
Views
167
Activity
3w
Trapped in a "Black Box": Almost 3 weeks of App Review silence and ignored support tickets. Is this normal?
Hi everyone, I am writing this out of pure desperation because I honestly don't know what else to do or where to turn. My app has been stuck in the review process since February 9th (almost 3 weeks now), and I feel like I am completely trapped in a black box. I have tried every official channel available to get at least some information: I submitted requests for an expedited review. I opened 3 separate tickets with Developer Support, but received absolutely zero response. Just complete, deafening silence. I finally called the developer hotline. The representative was polite but completely unhelpful. I was basically told that "everything is fine, this timeline is normal, we can't do anything to help you, just keep waiting." Just to put this wait time into perspective, while sitting in this Apple "black box" waiting for a single message, I have managed to do the following: Registered a Google Play Developer account and fully verified my identity. Almost completely rewrote my entire app from scratch, translating it from Swift to Kotlin. Uploaded the Android version, got it approved, and I am already on Day 8 of closed testing in Google Play! Meanwhile, in App Store Connect, absolutely nothing has changed. No status updates, no rejections, no questions from the reviewers, no replies to my tickets. Just an absolute void. Has anyone else experienced this kind of extreme delay and complete lack of communication recently? For context, here are my details: Submission ID: db1ffb72-8f0f-49bc-b215-179e1e179b7d Apple ID: 6756239914 Any advice or shared experiences would be deeply appreciated. Thank you.
Replies
2
Boosts
0
Views
201
Activity
4w
False Positive "Shared Device" Flag blocking Enrolment (Case 102822291300)
Dear All, I am writing to report what appears to be a bug in the Apple Developer App’s enrollment verification system that is incorrectly flagging personal hardware. Problem Statement: I am unable to complete enrollment for the Apple Developer Program. On the "Personal Information" or "Identity Verification" stage, the app displays: "Contact Us to Continue. There may be an issue with your account that needs to be resolved..." Hardware History: Device: iPhone 14 (iOS latest) The device was previously owned by a family member who never enrolled in the Developer Program. The device was factory reset/wiped before my account was added. Despite this, Support (Case 102822291300) states the device is flagged as a "shared device" and refuses to proceed with identity verification. The Impasse: Support has indicated they cannot provide further assistance. However, as an individual developer, this is my primary device. If the system is flagging a reset device as "shared" when there is no history of developer use, it suggests a flaw in the automated verification logic. Request: Can a moderator or support representative look into Case 102822291300 to verify why this specific hardware is being restricted? Is there a manual path for identity verification (Passport/National ID) that doesn't rely on the Apple Developer App's automated device check? Allow me to do manual verification and provide the resolution to the issue. I am eager to bring my app to the platform and would appreciate a path to resolution. Kind Regards, Pradeep Bangalore
Replies
0
Boosts
0
Views
38
Activity
3w
Final Cut Pro not loading my plugin
I have exhausted standard debugging approaches and need guidance on Final Cut Pro's AU plugin loading behavior. PLUGIN OVERVIEW My plugin is an AUv2 audio plugin built with JUCE 8. The plugin loads and functions correctly in: Pro Tools (AAX) Media Composer (AAX) Reaper (AU + VST3) Logic Pro (AU) GarageBand (AU) Audacity (AU) DaVinci Resolve / Fairlight (AU) Harrison Mixbus (AU) Ableton Live (AU) Cubase (VST3) Nuendo (VST3) It does NOT load in Final Cut Pro (tested on 10.8.x, macOS 14.6 and 15.2). DIAGNOSTICS COMPLETED auval passes cleanly: auval -v aufx Hwhy Manu → AU VALIDATION SUCCEEDED Plugin is notarized and stapled: xcrun stapler validate → The validate action worked spctl --assess --type exec → Note: returns 'not an app' which we understand is expected for .component bundles. Hardened Runtime is enabled on the bundle. We identified that our Info.plist contained a 'resourceUsage' dictionary in the AudioComponents entry. We found via system logs that this was setting au_componentFlags = 2 (kAudioComponentFlag_SandboxSafe), causing FCP to attempt loading the plugin in-process inside its sandbox, where our UDP networking is denied. We removed the resourceUsage dict, confirmed au_componentFlags = 0 in the CAReportingClient log, and FCP now loads the plugin out-of-process via AUHostingServiceXPC_arrow. Despite au_componentFlags = 0 and out-of-process loading confirmed, the plugin still does not appear in FCP's effects browser. We also identified and fixed a channel layout issue — our isBusesLayoutSupported was not returning true for 5-channel layouts (which FCP uses internally). This is now fixed. AU cache has been fully cleared: ~/Library/Caches/com.apple.audio.AudioComponentRegistrar /Library/Caches/com.apple.audio.AudioComponentRegistrar coreaudiod and AudioComponentRegistrar killed to force rescan auval -a run to force re-registration
Replies
0
Boosts
0
Views
215
Activity
4w
XCode 26.3 AI assistant's attach file feature doesn't work in a big project
It doesn't show any files, only warning or errors for me. Same behavior with @
Replies
4
Boosts
0
Views
114
Activity
3w
Will AI change Swift Student Challenge in future years?
With AI getting so powerful, do you think Swift Student Challenge will change a lot next year? Will Apple update the rules or change what they're judging? Would love to hear thoughts from anyone who's followed SSC for a while.
Replies
0
Boosts
0
Views
189
Activity
4w