I’d like to create a button on my iPhone Home Screen that changes settings if I am using an external lens mounted to my phone.
This is what I have along with notes on what I hope to do. But it does not work. I am only able to open either the camera settings or the camera preserve settings but nothing will toggle macro controls.
There is an additional step within the camera app as well but it can be skipped for brevity.
Name: Telephoto and macro lenses
Check if macro settings are already enabled or not. (Settings > Camera > Macro Control toggle) as well as (Settings > Camera > Preserve Settings > Macro Control toggle)
Non-working shortcut code:
Verification question: "Are you using an external lens?"
If yes,
Notification: "Turning on Macro Control and preserve settings for macro control." (Run below immediately)
prefs:root=CAMERA&path=Turn%20On%20Macro_Control
And
prefs:root=CAMERA&path=CameraPreserveSettingsSwitch&path=Turn%20On%20Macro_Control
If no,
Notification: "Turning off Macro Control and Preserve Settings for macro control." (Run below immediately)
prefs:root=CAMERA&path=Turn%20Off%20Macro_Control
And
prefs:root=CAMERA&path=CameraPreserveSettingsSwitch&path=Turn%20Off%20Macro_Control
Posts under iPhone tag
200 Posts
Sort by:
Post
Replies
Boosts
Views
Activity
(iOS 17.3)
I'm using the Apple supplied iOS sample project "ConfiguringAWiFiAccessoryToJoinTheUsersNetwork" as a base to write an App to configure an existing WiFi device using the NEHotspotConfiguration API's. I have almost everything working, and can join the network and send a packet to the device to configure it. I know that it is working as the device responds properly to what I send it. But I am not able to receive the response back from the device to the packet sent. (Only need 1 packet sent and 1 packet received)
However. If I run a packet sniffer on the phone before running my test App, then I do get a response. No packet sniffer running, no response.
When I do a debugDescription on the NWConnection after it reaches ".ready", I notice that when the sniffer is running I'm using loopback lo0:
[C1 connected 192.168.4.1:80 tcp, url: http://192.168.4.1:80, attribution: developer, path satisfied (Path is satisfied), viable, interface: lo0]
and I get a packet response in the NWConnection receiveMessage callback.
But with no sniffer running, I get interface en0:
[C1 connected 192.168.4.1:80 tcp, url: http://192.168.4.1:80, attribution: developer, path satisfied (Path is satisfied), viable, interface: en0[802.11], ipv4, dns, uses wifi]
and there is no callback to the receiveMessage handler and the NWconnection eventually times out.
The interface used seems to be the only difference that I can see when I have a sniffer running. Any ideas as to why I can't see a response in "normal" operation?
Hi, can someone explain me why there is so huge difference between download and install size? I get it that download is compressed but there difference is really huge.
I am creating an application that needs to connect to an Iot device, so i want to make a wifi hotspot with a custom SSID and password and WPA3.
Could you please provide an example code in Objective-C to get started?
I have a SPFx React application where I am printing the HTML page content using the javascript default window.print() functionality. Once I save the page as pdf from the print preview window and open it using Adobe Acrobat, the links(for eg -> Google) within the content are not clickable and appearing as plain text.
I have tried to print random pages post searching with any keywords in Google and saved the files as pdfs, but, unfortunately, the links are still not clickable there as well.
To check whether it is an Adobe Acrobat issue, I have performed the same print functionality from Android devices and shared the pdf file across the iOS devices and in that case, when opened using Adobe Acrobat, the links are appearing to be clickable.
I am wondering whether it is something related to how the default print functionality works for iPadOS and iOS devices. Any insights on this would be really helpful. Thanks!!!
Note: The links are clickable for MacOS as well as for Windows.
#ios #ipados #javascript #spfx #react
Hello Apple dev world and teams , I am experiencing anormal Waiting for Review for our app, now for more than a week and decided to post the issue here and hopefully get help.
We had successfully upgraded to organization membership and the process with Apple support was really enchanting really responsive, hoping to have the same experience with this issue for review also.
Topic:
App Store Distribution & Marketing
SubTopic:
App Review
Tags:
iPhone
App Submission
Apple Business Manager
Dears,
We are facing some issue in ios 18.4.1. Recently some of our end users who updated their ios devices to 18.4.1 have experienced random 403 errors in runtime. as per our analysis, We identified that these errors are associated with "CSRF token mismatch".
After successful login, the user's CSRF token is causing issue and it was changed in runtime, this causes the cookie mismatch, and the users is getting 403 errors, and the user session is getting invalid suddenly.
let me know if anyone facing the same issue in ios 18.4.1 and let me know Is there any workaround for this issue.
Thanks.
As we described on the title, the model that I have built completely works on iPhone 15 / A16 Bionic, on the other hand it does not run on iPhone 16 / A18 chip with the following error message.
E5RT encountered an STL exception. msg = MILCompilerForANE error: failed to compile ANE model using ANEF. Error=_ANECompiler : ANECCompile() FAILED.
E5RT: MILCompilerForANE error: failed to compile ANE model using ANEF. Error=_ANECompiler : ANECCompile() FAILED (11)
It consumes 1.5 ~ 1.6 GB RAM on the loading the model, then the consumption is decreased to less than 100MB on the both of iPhone 15 and 16. After that, only on iPhone 16, the above error is shown on the Xcode log, the memory consumption is surged to 5 to 6GB, and the system kills the app. It works well only on iPhone 15.
This model is built with the Core ML tools. Until now, I have tried the target iOS 16 to 18 and the compute units of CPU_AND_NE and ALL. But any ways have not solved this issue. Eventually, what kindof fix should I do?
minimum_deployment_target = ct.target.iOS18
compute_units = ct.ComputeUnit.ALL
compute_precision = ct.precision.FLOAT16
We are currently working on deploying a Java Card applet onto the eSIM (eUICC) inside an iPhone.
According to the GSMA SGP specifications, the eUICC is expected to support Java Card 3.0.5 Classic Edition. As defined in the Java Card 3.0.5 specification, the javacardx.crypto package should support standard algorithms including MessageDigest.ALG_HMAC_SHA_512.
However, during our testing on the iPhone's embedded eSIM, we found that ALG_HMAC_SHA_512 appears to be unsupported or disabled. The same applet functions correctly on external Java Card platforms that support Java Card 3.0.5, leading us to believe that this is a restriction specific to the iPhone’s eUICC implementation.
Our main questions are:
Why is ALG_HMAC_SHA_512, which is part of the standard Java Card 3.0.5 specification, not available on the iPhone eSIM?
Has Apple imposed any internal restrictions or exclusions on certain crypto algorithms for security, performance, or compliance reasons?
Is there a list or documentation of supported and unsupported Java Card APIs or algorithms on the eUICC used in iPhones?
Any insights from Apple engineers or other developers with experience on this topic would be greatly appreciated.
Thank you in advance!
I've developed an app in swift and UIKit. Its multi linguistic app supports English and Arabic. When I change the language from English to Arabic or Arabic to English. After changing language, when I navigate through different screens crash is happening randomly on different screens most of the time when I tap to navigate to a screen. And cash log is:
This time crashed with this error Exception NSException * "Not possible to remove variable:\t945: <unknown var (bug!) with engine as delegate:0x2824edf00>{id: 34210} colIndex:261 from engine <NSISEngine: 0x15c5dd5f0>{ delegate:0x15c594b50\nEngineVars:\n\t 0: objective{id: 31542} rowIndex:0\n\t 1: UIButton:0x15c6255b0.Width{id: 31545} rowIndex:1\n\t 2: 0x281c41450.marker{id: 31548} colIndex:1\n\t 3: UIButton:0x15c6255b0.Height{id: 31547} rowIndex:1073741824\n\t 4: 0x281c412c0.marker{id: 31546} colIndex:1073741825\n\t 5: UIButton:0x15c625a50.Width{id: 31549} rowIndex:11\n\t 6: 0x281c41270.marker{id: 31544} colIndex:2\n\t 7: UIButton:0x15c625a50.Height{id: 31551} rowIndex:1073741825\n\t 8: 0x281c414a0.marker{id: 31550} colIndex:1073741826\n\t 9: UILabel:0x15c625d10.Height{id: 31553} rowIndex:1073741826\n\t 10: 0x281c41590.marker{id: 31552} colIndex:1073741827\n\t 11: UIImageView:0x15c625870.Width{id: 31555} rowIndex:3\n\t 12: 0x281c41360.marker{id: 31554} colIndex:3\n\t 13: UIImageView:0x15c625870.Height{id: 31557} rowIndex:1073741827\n\t 14: 0x281c413b0.marker{id: 31556} colIndex:1073741828"... 0x0000000282fb11a0
For switching language I'm using this code snippet:
private func restartApp() {
guard let windowScene = UIApplication.shared.connectedScenes.first as? UIWindowScene,
let delegate = windowScene.delegate as? SceneDelegate,
let window = delegate.window else {
return
}
// Create a new root view controller
let vc : AppLoadingVC = AppRouter.instantiateViewController(storyboard: .Splash)
let nc = UINavigationController(rootViewController: vc)
ApplicationManager.sharedInstance.isUserLoggedIn = false
DispatchQueue.main.async {
if UserDefaults.isRTL {
UIView.appearance().semanticContentAttribute = .forceRightToLeft
SideMenuController.preferences.basic.forceRightToLeft = true
Localize.setCurrentLanguage("ar")
} else {
UIView.appearance().semanticContentAttribute = .forceLeftToRight
SideMenuController.preferences.basic.forceRightToLeft = false
Localize.setCurrentLanguage("en")
}
window.rootViewController = nc
window.makeKeyAndVisible()
}
}
Please anybody help me I've been stuck here since lot of days. I tried multiple things but all in vain.
I am developing the electronic part of product, it includes a Find-My features. I saw some forum that testing Find-My feature needs a CSR and testing token. Can anyone teach me how to apply CSR and testing token step by step?
Thank you very much.
Best regards
Sam Ng
I‘m getting tired of having to reinstall apps because of the amount of datas. I deleted alot of photos. Deleted apps. The thing that gets my storage full is the System data. I searched up for tutorials, it was full with clearing safari history or deleting old messages, offloading apps. I did them all, none of them decreased my system data and its at 12GB while my phone‘s GB is only up to 64GB. I use IOS 16 Iphone 11. Can someone help?
Many of us Bangladeshi iPhone users were upset when Apple changed the font to Bangla in the most recent iOS version (18.4.1). We prefer the old Bangla typeface. I want the old Bangla typeface to return, and so do we. Please consider this.
If I want to display my corporations digital position would it be considered legal to use an iPhone SIM card reader as it seems to have a “data” point that hints at the idea of stabilization of intelligence that be found both off line and online.
I did a little research and it said Apples proprietary patents are referring to the size, ejection, magnetic trigger etc which leads me to want to know if the idea of using a SIM card reader to position the corporation as it uses a $USD specimen data file to position the idea of the corporation into a tangible format.
The part I’m referring to is the iPhone 11 sim tray and reader. The picture refers to the idea of the point on the back being similar to the ***** eye as something similar to the dollar seal.
I want to be a winner in the industry, I’ve got 50 plus “winner web domains” to bring to life and it’s time to buckle down and evolve the digital economy but first I want to validate the legality of me using disassembled Apple products to configure a “digital unit” that displays a “cyber analog”.
DYLD, symbol '_CTRadioAccessTechnologyNR' not found, expected in '/System/Library/Frameworks/CoreTelephony.framework/CoreTelephony'
In the public release of iOS 18.4.1 and iPadOS 18.4.1, external input support for keyboards and mice is critically degraded. This issue affects both Apple-branded and third-party HID-compliant devices, over both wired USB-C and Bluetooth.
Tested Hardware:
• iPhone 16 Pro Max (256GB)
• iPad Pro (USB-C, latest gen), last gen iPad as well
Affected Devices:
• Apple Magic Mouse and Keys (wired USB-C/Bluetooth)
• Redragon K580RGBPRO (Bluetooth/wired USB-C)
• Razer Naga V2 Pro (Bluetooth/USB-C)
Symptoms:
• Severe keystroke delay and dropped input
• Modifier keys (Shift, Command, Option) fail intermittently
• Input degrades further with multiple HID devices connected
• Mouse input via Bluetooth exhibits pointer lag and jitter
• Occurs in all apps: Notes, Safari, Mail, text fields, password entries, etc.
• Identical results using Apple USB-C cables
Reproducibility:
100%. Clean boots, minimal background activity, and isolated environments (including Airplane Mode) do not resolve the issue. Identical behavior across both iPhone and iPad.
Expected Behavior:
All HID-compliant external input devices — particularly Apple-branded ones — should provide low-latency, reliable, and consistent input over both USB-C and Bluetooth, especially in a production iOS/iPadOS release.
Actual Behavior:
External keyboards and mice exhibit:
• Lag
• Dropped characters
• Failed modifiers
• Degraded mouse tracking
Even with the latest hardware and clean configurations.
Severity:
Critical.
This is a platform-level failure affecting I/O at the user interaction layer. Input reliability is non-negotiable — especially on $2000+ flagship devices using Apple’s own peripherals.
Closing Note (for Apple engineering & peer devs):
This is not a beta regression — it’s a public release flaw that undermines iOS and iPadOS usability for power users, professionals, and accessibility communities alike. That Apple Magic Keyboard, Redragon, and Razer gear all fail equally and consistently should be a wake-up call.
Apple: this needs to be escalated. Now.
External input — one of the most basic subsystems in any OS — is broken on your highest-end devices.
In our web application some functionalities will allow user to upload multiple images (More than 25 images) in a single page
It is working find in all OS and browsers except iOS
When user try to upload images directly from camera there will be some overlaps, duplication, missing etc.
This is happening in both Safari and Chrome, we had a thorough check in our application and found every thing is working fine from our end
You can reproduce the issue by creating a web page which accept more than 50 images (we tried the same in ASP MVC Core & PHP) and showing the images in order
access the page through your iPhone using Safari or Chrome
Try to upload images directly from your camera, try sequential images (Image of a stop watch, or some thing like that) so that you can easily identify the order of files uploaded
and check the listing page of uploaded image (Try these steps multiple times)
You can find some images are duplicated and some are missing
I’d love to see Apple implement a Bionic Reading feature as a system-wide accessibility option. This type of reading aid highlights the first part of each word in bold to help guide the eyes and improve comprehension.
It’s been shown to be especially helpful for people with ADHD, dyslexia, and other neurodivergent needs. Having a toggle in Settings > Accessibility would be life-changing.
Ideally, it could be:
• Enabled system-wide, or per-app
• Allow customization of how much of the word is bolded
• Available in Safari, Messages, Books, News, etc.
Hi all,
I’m running into a persistent linker error when building my Unity 6 project (IL2CPP, iOS target) that calls a Swift method via an Objective-C++ wrapper. Despite following all known steps, I keep getting:
Undefined symbols for architecture arm64:
"_placeGeoAnchor", referenced from:
_GeoAnchorTrigger_placeGeoAnchor in libGameAssembly.a
...
ld: symbol(s) not found for architecture arm64
I’m trying to place a persistent AR anchor at real-world GPS coordinates (so that the same asset can appear at the same location for a returning user). Since I’m targeting iOS, I can’t use Google’s geospatial anchors (but I sooo wish I could--please apple I beg of you stop being so selfish lol).
I've already done these things:
Swift file is added to Unity-iPhone target.
.mm and .h files are in Unity-iPhone target under Compile Sources.
Bridging header is set to Unity-iPhone-Bridging-Header.h.
Generated header name is correct (GeoTest-Swift.h).
Build Active Architecture Only set to No.
Function has attribute((visibility("default"))).
Unity project uses IL2CPP scripting backend.
Yet I'm still getting the same linker error — it appears Unity (via IL2CPP) references the function, but Xcode doesn't link it.
It’s something small that’s being missed with how IL2CPP links native symbols? Or maybe I need to explicitly include something in Link Binary With Libraries? I’ve verified symbol visibility and targets repeatedly.
I’ve built AR features in Unity before (for Quest), but this is my first time trying to bridge C# → Objective-C++ → Swift in this way for a geolocation-based AR anchor for an iphone.
I'm going crazy, I’ve been stuck on this for 12+ hours now, so any insight or nudge would be deeply appreciated.
SPECS:
Macbook Pro M4 Pro--Sequoia 15.4
Unity 6000.0.45f1
IPhone 11 iOS 18.4
Xcode 15
Dear Apple Engineers,
Recently, I have switched from the iPhone 13 Pro to the iPhone 14 Pro, and every single notification has not lit up my screen even though I heard the notification sounds as well as the vibrations. I have made sure dnd is off, no other focus mode is on, attention aware features are off, no Apple Watch connected, forced restart twice, normal restart multiple times.
Upon more research, I have discovered that it has been a prominent issue since September 2023 as you can see from this Apple discussion forum.
https://discussions.apple.com/thread/255133907?answerId=260178222022&sortBy=newest_first#260178222022
I find it deeply concerning that, despite reports from multiple users, Apple has yet to provide a permanent solution or acknowledge the defect publicly. This lack of action undermines consumer trust. I hope to resolve this issue amicably and continue supporting Apple as a brand I trust.
Thank you for your immediate attention to this matter.