According to the documents, Cellebrite could not unlock any iPhones running iOS 17.4 or newer as of April 2024, labeling them as “In Research.” For iOS versions 17.1 to 17.3.1, the company could unlock the iPhone XR and iPhone 11 series using their “Supersonic BF” (brute force) capability. However, iPhone 12 and newer models running these iOS versions were listed as “Coming soon.”
The Android support matrix showed broader coverage for locked Android devices, though some limitations remained. Notably, Cellebrite could not brute force Google Pixel 6, 7, or 8 devices that had been powered off. The document also specifically mentioned GrapheneOS, a privacy-focused Android variant reportedly gaining popularity among security-conscious users.
Links to the docs:
GrapheneOS has a thread about this on Mastodon, which adds a bit more detail:
Cellebrite was a few months behind on supporting the latest iOS versions. It’s common for them to fall a few months behind for the latest iOS and quarterly/yearly Android releases. They’ve had April, May, June and July to advance further. It’s wrong to assume it didn’t change.
404media published an article about the leaked documentation this week but it doesn’t go into depth analyzing the leaked information as we did, but it didn’t make any major errors. Many news publications are now writing highly inaccurate articles about it following that coverage.
The detailed Android table showing the same info as iPhones for Pixels wasn’t included in the article. Other news publications appear to be ignoring the leaked docs and our thread linked by 404media with more detail. They’re only paraphrasing that article and making assumptions.
We received Cellebrite’s April 2024 Android and iOS support documents in April and from another source in May before publishing it. Someone else shared those and more documents on our forum. It didn’t help us improve GrapheneOS, but it’s good to know what we’re doing is working.
It would be a lot more helpful if people leaked the current code for Cellebrite, Graykey and XRY to us. We’ll report all of the Android vulnerabilities they use whether or not they can be used against GrapheneOS. We can also make suggestions on how to fix vulnerability classes.
In April, Pixels added a reset attack mitigation feature based on our proposal ruling out the class of vulnerability being used by XRY.
In June, Pixels added support for wipe-without-reboot based on our proposal to prevent device admin app wiping bypass being used by XRY.
In Cellebrite’s docs, they show they can extract the iOS lock method from memory on an After First Unlock device after exploiting it, so the opt-in data classes for keeping data at rest when locked don’t really work. XRY used a similar issue in their now blocked Android exploit.
GrapheneOS zero-on-free features appear to stop that data from being kept around after unlock. However, it would be nice to know what’s being kept around. It’s not the password since they have to brute force so it must be the initial scrypt-derived key or one of the hashes of it.
They’re exploiting vulnerabilities and back doors not brute forcing your passcode. The only way you’re keeping them out is with hardware encryption which the iPhone has and probably why it’s the only one not vulnerable. Hardware encryption also won’t matter if your vendor shares their keys with law enforcement. As far as I’m aware, Apple is the only one that’s gone to court and successfully defended their right to refuse access to encryption keys.
Don’t put anything incriminating on your phones.
Huh? None of this makes sense to me. Apple and Google do not have encryption keys for your device, regardless of hardware or software. When your phone is unlocked, the key and/or hash is in memory, and may be key wrapped, regardless of hardware or software. iOS is listed as vulnerable to the attacks, minus the latest devices and versions, which is pretty much the norm with lagging development. Same woth latest Pixels. Where are you getting this info from?
https://support.apple.com/guide/security/hardware-security-overview-secf020d1074/web
The FBI wanted access to Apple’s encryption keys which they use to sign their software. They don’t have ‘your’ encryption keys, they have their own that the FBI wanted to use to bypass these features. They eventually dropped it because they found a zero day exploit which apple fixed in later versions. That is why the newer phones aren’t vulnerable (yet).
https://en.wikipedia.org/wiki/Apple–FBI_encryption_dispute
Familiar, but based on your first comment about the benefits of hardware encryption over software encryption, and thus iOS being better than Android, perhaps you’re misinterpretting the specifics?
For the first point, the SE only stores keys at rest. The keys and hashes are still in memory when booted, otherwise the device wouldn’t be able to function. This works the same as software encryption, the key itself is just encrypted and stored on “disk” instead of in flash when off.
For the second, Apple’s software signing keys would not give the FBI access to a device. There is nothing to “turn over”. The signing of new software to bypass the lock was to remove the 10 retry reset. As above, there is no benefit to hardware encryption over software here.
The benefit hardware encryption brings is potential speed (which is certainly valuable, but not necessarily more secure or harder to crack).
I’m not claiming iPhones are superior. I don’t care about dumb OS wars, just don’t put things on your phone expecting that they can’t be retrieved. That’s the only point I’m trying to make here.
And the keys absolutely would give them access since those keys are used to sign Apple software which runs with enough privileges to access the encryption keys stored in the “Secure Enclave”. Anything you entrust to a company’s software is only as secure as the company wants to make it, and the only company to publicly resist granting that acces is Apple (so far)
This was the hardware vs software comment I was debating, not the rest.
Also, software signing keys (like those requested by the FBI) would work for enabling brute force since that’s a change to the software, but not for direct access into SE. That would be like saying a firmware update could grant access to a LUKS partition without the passphrase. Not possible. If it was, no open source encryption would ever work.
The only thing that has successfully managed to thwart the FBI in their attempts to break into a phone was Apple’s hardware based encryption. To such an extent that they took legal and legislative actions to try and circumvent it. The specifics of how the encryption works is irrelevant to this argument, and you are more than welcome to consider that point conceded.
If it’s not the “Apple’s” and also not the “hardware based encryption”, what’s the argument then? As you pointed it out FBI only needed apple’s help since they didn’t have a working exploit and dropped it once a new one was found. In the latest case with android once again their existing tools didn’t work but cellibrite had an unreleased one ready to use, so they didn’t need to go to court in the meantime.
Exactly, thank you. They seem to think hardware encryption is a form of magic here, and are missing the key point. Sure, we are needlessly debating on the interwebs, but this type of thing is what leads to those that don’t understand the specifics gravitating towards the wrong solutions for their needs.
In a before-first-unlock state they absolutely are bruteforcing, since the filesystem is encrypted. The exploits are for bypassing the retry limit in that case.
And the manufacturers don’t have your encryption keys. They’re unique.