... and fingerprint authentication has problems too

Joseph Sinclair plug-discussion at stcaz.net
Thu Aug 4 03:36:52 MST 2016


This is a bit long, to here's the TL;DR:
If you just want to stop thieves, use tokens and biometrics for convenience.
If you want to protect privacy, use a password.
If you want to protect against both, start with a password, and add token, if you want more security then add a biometric on top of that.

If you have something encrypted with a modern symmetric algorithm (e.g. AES/Rijndael, Twofish, Serpent), then you don't need to worry about it being decrypted without your 128-bit key.
If you are expecting quantum computers to become cheap in the near future, or you're worried about Starfleet being willing to consume the rest-mass energy of a galaxy to decrypt your file, use a 256-bit key.
If you're really worried about a state willing to spend a few million USD on a side-channel attack (which goes around the cryptography, so your cipher and key don't matter much) against your one file or device, then keep your data somewhere other than a computer or phone and far away from anything resembling a network.

----------------------------------------------------------------------------------------------------
First, as to the biometric issue and state compulsion:
1) The Virginia decision is pretty unfortunate, they're basically saying that the fingerprints are like keys (which are not, of course, testimonial).  This goes with existing caselaw, however, and actually makes sense if you think about it like a lawyer would.
   Imagine you have private papers, sealed in a vault in your home.  The state cannot force you to tell them the combination (knowledge), but they can demand the key (if there is one) or cut open the safe if necessary. All that the state needs is a search warrant and the time/money to break open the safe, because it's the 4th amendment, which acknowledges your property rights, that applies in that case.
   Your phone is considered just another sealed vault where you keep your private papers.  You're expected to protect them from thieves, but the state can, and will, compel you to produce the keys to unlock it if they can.
2) Using a biometric or token without an accompanying password is fine for security, but bad for privacy.  All "good" digital security, *when privacy is the goal*, requires a password (or equivalent, "something you know").
   That locked-in-your-mind characteristic is the only thing that's protected by the rights acknowledged in the 5th amendment.
   A search warrant is the only requirement for something locked by a key (token, "something you have"), and the state may use any reasonable means to get in without the proper key (subject to warrant requirements).
   It seems clear, to me, from recent cases that biometrics ("something you are") are going to be treated the same way as a physical key by the courts, at best.
3) The more concerning issue is some recent cases where courts did not require a search warrant to break into and examine the contents of a smartphone.  That issue is far more worrisome, and it's an area where there is very valid basis to oppose (through the political process) the overreach of the state.
4) The use of biometrics on phones is predicated on the assumption that you have no privacy interest in your phone; the security of the phone is against theft, not against snoops or law enforcement.
   If you are trying to protect privacy, in addition to property (and intend to use biometrics), then after you've proved knowledge, and provided a physical key/token, you're challenged with a unique identifier you cannot change (yourself) and, ideally, cannot replicate as a final challenge.
   In the case of a phone, the passcode is knowledge, the phone is a token, and your biometric would be the final layer of authentication.
   If you are concerned about privacy on your phone (not just theft), then it currently falls on you to require a passcode/pin, instead of using the more convenient token or biometric options.

If you think about it, we've used a combination of token and biometric for quite some time (driver's license with photo or other state-issued document with photo) for government authentication, so anything you don't want the government to have should use something other than those two factors (possibly in combination with those two factors as added protection against thieves).
I may not like that the state intends to compel production of a biometric, but it doesn't, as far as I can tell, significantly conflict with long-standing jurisprudence either.

----------------------------------------------------------------------------------------------------
Second, as to encryption. If you're afraid of being targeted by a state with massive resources, note the following regarding data encryption:
1) Full disk encryption (any bulk encryption really, including the actual data transfer in TLS) uses a symmetric key, not an RSA key (which is only used to authenticate and encrypt the symmetric key, when used at all).
2) The most secure (general bulk data) encryption is actually symmetric key (e.g. AES/Rijndael, Serpent, or Twofish), public-key encryption (e.g. RSA or Elliptic-Curve) is really only useful (due to key sizes, computational complexity, cipher design limitations, etc...) in the key-exchange portion of a remote communication, or in some types of authentication (e.g. token).
3) For modern symmetric key ciphers you need only 128 bits for the key, 256 if you're truly paranoid.
4) Rijndael, Serpent, and Twofish are unbroken and considered extremely secure.  The full ciphers are not subject to any known attack better than brute-force.  There are known attacks against reduced-security versions of the algorithms, with Serpent having the largest security margin (difference between attackable versions and the full cipher) among them.
   In all cases, even the best known attacks against *weakened implementations* require more computing power than the entire planet can provide within the foreseeable future.

Basically, decrypting a well-studied block encryption cipher (such as Twofish or Serpent) with an open-source implementation and a decent key (20-30 well chosen characters) is currently beyond the capabilities of even massive state actors (such as China, Russia, or the USA), or even the entire planet working in concert.

Side channel attacks (such as those used to break ATM's, iPhones, or SSLv3) are the current preferred means, as they're actually possible, but incredibly expensive (in most cases, but not for SSLv3; seriously, don't use SSLv3 for anything).  Unless your data is worth the millions required to accomplish a side-channel attack against your system, you probably don't need to worry about these.
No key or cipher will protect from side-channel attacks as the side-channel attacks (by definition) bypass the cipher and attack the processing environment. Some newer (and less studied) ciphers like Threefish may limit the usefulness of some side-channel attacks, however, but even that is very limited added security.
If you're concerned about side-channel attacks (e.g. your data is worth more than 1 million USD to a state or other large organization wanting to read it), then you'd best keep that data off of anything using an SSD, CPU, or magnetic disk, and away from any device connected to a network.

----------------------------------------------------------------------------------------------------
Some basic numbers, for the curious:
Brute-force of a 128 bit key requires a median of 2^127 trials, which, at 1 microjoule/trial (far below current processors' power cost), would require roughly 1 billion times the total energy consumption (in all forms) of the world in 2010.  Even assuming there existed incomprehensible numbers of CPU's available to run the trials, it's beyond current human capability to power them all.
Brute-force of a 256 bit key would require an amount of energy exceeding the estimated theoretical rest-mass energy (including dark matter and dark energy) of the observable universe.
Quantum computing can reduce the effective key size by half (Grover's Algorithm), however, so if you expect quantum computers to become useful for this purpose in the next decade or so (unlikely), then you might prefer a 256 bit key. Even at 64 effective bits, however, it would take millions of quantum CPU-years to complete the process, for a single encrypted file.
The best side-channel attacks I know of are against the CPU (SSD is the next target, followed by magnetic disk and network connections), it requires, on average, 3 weeks of time using extremely specialized equipment and highly skilled engineers.  The process involves tens of thousands of iterations and only works against relatively short keys (such as a PIN).
The best side-channel attacks I know of against actively defended systems (systems that self-destruct in some way when an attack is detected, such as by deleting all data after some number of failed logins) involve painstaking work with an AFM and some extremely careful interaction with the active system cooled to extremely low temperatures.  These are also limited to relatively short keys (such as a PIN) and take considerable time.


On 08/02/2016 02:08 PM, Mike Bushroe wrote:
> This is scary. I would hope that as caselaw becomes more extensive and
> complete they split this into two parts. I have no qualms about allowing
> police to compel finger prints of any degree of fidelity. It is already
> standard practice to photograph and finger print every arrested person, so
> this is little change from decades worth of standard practice. However, I
> think they should split this when it comes to the step of fabricating a
> fake finger (or compelling the defendant to use their own finger) to unlock
> any secure data store (phone, computer, network file server(that is a scary
> hole in the security system itself!), safe, etc. If they have gone to the
> effort of securing access with a finger print then the information inside
> is clealry not 'public domain'. It should be as secure as their own
> testimony. It should be illegal for them to 'force' entry with a fake
> finger, and they should not be able to use any data inside, nor any further
> leads developed from data learned by that process. However, I am not
> confident that the world will be that reasonable.
> 
> A second weakness in my argument is that on TV, they routinely hack into
> someone's encrypted files and disks. If they can actually use anything that
> they can hack out of your computer, then faking a finger print to gain
> access is just a partially physical method of doing the same encryption
> hacking.
> 
> If this turns out to be the case, or becomes the case, then there would be
> no safe way to store *any* information digitally. And since having a
> passphrase to unlock a large encryption key is no more secure than the
> passphrase and encryption key storage program, I begin to wonder about
> encrypting entire disks with a 2048 bit key that IS the passphrase! Now we
> just need to learn how to do passphrases with 2048 bits of significant
> data. ugh! Upper and lowercase letters, digits, punctuation only give about
> 6 1/2 bits per character. That would need a 315 character pass phrase to
> remember and type in each time to get maximum security. And don't even
> *think* about writing it down somewhere! :)
> 
> Mike
> 
> 
>> The Smartphone versus the Fifth Amendment," Berkeley Technology Law
>> Journal, 21 Dec 2014[3]
>>
>>> in the aftermath of Virginia v. Baust, many smartphone users may soon
>>> reconsider their reliance on fingerprint ID technology.
>>>
>>
>> In October [2014], a Virginia trial judge ruled [in Virginia v. Baust]
>>> that unlike a passcode, the production of one's fingerprint is not
>>> "testimonial communication", and therefore, the Fifth Amendment privilege
>>> against self-incrimination cannot be invoked. Rather, the government may
>>> properly compel the production of a smartphone user's fingerprint to unlock
>>> the user's device. This force compulsion would ostensibly extend to any
>>> applications within a device that can be opened via fingerprint.
>>>
>>
>> However,
>>
>> As a trial court, the ruling in Virginia v. Baust is not mandatory law.
>>> However, as with any early caselaw in a novel and undeveloped area of the
>>> law, this opinion will likely be cited as a persuasive authority.
>>
>>
>>
>>
>> ---------------------------------------------------
>> PLUG-discuss mailing list - PLUG-discuss at lists.phxlinux.org
>> To subscribe, unsubscribe, or to change your mail settings:
>> http://lists.phxlinux.org/mailman/listinfo/plug-discuss

-------------- next part --------------
A non-text attachment was scrubbed...
Name: signature.asc
Type: application/pgp-signature
Size: 473 bytes
Desc: OpenPGP digital signature
URL: <http://lists.phxlinux.org/pipermail/plug-discuss/attachments/20160804/b9fa9848/attachment.pgp>


More information about the PLUG-discuss mailing list