The impulse for writing my first post in this thread was frustration about a "technological" treatment that privacy often receives, and about a lobby that tries to tell everybody to encrypt everything, whether sensitive or not (I think I've seen this on this list, but I don't have the time to research now). The argument for using encryption seems to go like this: "privacy (value) is always good, privacy (secrecy) is achieved by encryption, therefore encryption is always desirable". I'm bothered when I read "privacy is achieved by encryption" (johnny1) - okay, maybe they use it as a synonym of "secrecy", but then it blurs the distinction between the two.
In my posts I just wanted to articulate one reason why for some people (from "out in the field") encryption is a non-choice, and I gave some context in which such attitude might be understandable. In further discussion I tried to describe Privacy as a social mechanism and show it's not equivalent to secrecy (i.e. information leakage does not necessarily mean loss of privacy), and therefore the above argument is non sequitur. I don't feel the best person to discuss these things, but I thought someone must first say an idea, so that others can use it and test it. Before writing my first post, I had read the Gaw et al. article (by your recommendation, actually), and I was a little less than satisfied (perhaps because authors didn't make an effort to hide their opinions). For this writing, I have read all the other articles mentioned in someone's earlier post. Please, allow me to add one more thought, which is relevant further down. It might sound paradoxical, but openness is what protects us in our lives. We require transparency from the government, from institutions, from private companies (in certain circumstances), and therefore we practice a culture of openness ourselves. This is how we keep institutions under control, and this is what keeps society together. People are more likely to support a transparent organisation. People are upset when public institutions become too secretive. Facing an opponent, to avoid fight, we may run away (nothing wrong, a normal defense), or challenge them (it often works). I think I can understand why some people (e.g. Jenny) feel encrypting "public" information is not appropriate: it's a challenge. "We had a meeting at 9, coffee was served, so what? We are worried about new law - every citizen has the right to be concerned. We investigate what BigCorp does - of course, we're ActivistCorp, it's our job, that's what our supporters pay us to do. Now, what are _you_ up to?". Secrecy would probably be not adequate, because then police could use any pretense to enter the offices to hamper the activity. Transparency also helps keep internal discipline (don't do stupid things). "Robert J. Hansen" <r...@sixdemonbag.org> wrote: > On 8/26/12 5:37 PM, Stan Tobias wrote: > > In the works cited before (this thread and other discussions), > > one recurring concern could be formulated as: "Why Johnny doesn't > > encrypt his Christmas greetings to his granny?", with an implicit > > assumption/expectation that everybody ought to use cryptography by > > default for any and everything. I'll concentrate on the encryption only. > > Well, speaking just for myself, I try not to make that assumption. I'm > interested in knowing why Johnny can't encrypt, and then further why > Johnny *doesn't* encrypt. These are two different questions which have > very different answers. I didn't adapt the title without a reason, my answer was directed towards that attitude. "What will it take to make the use of encrypted e-mail universal and routine?" is a quote from Gaw et al. Do we really have evidence people can't encrypt? For me the "johnny" articles were not quite clear about it (they seemed to investigate a different aspect). I don't believe people are stupid. They can learn to use cryptography, just as they have learned many other things in their lives. Another matter is what "can" does mean. I can fly an Airbus A380. (Sure :-^, I only have to find that "A/P" button.) - That was my conclusion having read "All users were able to decrypt. This is because PGP automatically decrypts emails when they appear in Outlook Express." (Sheng et al.). I think what is missing from many discussions is that to be able to _effectively_ use cryptography on computers, one has to know much more than how to use cryptography. (I could tell a funny story of a security failure, not because of wrong use of GnuPG, but because they didn't realize how a certain file system worked.) Can you imagine a responsible person exchanging sensitive information, while not being certain what he does is safe? It's a matter of personal integrity, it's not enough to tell a user "click here and there, and you're fine"; we have to first convince ourselves what we do is right. The upshot is that you cannot make cryptography easier for users, they will have to study and understand it themselves anyway. > "Why Johnny can't encrypt" is a human-computer interaction (HCI) > problem. HCI problems are eminently solvable. The papers have a lot of > exploration of this problem: see, e.g., "Why Johnny Can't Encrypt", > "Johnny 2", and "Why Johnny Still Can't Encrypt" for three examples of > really good peer-reviewed papers that explore this. One reason I didn't like those papers is that they concentrated on a particular version of a particular implementation, and they seemed to make mostly ignorant users to work with it (and by my standards, they were actually quite successful in that). Their methods and findings could be applied to any graphical software. As I said, for me to be able to use encryption means more than knowing which buttons to click. > "Why Johnny doesn't encrypt" is a social problem. Social problems are > notoriously intractable. I disagree here. They might be difficult to quantify, but we can discuss those issues and try to reach some conclusions, if not solutions. One reason I like to read Schneier's blog (if I have the time) is that he often discusses social aspects to security. > See, e.g., Gaw, Felten and Fernandez-Kelly's > paper. They found that even when people were aware of the dangers they > were facing, knew those dangers were real, had easy access to crypto > software and had been trained in its use, they *still* weren't using > crypto... principally because they didn't want to be seen as paranoid. In the article I didn't find anybody who said they didn't want to be _seen_ as paranoid, they only described certain behaviour as paranoid. I think people use the word "paranoid" when something conflicts with their perception of the world, and they don't know how to phrase it. Cryptography might not be difficult to apply, but is not without its own problems *around* it. In my small experience, it requires a lot of planning: what and why must be encrypted; what passwords will be used (with many encrypted files) and how to ensure I don't forget those passwords; how to ensure the encrypted files don't become corrupted (so that data doesn't become irrecoverable), how do I check that they're not corrupted, how many times will I check the files, where and how to make backups (of course, check goes before the backups, but you have to remember this), and many other small details in a long decision tree. Recently I helped a friend to recover data from a broken NTFS partition - mainly family pictures since many years ago. Had the disk been encrypted, the chances of recovering anything (by me, at least) would be close to zero. One has to be able to balance the risk of leaking information against completely losing it, and it's a big headache, especially that we don't realize all the factors that come into play. Using cryptography to protect secrets is a serious intellectual effort. Abe (in Gaw) described it as a "chore", and I think I can understand what he meant by that. If you want others to use cryptography communicating with you, you want to put the same burden on their backs. "Paranoid" in this case does not mean "tin-hat"; it just means that the effort you put into message exchange is not proportional to its value. As for the people you mentioned, I don't see exactly which you mean, I didn't find any egregious carelessness (except that users didn't understand digital signatures, but that wasn't a big issue either). Ultimately, it's the organization management's responsibility to decide to what degree information must be protected. Should cleaners be required to encrypt their emails? What about the plumbers? And encrypting often isn't the only (or most important) matter to think about (WikiLeaks "thank you" email fiasco). > I really don't want to rain on people's parades. A lot of these ideas > of "what the problem is" are deeply interesting. But until you actually > go out into the world and ask real users the question, and observe > workers in their natural environment, then it's a bunch of discussion > over how many angels can dance on the head of a pin. Facts by themselves are not knowledge. We gain insight by discussing facts. It's important to discuss before a next round of interviews, because then we can know better what questions to ask, and how to ask. My experience is generally in agreement with the findings of the articles. My only addition here is that I try to rationalize certain behaviour, which is something that the interviewees could not do on the spot, because they probably acted on instinct rather than calculated risk. I might be wrong about Jane, but I'm not wrong about myself, and my writing here is another testimony. There's no need to go far, this mailing list is a mine of issues people deal with, of reasons why they do or don't encrypt, and of their good and bad perceptions. If someone had the time to sort all those things out, it might result in another great scientific paper. Best regards, Stan. P.S.1. Having an occasion now, I just want to say to you, Robert, a big and sincere "Thank You!" for your articles on this mailing list. P.S.2. I mentioned British police once - they still don't wear guns: http://www.bbc.co.uk/news/magazine-19641398 P.S.3. Thanks to others who responded, and especially to Marco - after my second writing I got your reply first, and it was very nice and encouraging. It's not very important what I have to say, and only slightly topical, but because you've asked, here goes, but briefly: Facebook users have often been accused of carelessness about their privacy, to plain foolishness (by "facebook" I understand any web-page where users publish themselves, e.g. Wikipedia). Once I read a discussion on sexting among teenagers; the conclusion why they do it was that as they grow, they announce this fact to others, it's in their nature. (Some were harshly punished, but IMO not by life, but rather by self-righteous adults, who disregarded their normality.) I think the same applies to grown-ups; I've seen people publishing uninteresting things about themselves without purpose, I can't explain it other than by a need to announce one's presence to the world - it's something in the human nature. Much "Internet" time has passed, and I haven't seen any privacy disaster yet. People reveal a lot, but not everything; they make (sometimes funny) mistakes, but they also learn from those mistakes. Privacy is not all black-and-white, and we have room to try how much we can reveal, and when it becomes too much. I think a lot of good results from people publishing themselves en masse: we learn about other people, but most importantly *we learn about ourselves*. This helps _break social taboos_, and bring down barriers between people. We keep in contact with other people, share ideas, organize ourselves, we can influence political changes. People seek other people, it's in their nature. "Foolishness" is part of human lives; I think what attracts people, is that they can make a mistake, look foolish, and still maintain dignity, because everyone else around is equally "foolish". These are extremely important things, they help us to grow up, and they change our (global) culture. Ultimately, it may occur that there are not so many things really private, because essentially we all look the same, and do the same things. I think we sometimes overestimate the negatives when people publish their lives. When one person comes out naked into the street, it's a sensation. When one thousand become naked, then effectively noone is. Things that were inappropriate twenty years ago, are not so today. One concern does remain: the published information remains, and we can never be sure that it won't be used in future against us in ways yet unknown to us. Well, life is a risk. We must evaluate what's more important, creating more good and freedom for ourselves, or avoiding the risk. I see it as a race who will be there first, ordinary people establishing a new standard of normality, or the self-righteous - will they sense change in time and start regulating our lives again? Last: here on this list we reveal a lot about ourselves, too. If you ask questions, or help someone with understanding cryptography, you reveal you know something about it. This information is potentially more sensitive than what someone ate, or where they were on vacation, and could be used against you. So calling facebook people foolish on this list is... well, paradoxical at least. _______________________________________________ Gnupg-users mailing list Gnupg-users@gnupg.org http://lists.gnupg.org/mailman/listinfo/gnupg-users