CRYPTO-GRAM
March 15, 2009
by Bruce Schneier
Chief Security Technology Officer, BT
schneier@schneier.com
http://www.schneier.com
A free monthly newsletter providing summaries, analyses, insights, and
commentaries on security: computer and otherwise.
For back issues, or to subscribe, visit
<http://www.schneier.com/crypto-gram.html>.
You can read this issue on the web at
<http://www.schneier.com/crypto-gram-0903.html>. These same essays
appear in the "Schneier on Security" blog:
<http://www.schneier.com/blog>. An RSS feed is available.
** *** ***** ******* *********** *************
In this issue:
Perverse Security Incentives
Privacy in the Age of Persistence
News
Insiders
The Doghouse: Singularics
Three Security Anecdotes from the Insect World
The Kindness of Strangers
New eBay Fraud
Schneier News
IT Security: Blaming the Victim
Balancing Security and Usability in Authentication
Comments from Readers
** *** ***** ******* *********** *************
Perverse Security Incentives
An employee of Whole Foods in Ann Arbor, Michigan, was fired in 2007 for
apprehending a shoplifter. More specifically, he was fired for touching
a customer, even though that customer had a backpack filled with stolen
groceries and was running away with them.
I regularly see security decisions that, like the Whole Foods incident,
seem to make absolutely no sense. However, in every case, the decisions
actually make perfect sense once you understand the underlying
incentives driving the decision. All security decisions are trade-offs,
but the motivations behind them are not always obvious: They're often
subjective, and driven by external incentives. And often security
trade-offs are made for nonsecurity reasons.
Almost certainly, Whole Foods has a no-touching-the-customer policy
because its attorneys recommended it. "No touching" is a security
measure as well, but it's security against customer lawsuits. The cost
of these lawsuits would be much, much greater than the $346 worth of
groceries stolen in this instance. Even applied to suspected
shoplifters, the policy makes sense: The cost of a lawsuit resulting
from tackling an innocent shopper by mistake would be far greater than
the cost of letting actual shoplifters get away. As perverse it may
seem, the result is completely reasonable given the corporate incentives
-- Whole Foods wrote a corporate policy that benefited itself.
At least, it works as long as the police and other factors keep
society's shoplifter population down to a reasonable level.
Incentives explain much that is perplexing about security trade-offs.
Why does King County, Washington, require one form of ID to get a
concealed-carry permit, but two forms of ID to pay for the permit by
check? Making a mistake on a gun permit is an abstract problem, but a
bad check actually costs some department money.
In the decades before 9/11, why did the airlines fight every security
measure except the photo-ID check? Increased security annoys their
customers, but the photo-ID check solved a security problem of a
different kind: the resale of nonrefundable tickets. So the airlines
were on board for that one.
And why does the TSA confiscate liquids at airport security, on the off
chance that a terrorist will try to make a liquid explosive instead of
using the more common solid ones? Because the officials in charge of
the decision used CYA security measures to prevent specific, known
tactics rather than broad, general ones.
The same misplaced incentives explain the ongoing problem of innocent
prisoners spending years in places like Guantanamo and Abu Ghraib. The
solution might seem obvious: Release the innocent ones, keep the guilty
ones, and figure out whether the ones we aren't sure about are innocent
or guilty. But the incentives are more perverse than that. Who is
going to sign the order releasing one of those prisoners? Which
military officer is going to accept the risk, no matter how small, of
being wrong?
I read almost five years ago that prisoners were being held by the
United States far longer than they should, because "no one wanted to be
responsible for releasing the next Osama bin Laden." That incentive to
do nothing hasn't changed. It might have even gotten stronger, as these
innocents languish in prison.
In all these cases, the best way to change the trade-off is to change
the incentives. Look at why the Whole Foods case works. Store
employees don't have to apprehend shoplifters, because society created a
special organization specifically authorized to lay hands on people the
grocery store points to as shoplifters: the police. If we want more
rationality out of the TSA, there needs to be someone with a broader
perspective willing to deal with general threats rather than specific
targets or tactics.
For prisoners, society has created a special organization specifically
entrusted with the role of judging the evidence against them and
releasing them if appropriate: the judiciary. It's only because the
George W. Bush administration decided to remove the Guantanamo prisoners
from the legal system that we are now stuck with these perverse
incentives. Our country would be smart to move as many of these people
through the court system as we can.
This essay originally appeared on Wired.com.
http://www.wired.com/politics/security/commentary/securitymatters/2009/02/securitymatters_0226
or http://tinyurl.com/aku6bf
Whole Foods incident:
http://www.mlive.com/news/index.ssf/2007/12/grocery_worker_fired_for_stopp.html
or http://tinyurl.com/3dma49
King County ID checks:
http://www.kingcounty.gov/safety/sheriff/Services/Gun.aspx
Terrorists as liquid bombers:
http://www.schneier.com/blog/archives/2007/08/details_on_the_1.html
CYA security:
http://www.schneier.com/blog/archives/2007/02/cya_security_1.html
The perverse incentives of holding terrorist suspects in custody:
http://query.nytimes.com/gst/fullpage.html?res=9C00E3DF133EF934A15756C0A9629C8B63&sec=&spon=&pagewanted=all
or http://tinyurl.com/cgh86n
** *** ***** ******* *********** *************
Privacy in the Age of Persistence
(Note: This isn't the first time I have written about this topic, and it
surely won't be the last. I think I did a particularly good job
summarizing the issues this time, which is why I am reprinting it.)
Welcome to the future, where everything about you is saved. A future
where your actions are recorded, your movements are tracked, and your
conversations are no longer ephemeral. A future brought to you not by
some 1984-like dystopia, but by the natural tendencies of computers to
produce data.
Data is the pollution of the information age. It's a natural byproduct
of every computer-mediated interaction. It stays around forever, unless
it's disposed of. It is valuable when reused, but it must be done
carefully. Otherwise, its after effects are toxic.
And just as 100 years ago people ignored pollution in our rush to build
the Industrial Age, today we're ignoring data in our rush to build the
Information Age.
Increasingly, you leave a trail of digital footprints throughout your
day. Once you walked into a bookstore and bought a book with cash. Now
you visit Amazon, and all of your browsing and purchases are recorded.
You used to buy a train ticket with coins; now your electronic fare card
is tied to your bank account. Your store affinity cards give you
discounts; merchants use the data on them to reveal detailed purchasing
patterns.
Data about you is collected when you make a phone call, send an e-mail
message, use a credit card, or visit a website. A national ID card will
only exacerbate this.
More computerized systems are watching you. Cameras are ubiquitous in
some cities, and eventually face recognition technology will be able to
identify individuals. Automatic license plate scanners track vehicles in
parking lots and cities. Color printers, digital cameras, and some
photocopy machines have embedded identification codes. Aerial
surveillance is used by cities to find building permit violators and by
marketers to learn about home and garden size.
As RFID chips become more common, they'll be tracked, too. Already you
can be followed by your cell phone, even if you never make a call. This
is wholesale surveillance; not "follow that car," but "follow every car."
Computers are mediating conversation as well. Face-to-face conversations
are ephemeral. Years ago, telephone companies might have known who you
called and how long you talked, but not what you said. Today you chat in
e-mail, by text message, and on social networking sites. You blog and
you Twitter. These conversations -- with family, friends, and colleagues
-- can be recorded and stored.
It used to be too expensive to save this data, but computer memory is
now cheaper. Computer processing power is cheaper, too; more data is
cross-indexed and correlated, and then used for secondary purposes. What
was once ephemeral is now permanent.
Who collects and uses this data depends on local laws. In the US,
corporations collect, then buy and sell, much of this information for
marketing purposes. In Europe, governments collect more of it than
corporations. On both continents, law enforcement wants access to as
much of it as possible for both investigation and data mining.
Regardless of country, more organizations are collecting, storing, and
sharing more of it.
More is coming. Keyboard logging programs and devices can already record
everything you type; recording everything you say on your cell phone is
only a few years away.
A "life recorder" you can clip to your lapel that'll record everything
you see and hear isn't far behind. It'll be sold as a security device,
so that no one can attack you without being recorded. When that happens,
will not wearing a life recorder be used as evidence that someone is up
to no good, just as prosecutors today use the fact that someone left his
cell phone at home as evidence that he didn't want to be tracked?
You're living in a unique time in history: the technology is here, but
it's not yet seamless. Identification checks are common, but you still
have to show your ID. Soon it'll happen automatically, either by
remotely querying a chip in your wallets or by recognizing your face on
camera.
And all those cameras, now visible, will shrink to the point where you
won't even see them. Ephemeral conversation will all but disappear, and
you'll think it normal. Already your children live much more of their
lives in public than you do. Your future has no privacy, not because of
some police-state governmental tendencies or corporate malfeasance, but
because computers naturally produce data.
Cardinal Richelieu famously said: "If one would give me six lines
written by the hand of the most honest man, I would find something in
them to have him hanged." When all your words and actions can be saved
for later examination, different rules have to apply.
Society works precisely because conversation is ephemeral; because
people forget, and because people don't have to justify every word they
utter.
Conversation is not the same thing as correspondence. Words uttered in
haste over morning coffee, whether spoken in a coffee shop or thumbed on
a BlackBerry, are not official correspondence. A data pattern indicating
"terrorist tendencies" is no substitute for a real investigation. Being
constantly scrutinized undermines our social norms; furthermore, it's
creepy. Privacy isn't just about having something to hide; it's a basic
right that has enormous value to democracy, liberty, and our humanity.
We're not going to stop the march of technology, just as we cannot
un-invent the automobile or the coal furnace. We spent the industrial
age relying on fossil fuels that polluted our air and transformed our
climate. Now we are working to address the consequences. (While still
using said fossil fuels, of course.) This time around, maybe we can be a
little more proactive.
Just as we look back at the beginning of the previous century and shake
our heads at how people could ignore the pollution they caused, future
generations will look back at us -- living in the early decades of the
information age -- and judge our solutions to the proliferation of data.
We must, all of us together, start discussing this major societal change
and what it means. And we must work out a way to create a future that
our grandchildren will be proud of.
This essay originally appeared on the BBC.com website.
http://news.bbc.co.uk/1/hi/technology/7897892.stm
National ID cards:
http://www.schneier.com/essay-160.html
Surveillance cameras:
http://www.schneier.com/essay-225.html
RFID chips:
http://epic.org/privacy/rfid/
Cell phone surveillance:
http://computerworld.com/action/article.do?command=viewArticleBasic&articleId=9127462
or http://tinyurl.com/au2f4n
Wholesale surveillance:
http://www.schneier.com/essay-147.html
Data mining:
http://www.schneier.com/essay-108.html
The future of surveillance:
http://www.schneier.com/essay-109.html
Face recognition:
http://epic.org/privacy/facerecognition/
Privacy and the younger generation:
http://nymag.com/news/features/27341/
Ill effects of constant surveillance:
http://news.bbc.co.uk/1/hi/uk_politics/7872425.stm
The value of privacy:
http://www.schneier.com/essay-114.html
** *** ***** ******* *********** *************
News
Uni-ball is using fear to sell its hard-to-erase pen -- but it's the
wrong fear. They're confusing check-washing fraud, where someone takes
a check and changes the payee and maybe the amount, with identity theft.
And how can someone steal money from me by erasing and changing
information on a tax form? Are they going to cause my refund check to
be sent to another address? This is getting awfully Byzantine.
http://videogum.com/archives/commercials/s-epatha-merkerson-will-terrif_045001.html
or http://tinyurl.com/7jcful
http://www.schneier.com/blog/archives/2007/09/using_fear_to_s.html
Los Alamos has lost 80 computers: no idea if they're stolen, or just
misplaced. Typical story -- not even worth commenting on -- but this
great comment explains a lot about what was wrong with their security
policy: "The letter, addressed to Department of Energy security
officials, contends that 'cyber security issues were not engaged in a
timely manner' because the computer losses were treated as a 'property
management issue.'" The real risk in computer losses is the data, not
the hardware. I thought everyone knew that.
http://www.google.com/hostednews/afp/article/ALeqM5jXipyrzU1GKO4KQ3f4hhKyLvJvTA
or http://tinyurl.com/d7oxy5
Difficult-to-pronounce things are judged to be more risky than
easy-to-pronounce things:
http://www.ncbi.nlm.nih.gov/pubmed/19170941
New paper: "WiFi networks and malware epidemiology," by Hao Hu, Steven
Myers, Vittoria Colizza, and Alessandro Vespignani. Honestly, I'm not
sure I understood most of the article. And I don't think that their
model is all that great. But I like to see these sorts of methods
applied to malware and infection rates.
http://www.pnas.org/content/early/2009/01/26/0811973106
http://arxiv.org/abs/0706.3146
HIPAA accountability in U.S. stimulus bill:
http://www.schneier.com/blog/archives/2009/02/hipaa_accountab.html
Terrorism common sense from MI6:
http://www.theregister.co.uk/2009/02/11/mi6_spy_rubbishes_terrorism_fear/
or http://tinyurl.com/cxfl8s
Here's an analysis of 30,000 passwords from phpbb.com.
http://www.darkreading.com/blog/archives/2009/02/phpbb_password.html
It's similar to my analysis of 34,000 MySpace passwords.
http://www.schneier.com/blog/archives/2006/12/realworld_passw.html
Seems like we still can't choose good passwords. Conficker.B exploits
this, trying about 200 common passwords to help spread itself.
http://www.sophos.com/blogs/gc/g/2009/01/16/passwords-conficker-worm/
Blog entry:
http://www.schneier.com/blog/archives/2009/02/another_passwor.html
Evidence of the effectiveness of the "broken windows" theory of crime
fighting:
http://www.boston.com/news/local/massachusetts/articles/2009/02/08/breakthrough_on_broken_windows/
or http://tinyurl.com/cslqo5
http://www.ncjrs.gov/App/publications/Abstract.aspx?id=246202
The NSA wants help eavesdropping on Skype:
http://www.theregister.co.uk/2009/02/12/nsa_offers_billions_for_skype_pwnage/
or http://tinyurl.com/a9hn2n
I'm sure this is a real problem. Here's an article claiming that
Italian criminals are using Skype more than the telephone because of
eavesdropping concerns.
http://www.theregister.co.uk/2009/02/16/italian_crooks_skype/
A study from New Jersey shows that Megan's Law -- laws designed to
identity sex offenders to the communities they live in -- is ineffective
in reducing sex crimes or deterring recidivists.
http://www.nj.com/news/index.ssf/2009/02/study_finds_megans_law_fails_t_1.html
or http://tinyurl.com/b2mql2
Another Conficker variant: Conficker B++. This is one well-designed
piece of malware.
http://www.schneier.com/blog/archives/2009/02/new_conficker_v.html
President Obama has tasked Melissa Hathaway with conducting a 60-day
review of the nation's cybersecurity policies.
http://www.usatoday.com/tech/2009-02-16-cybersecurity-expert-obama_N.htm
or http://tinyurl.com/cx3kon
http://www.computerworld.com/action/article.do?command=viewArticleBasic&articleId=9127682&intsrc=news_ts_head
or http://tinyurl.com/d2ygpp
This interview, conducted last year, will give you a good idea of how
she thinks.
http://www2.computer.org/portal/web/computingnow/1208/whatsnew/securityandprivacy
or http://tinyurl.com/by28l7
Maine man tries to build a dirty bomb and no one cares, probably because
he isn't Muslim. White supremacist terrorism just isn't sexy these days.
http://jonathanstray.com/maine-man-tries-to-build-dirty-bomb
There are rumors of prototype electromagnetic pulse grenades:
http://www.theregister.co.uk/2009/02/12/electropulse_grenades/
TrapCall is a new service that reveals the caller ID on anonymous or
blocked calls.
http://blog.wired.com/27bstroke6/2009/02/trapcall.html
Judge orders defendant to decrypt laptop: interesting Fifth Amendment case.
http://news.cnet.com/8301-13578_3-10172866-38.html
Use this shower mirror with a hidden camera to catch the lovers of
cheating spouses:
http://www.dpl-surveillance-equipment.com/100611.html
The site has a wide variety of hidden cameras in common household objects.
http://www.dpl-surveillance-equipment.com/wireless_hidden_cameras.html
University of Miami law professor Michael Froomkin writes about ID cards
and society in "Identity Cards and Identity Romanticism."
http://papers.ssrn.com/sol3/papers.cfm?abstract_id=1309222
http://www.schneier.com/blog/archives/2009/03/michael_froomki.html
This commentary on the UK government national security strategy is
scary: "Sir David Omand, the former Whitehall security and intelligence
co-ordinator, sets out a blueprint for the way the state will mine data
-- including travel information, phone records and emails -- held by
public and private bodies and admits: 'Finding out other people's
secrets is going to involve breaking everyday moral rules.'" In short:
it's immoral, but we're going to do it anyway.
http://www.guardian.co.uk/uk/2009/feb/25/personal-data-terrorism-surveillance
or http://tinyurl.com/c5ll6r
Programs "staple" and "unstaple" perform all-or-nothing encryption.
Just demonstration code, but interesting all the same.
http://sysnet.ucsd.edu/projects/staple/
Interesting paper: "Optimised to Fail: Card Readers for Online Banking,"
by Saar Drimer, Steven J. Murdoch, and Ross Anderson.
http://www.cl.cam.ac.uk/~sjm217/papers/fc09optimised.pdf
http://www.lightbluetouchpaper.org/2009/02/26/optimised-to-fail-card-readers-for-online-banking/
or http://tinyurl.com/bdnafk
I'm sure you need some skill to actually use this self-defense pen, and
I'm also sure it'll get through airport security checkpoints just fine.
http://www.botachtactical.com/kzxtremepen.html
This article gives an overview of U.S. military robots, and discusses
some of the issues regarding the ethics of their use in war.
http://www.thenewatlantis.com/publications/military-robots-and-the-laws-of-war
or http://tinyurl.com/csoj98
The article was adapted from his book Wired for War: The Robotics
Revolution and Conflict in the 21st Century, published this year. I
bought the book, but I have not read it yet. Related is this paper on
the ethics of autonomous military robots.
http://www.schneier.com/blog/archives/2008/01/ethics_of_auton.html
Blog entry:
http://www.schneier.com/blog/archives/2009/03/history_and_eth.html
Secret NATO documents about the war in Afghanistan leaked due to bad
password:
https://secure.wikileaks.org/wiki/N1
Security theater scare mongering, in hotels and churches:
http://news.bbc.co.uk/1/hi/england/london/7933004.stm
http://www.cnn.com/2009/CRIME/03/09/church.security/index.html
http://www.schneier.com/blog/archives/2009/03/security_theate_2.html
Fascinating history of the techniques used to distribute child porn
throughout the world:
http://wikileaks.org/wiki/My_life_in_child_porn
http://www.schneier.com/blog/archives/2009/03/the_techniques.html#c356628
or http://tinyurl.com/asnc63
Google Maps spam:
http://www.schneier.com/blog/archives/2009/03/google_map_spam.html
This story of the world's largest diamond heist reads like a movie plot:
http://www.wired.com/politics/law/magazine/17-04/ff_diamonds?currentPage=all
or http://tinyurl.com/ak8hrx
Many Sentex keypads, which are used to secure doors everywhere, can be
opened with a default admin password:
http://www.schneier.com/blog/archives/2009/03/the_doghouse_se_1.html
** *** ***** ******* *********** *************
Insiders
Rajendrasinh Makwana was a UNIX contractor for Fannie Mae. On October
24, he was fired. Before he left, he slipped a logic bomb into the
organization's network. The bomb would have "detonated" on January 31.
It was programmed to disable access to the server on which it was
running, block any network monitoring software, systematically and
irretrievably erase everything -- and then replicate itself on all 4,000
Fannie Mae servers. Court papers claim the damage would have been in
the millions of dollars, a number that seems low. Fannie Mae would have
been shut down for at least a week.
Luckily -- and it does seem it was pure luck -- another programmer
discovered the script a week later, and disabled it.
Insiders are a perennial problem. They have access, and they're known
by the system. They know how the system and its security works, and its
weak points. They have opportunity. Bank heists, casino thefts,
large-scale corporate fraud, train robberies: many of the most
impressive criminal attacks involve insiders. And, like Makwana's
attempt at revenge, these insiders can have pretty intense motives --
motives that can only intensify as the economy continues to suffer and
layoffs increase.
Insiders are especially pernicious attackers because they're trusted.
They have access because they're *supposed* to have access. They have
opportunity, and an understanding of the system, because they use it --
or they designed, built, or installed it. They're already inside the
security system, making them much harder to defend against.
It's not possible to design a system without trusted people. They're
everywhere. In offices, employees are trusted people given access to
facilities and resources, and allowed to act -- sometimes broadly,
sometimes narrowly -- in the company's name. In stores, employees are
allowed access to the back room and the cash register; and customers are
trusted to walk into the store and touch the merchandise. IRS employees
are trusted with personal tax information; hospital employees are
trusted with personal health information. Banks, airports, and prisons
couldn't operate without trusted people.
Replacing trusted people with computers doesn't make the problem go
away; it just moves it around and makes it even more complex. The
computer, software, and network designers, implementers, coders,
installers, maintainers, etc. are all trusted people. See any analysis
of the security of electronic voting machines, or some of the frauds
perpetrated against computerized gambling machines, for some graphic
examples of the risks inherent in replacing people with computers.
Of course, this problem is much, much older than computers. And the
solutions haven't changed much throughout history, either. There are
five basic techniques to deal with trusted people:
1. Limit the number of trusted people. This one is obvious. The fewer
people who have root access to the computer system, know the combination
to the safe, or have the authority to sign checks, the more secure the
system is.
2. Ensure that trusted people are also trustworthy. This is the idea
behind background checks, lie detector tests, personality profiling,
prohibiting convicted felons from getting certain jobs, limiting other
jobs to citizens, the TSA's no-fly list, and so on, as well as behind
bonding employees, which means there are deep pockets standing behind
them if they turn out not to be trustworthy.
3. Limit the amount of trust each person has. This is
compartmentalization; the idea here is to limit the amount of damage a
person can do if he ends up not being trustworthy. This is the concept
behind giving people keys that only unlock their office or passwords
that only unlock their account, as well as "need to know" and other
levels of security clearance.
4. Give people overlapping spheres of trust. This is what security
professionals call defense in depth. It's why it takes two people with
two separate keys to launch nuclear missiles, and two signatures on
corporate checks over a certain value. It's the idea behind bank
tellers requiring management overrides for high-value transactions,
double-entry bookkeeping, and all those guards and cameras at casinos.
It's why, when you go to a movie theater, one person sells you a ticket
and another person standing a few yards away tears it in half: It makes
it much harder for one employee to defraud the system. It's why key
bank employees need to take their two-week vacations all at once -- so
their replacements have a chance to uncover any fraud.
5. Detect breaches of trust after the fact and prosecute the guilty.
In the end, the four previous techniques can only do so well. Trusted
people can subvert a system. Most of the time, we discover the security
breach after the fact and then punish the perpetrator through the legal
system: publicly, so as to provide a deterrence effect and increase the
overall level of security in society. This is why audit is so vital.
These security techniques don't only protect against fraud or sabotage;
they protect against the more common problem: mistakes. Trusted people
aren't perfect; they can inadvertently cause damage. They can make a
mistake, or they can be tricked into making a mistake through social
engineering.
Good security systems use multiple measures, all working together.
Fannie Mae certainly limits the number of people who have the ability to
slip malicious scripts into their computer systems, and certainly limits
the access that most of these people have. It probably has a hiring
process that makes it less likely that malicious people come to work at
Fannie Mae. It obviously doesn't have an audit process by which a
change one person makes on the servers is checked by someone else; I'm
sure that would be prohibitively expensive. Certainly the company's IT
department should have terminated Makwana's network access as soon as he
was fired, and not at the end of the day.
In the end, systems will always have trusted people who can subvert
them. It's important to keep in mind that incidents like this don't
happen very often; that most people are honest and honorable. Security
is very much designed to protect against the dishonest minority. And
often little things -- like disabling access immediately upon
termination -- can go a long way.
This essay originally appeared on the Wall Street Journal website.
http://online.wsj.com/article/SB123447990459779609.html
Makwana:
http://blogs.zdnet.com/BTL/?p=11905
http://www.theregister.co.uk/2009/01/29/fannie_mae_sabotage_averted/
http://blog.wired.com/27bstroke6/2009/01/fannie.html
Economic downturn increases insider threat:
http://news.bbc.co.uk/1/hi/technology/7875904.stm
Hospital employees illegally accessing patient data:
http://www.schneier.com/blog/archives/2007/10/27_suspended_fo.html
Insecurity in electronic voting machines:
http://www.schneier.com/blog/archives/2006/11/voting_technolo.html
http://www.nytimes.com/2008/01/06/magazine/06Vote-t.html
http://www.schneier.com/essay-101.html
http://freedom-to-tinker.com/blog/dwallach/vendor-misinformation-e-voting-world
or http://tinyurl.com/5c7kxn
http://www.schneier.com/blog/archives/2008/08/diebold_finally.html
http://blog.wired.com/27bstroke6/2009/01/diebold-audit-l.html
http://www.schneier.com/essay-068.html
http://www.crypto.com/blog/ohio_voting/
http://www.huffingtonpost.com/kirsten-anderson/an-interview-with-david-w_b_64063.html
or http://tinyurl.com/ad6rn3
Computerized gambling machine fraud:
http://www.reviewjournal.com/lvrj_home/1998/Jan-10-Sat-1998/news/6745681.html
or http://tinyurl.com/xswg
Replacing people with computers:
http://www.schneier.com/blog/archives/2008/12/comparing_the_s.html
Audit:
http://www.schneier.com/blog/archives/2008/12/audit.html
** *** ***** ******* *********** *************
The Doghouse: Singularics
This is priceless:
"Our advances in Prime Number Theory have led to a new branch of
mathematics called Neutronics. Neutronic functions make possible for the
first time the ability to analyze regions of mathematics commonly
thought to be undefined, such as the point where one is divided by zero.
In short, we have developed a new way to analyze the undefined point at
the singularity which appears throughout higher mathematics.
"This new analytic technique has given us profound insight into the way
that prime numbers are distributed throughout the integers. According to
RSA's website, there are over 1 billion licensed instances of RSA
public-key encryption in use in the world today. Each of these instances
of the prime number based RSA algorithm can now be deciphered using
Neutronic analysis. Unlike RSA, Neutronic Encryption is not based on two
large prime numbers but rather on the Neutronic forces that govern the
distribution of the primes themselves. The encryption that results from
Singularic's Neutronic public-key algorithm is theoretically impossible
to break."
You'd think that anyone who claims to be able to decrypt RSA at the key
lengths in use today would, maybe, um, demonstrate that at least once.
Otherwise, this can all be safely ignored as snake oil.
The founder and CTO also claims to have proved the Riemann Hypothesis,
if you care to wade through the 63-page paper.
http://www.singularics.com/products/encryption/
Snake oil:
http://www.schneier.com/crypto-gram-9902.html#snakeoil
Riemann Hypothesis "proof":
http://www.singularics.com/science/mathematics/OnNeutronicFunctions.pdf
or http://tinyurl.com/agmoy9
** *** ***** ******* *********** *************
Three Security Anecdotes from the Insect World
Beet armyworm caterpillars react to the sound of a passing wasp by
freezing in place, or even dropping off the plant. Unfortunately,
armyworm intelligence isn't good enough to tell the difference between
enemy aircraft (the wasps that prey on them) and harmless commercial
flights (bees); they react the same way to either. So by producing
nectar for bees, plants not only get pollinated, but also gain some
protection against being eaten by caterpillars.
The small hive beetle lives by entering beehives to steal combs and
honey. They home in on the hives by detecting the bees' own alarm
pheromones. They also track in yeast that ferments the pollen and
releases chemicals that spoof the alarm pheromones, attracting more
beetles and more yeast. Eventually the bees abandon the hive, leaving
the beetles and yeast to finish off the pollen and honey.
Mountain alcon blue caterpillars get ants to feed them by spoofing a
biometric: the sounds made by the queen ant.
http://scienceblogs.com/notrocketscience/2008/12/buzzing_bees_scare_caterpillars_away_from_plants.php
or http://tinyurl.com/b2fp7m
http://scienceblogs.com/notrocketscience/2009/01/beetle_and_yeast_team_up_against_bees.php
or http://tinyurl.com/96kdea
http://scienceblogs.com/notrocketscience/2009/02/butterflies_scrounge_off_ants_by_mimicking_the_music_of_quee.php
or http://tinyurl.com/cxu8cm
** *** ***** ******* *********** *************
The Kindness of Strangers
When I was growing up, children were commonly taught: "don't talk to
strangers." Strangers might be bad, we were told, so it's prudent to
steer clear of them.
And yet most people are honest, kind, and generous, especially when
someone asks them for help. If a small child is in trouble, the
smartest thing he can do is find a nice-looking stranger and talk to him.
These two pieces of advice may seem to contradict each other, but they
don't. The difference is that in the second instance, the child is
choosing which stranger to talk to. Given that the overwhelming majority
of people will help, the child is likely to get help if he chooses a
random stranger. But if a stranger comes up to a child and talks to him
or her, it's not a random choice. It's more likely, although still
unlikely, that the stranger is up to no good.
As a species, we tend help each other, and a surprising amount of our
security and safety comes from the kindness of strangers. During
disasters: floods, earthquakes, hurricanes, bridge collapses. In times
of personal tragedy. And even in normal times.
If you're sitting in a café working on your laptop and need to get up
for a minute, ask the person sitting next to you to watch your stuff.
He's very unlikely to steal anything. Or, if you're nervous about that,
ask the three people sitting around you. Those three people don't know
each other, and will not only watch your stuff, but they'll also watch
each other to make sure no one steals anything.
Again, this works because you're selecting the people. If three people
walk up to you in the cafe and offer to watch your computer while you go
to the bathroom, don't take them up on that offer. Your odds of getting
three honest people are much lower.
Some computer systems rely on the kindness of strangers, too. The
Internet works because nodes benevolently forward packets to each other
without any recompense from either the sender or receiver of those
packets. Wikipedia works because strangers are willing to write for, and
edit, an encyclopedia with no recompense.
Collaborative spam filtering is another example. Basically, once someone
notices a particular e-mail is spam, he marks it, and everyone else in
the network is alerted that it's spam. Marking the e-mail is a
completely altruistic task; the person doing it gets no benefit from the
action. But he receives benefit from everyone else doing it for other
e-mails.
Tor is a system for anonymous Web browsing. The details are complicated,
but basically, a network of Tor servers passes Web traffic among each
other in such a way as to anonymize where it came from. Think of it as a
giant shell game. As a Web surfer, I put my Web query inside a shell and
send it to a random Tor server. That server knows who I am but not what
I am doing. It passes that shell to another Tor server, which passes it
to a third. That third server -- which knows what I am doing but not who
I am -- processes the Web query. When the Web page comes back to that
third server, the process reverses itself and I get my Web page.
Assuming enough Web surfers are sending enough shells through the
system, even someone eavesdropping on the entire network can't figure
out what I'm doing.
It's a very clever system, and it protects a lot of people, including
journalists, human rights activists, whistleblowers, and ordinary people
living in repressive regimes around the world. But it only works because
of the kindness of strangers. No one gets any benefit from being a Tor
server; it uses up bandwidth to forward other people's packets around.
It's more efficient to be a Tor client and use the forwarding
capabilities of others. But if there are no Tor servers, then there's no
Tor. Tor works because people are willing to set themselves up as
servers, at no benefit to them.
Alibi clubs work along similar lines. You can find them on the Internet,
and they're loose collections of people willing to help each other out
with alibis. Sign up, and you're in. You can ask someone to pretend to
be your doctor and call your boss. Or someone to pretend to be your boss
and call your spouse. Or maybe someone to pretend to be your spouse and
call your boss. Whatever you want, just ask and some anonymous stranger
will come to your rescue. And because your accomplice is an anonymous
stranger, it's safer than asking a friend to participate in your ruse.
There are risks in these sorts of systems. Regularly, marketers and
other people with agendas try to manipulate Wikipedia entries to suit
their interests. Intelligence agencies can, and almost certainly have,
set themselves up as Tor servers to better eavesdrop on traffic. And a
do-gooder could join an alibi club just to expose other members. But for
the most part, strangers are willing to help each other, and systems
that harvest this kindness work very well on the Internet.
This essay originally appeared on the Wall Street Journal website.
http://online.wsj.com/article/SB123567809587886053.html
Tor:
http://www.torproject.org/torusers.html.en
http://www.torproject.org
Alibi clubs:
http://www.nytimes.com/2004/06/26/technology/26ALIB.html?hp
http://www.alibinetwork.com/index.jsp
** *** ***** ******* *********** *************
New eBay Fraud
Here's a clever fraud, exploiting relative delays in eBay, PayPal, and
UPS shipping.
"The buyer reported the item as 'destroyed' and demanded and got a
refund from Paypal. When the buyer shipped it back to Chad and he opened
it, he found there was nothing wrong with it -- except that the scammer
had removed the memory, processor and hard drive. Now Chad is out $500
and left with a shell of a computer, and since the item was 'received'
Paypal won't do anything."
Very clever. The seller accepted the return from UPS after a visual
inspection, so UPS considered the matter closed. PayPal and eBay both
considered the matter closed. if the amount was large enough, the
seller could sue, but how could he prove that the computer was
functional when he sold it?
It seems to me that the only way to solve this is for PayPal to not
process refunds until the seller confirms what he received back is the
same as what he shipped. Yes, then the seller could commit similar
fraud, but sellers (certainly professional ones) have a greater
reputational risk.
http://consumerist.com/5159479/ebay-scammer-says-pc-destroyed-in-mail-takes-500-sends-back-destroyed-pc-minus-parts
or http://tinyurl.com/czj2bu
** *** ***** ******* *********** *************
Schneier News
Schneier is speaking at MinneWebCon on April 6 in Minneapolis.
http://minnewebcon.umn.edu/
Schneier is speaking at the 3rd Annual Asia-Pacific Programme for Senior
National Security Officers (APPSNO) on April 14 in Singapore.
http://www.rsis.edu.sg/cens/events/upcoming_events.html
** *** ***** ******* *********** *************
IT Security: Blaming the Victim
Blaming the victim is common in IT: users are to blame because they
don't patch their systems, choose lousy passwords, fall for phishing
attacks, and so on. But, while users are, and will continue to be, a
major source of security problems, focusing on them is an unhelpful way
to think.
People regularly don't do things they are supposed to: changing the oil
in their cars, going to the dentist, replacing the batteries in their
smoke detectors. Why? Because people learn from experience. If something
is immediately harmful, e.g., touching a hot stove or petting a live
tiger, they quickly learn not to do it. But if someone skips an oil
change, ignores a computer patch, or chooses a lousy password, it's
unlikely to matter. No feedback, no learning.
We've tried to solve this in several ways. We give people rules of
thumb: oil change every 5,000 miles; secure password guidelines. Or we
send notifications: smoke alarms beep at us, dentists send postcards,
Google warns us if we are about to visit a website suspected of hosting
malware. But, again, the effects of ignoring these aren't generally felt
immediately.
This makes security primarily a hindrance to the user. It's a recurring
obstacle: something that interferes with the seamless performance of the
user's task. And it's human nature, wired into our reasoning skills, to
remove recurring obstacles. So, if the consequences of bypassing
security aren't obvious, then people will naturally do it.
This is the problem with Microsoft's User Account Control (UAC).
Introduced in Vista, the idea is to improve security by limiting the
privileges applications have when they're running. But the security
prompts pop up too frequently, and there's rarely any ill-effect from
ignoring them. So people do ignore them.
This doesn't mean user education is worthless. On the contrary, user
education is an important part of any corporate security program. And at
home, the more users understand security threats and hacker tactics, the
more secure their systems are likely to be. But we should also recognise
the limitations of education.
The solution is to better design security systems that assume uneducated
users: to prevent them from changing security settings that would leave
them exposed to undue risk, or even better to take security out of their
hands entirely.
For example, we all know that backups are a good thing. But if you
forget to do a backup this week, nothing terrible happens. In fact,
nothing terrible happens for years on end when you forget. So, despite
what you know, you start believing that backups aren't really that
important. Apple got the solution right with its backup utility Time
Machine. Install it, plug in an external hard drive, and you are
automatically backed up against hardware failure and human error. It's
easier to use it than not.
For its part, Microsoft has made great strides in securing its operating
system, providing default security settings in Windows XP and even more
in Windows Vista to ensure that, when a naive user plugs a computer in,
it's not defenceless.
Unfortunately, blaming the user can be good business. Mobile phone
companies save money if they can bill their customers when a calling
card number is stolen and used fraudulently. British banks save money by
blaming users when they are victims of chip-and-pin fraud. This is
continuing, with some banks going so far as to accuse the victim of
perpetrating the fraud, despite evidence of large-scale fraud by
organised crime syndicates.
The legal system needs to fix the business problems, but system
designers need to work on the technical problems. They must accept that
security systems that require the user to do the right thing are doomed
to fail. And then they must design resilient security nevertheless.
This essay originally appeared in The Guardian.
http://www.guardian.co.uk/technology/2009/mar/12/read-me-first
Users are a problem:
http://www.informationweek.com/news/security/client/showArticle.jhtml?articleID=213002007
or http://tinyurl.com/ab8pux
http://www.informationweek.com/news/security/attacks/showArticle.jhtml?articleID=212700890
or http://tinyurl.com/b2s2ep
Lousy passwords:
http://www.schneier.com/essay-144.html
Choosing good passwords:
http://www.schneier.com/essay-148.html
Microsoft's UAC problems:
http://arstechnica.com/security/news/2008/04/vistas-uac-security-prompt-was-designed-to-annoy-you.ars
or http://tinyurl.com/cxazee
The limits of education:
http://www.schneier.com/essay-139.html
Blaming the user:
http://www.schneier.com/blog/archives/2005/12/cell_phone_comp.html
http://news.bbc.co.uk/1/hi/programmes/newsnight/7265437.stm
Large-scale chip-and-pin fraud:
http://www.telegraph.co.uk/news/newstopics/politics/lawandorder/3173346/Chip-and-pin-scam-has-netted-millions-from-British-shoppers.html
or http://tinyurl.com/4xuk69
** *** ***** ******* *********** *************
Balancing Security and Usability in Authentication
Since January, the Conficker.B worm has been spreading like wildfire
across the Internet: infecting the French Navy, hospitals in Sheffield,
the court system in Houston, and millions of computers worldwide. One
of the ways it spreads is by cracking administrator passwords on
networks. Which leads to the important question: Why in the world are
IT administrators still using easy-to-guess passwords?
Computer authentication systems have two basic requirements. They need
to keep the bad guys from accessing your account, and they need to allow
you to access your account. Both are important, and every
authentication system is a balancing act between the two. Too little
security, and the bad guys will get in too easily. But if the
authentication system is too complicated, restrictive, or hard to use,
you won't be able to -- or won't bother to -- use it.
Passwords are the most common authentication system, and a good place to
start. They're very easy to implement and use, which is why they're so
popular. But as computers have become faster, password guessing has
become easier. Most people don't choose passwords that are complicated
enough to remain secure against modern password-guessing attacks.
Conficker.B is even less clever; it just tries a list of about 200
common passwords.
To combat password guessing, many systems force users to choose
harder-to-guess passwords -- requiring minimum lengths, non
alpha-numeric characters, etc. -- and change their passwords more
frequently. The first makes guessing harder, and the second makes a
guessed password less valuable. This, of course, makes the system more
annoying, so users respond by writing their passwords down and taping
them to their monitors, or simply forgetting them more often. Smarter
users write them down and put them in their wallets, or use a secure
password database like Password Safe.
Users forgetting their passwords can be expensive -- sysadmins or
customer service reps have to field phone calls and reset password -- so
some systems include a backup authentication system: a secret question.
The idea is that if you forget your password, you can authenticate
yourself with some personal information that only you know. Your
mother's maiden name was traditional, but these days there are all sorts
of secret questions: your favourite schoolteacher, favourite colour,
street you grew up on, name of your first pet, and so on. This might
make the system more usable, but it also makes it much less secure:
answers can be easily guessable, and are often known by people close to you.
A common enhancement is a one-time password generator, like a SecurID
token. This is a small device with a screen that displays a password
that changes automatically once a minute. Adding this is called
two-factor authentication, and is much more secure, because this token
-- "something you have" -- is combined with a password -- "something you
know." But it's less usable, because the tokens have to be purchased
and distributed to all users, and far too often it's "something you lost
or forgot." And it costs money. Tokens are far more frequently used in
corporate environments, but banks and some online gaming worlds have
taken to using them -- sometimes only as an option, because people don't
like them.
In most cases, how an authentication system works when a legitimate user
tries to log on is much more important than how it works when an
impostor tries to log on. No security system is perfect, and there is
some level of fraud associated with any of these authentication methods.
But the instances of fraud are rare compared to the number of times
someone tries to log on legitimately. If a given authentication system
let the bad guys in one in a hundred times, a bank could decide to live
with the problem -- or try to solve it in some other way. But if the
same authentication system prevented legitimate customers from logging
on even one in a thousand times, the number of complaints would be
enormous and the system wouldn't survive one week.
Balancing security and usability is hard, and many organizations get it
wrong. But it's also evolving; organizations needing to tighten their
security continue to push more involved authentication methods, and more
savvy Internet users are willing to accept them. And certainly IT
administrators need to be leading that evolutionary change.
A version of this essay was originally published in The Guardian.
http://www.guardian.co.uk/technology/2009/feb/19/insecure-passwords-conflickerb-worm
or http://tinyurl.com/awd5np
Conficker.B:
http://www.crn.com/security/212902319
http://www.telegraph.co.uk/news/worldnews/europe/france/4547649/French-fighter-planes-grounded-by-computer-virus.html
or http://tinyurl.com/bbku57
http://www.smarthealthcare.com/sheffield-conficker
http://www.theregister.co.uk/2009/02/09/houston_malware_infection/
http://arstechnica.com/security/news/2009/01/conficker-worm-spikes-infects-1-1-million-pcs-in-24-hours.ars
or http://tinyurl.com/dmvd8d
http://securitywatch.eweek.com/virus_and_spyware/experts_-_conficker_usb_worm_spreading_quickly.html
or http://tinyurl.com/bk5fs9
http://voices.washingtonpost.com/securityfix/2009/01/tricky_windows_worm_wallops_mi.html
or http://tinyurl.com/8e8fbg
http://bt.counterpane.com/Risk_Assessment_W32.Conficker_Worm_Update2.pdf
or http://tinyurl.com/detvm5
http://www.microsoft.com/security/portal/Entry.aspx?Name=Worm:Win32/Conficker.B
or http://tinyurl.com/9vpbxs
http://www.sophos.com/blogs/gc/g/2009/01/16/passwords-conficker-worm/
Guessing passwords:
http://www.schneier.com/essay-246.html
http://www.schneier.com/essay-148.html
Password Safe:
http://www.schneier.com/passsafe.html
Security problems with secret questions:
http://www.schneier.com/blog/archives/2005/02/the_curse_of_th.html
** *** ***** ******* *********** *************
Comments from Readers
There are hundreds of comments -- many of them interesting -- on these
topics on my blog. Search for the story you want to comment on, and join in.
http://www.schneier.com/blog
Sunday, 15 March 2009
Posted by Britannia Radio at 11:54