Since Thompson had created the original Unix compiler - which was used to compile every other compiler and thus every other flavor of Unix - this was a pretty wild thought experiment, especially since he didn't outright deny having done it.
7/
Top-level
Since Thompson had created the original Unix compiler - which was used to compile every other compiler and thus every other flavor of Unix - this was a pretty wild thought experiment, especially since he didn't outright deny having done it. 7/ 18 comments
Okay, so maybe you are sure the OS isn't compromised, but how does the OS know if it's even running on the "bare metal" of your computer. Maybe it is running inside a virtual machine, and the *actual* OS on the computer is a malicious program that sits between your OS and the chips and circuits, distorting the data it sends and receives. This is called a "rootkit," and it's a deadass nightmare that actually exists in the actual world. 9/ A computer with a rootkit is a brain in a jar, a human battery in the Matrix. You, the computer user, can ask the operating system questions about its operating environment that it will answer faithfully and truthfully, and *those answers will all be wrong*, because the *actual computer* is being controlled by the rootkit and it only tells your operating system what it wants it to know. 10/ 20 years ago, some clever Microsoft engineers proposed a solution to this conundrum: "Trusted Computing." They proposed adding a second computer to your system, a sealed, secure chip with very little microcode, so little that it could all be audited in detail and purged of bugs. 11/ My then-colleague Seth Schoen - EFF's staff technologist, the most technically sophisticated person to have been briefed on the technology without signing an NDA - made several pointed critiques of Palladium: https://web.archive.org/web/20020802145913/http://vitanuova.loyalty.org/2002-07-05.html And suggested a hypothetical way to make sure it only served computer users, and not corporations or governments who wanted to control them: https://www.linuxjournal.com/article/7055 13/ But his most salient concern was this: "what if malware gets into the trusted computing chip?" The point of trusted computing was to create a nub of certainty, a benevolent God whose answers to your questions could always be trusted. The output from a trusted computing element would be ground truth, axiomatic, trusted without question. By having a reliable external observer of your computer and its processes, you could always tell whether you were in the Matrix or in the world. 14/ It was a red pill for your computer. What if it was turned? What if some villain convinced it to switch sides, by subverting its code, or by subtly altering it at the manufacturer? That is, what if Descartes' God was a sadist who *wanted* to torment him? This was a nightmare scenario in 2002, one that the trusted computing advocates never adequately grappled with. In the years since, it's only grown more salient, as trusted computing variations have spread to many kinds of computer. 15/ The most common version is the UEFI - ("Unified Extensible Firmware Interface") - a separate operating system, often running on its own chip (though sometimes running in a notionally "secure" region of your computer's main processors) that is charged with observing and securing your computer's boot process. 16/ UEFI poses lots of dangers to users; it can (and is) used by manufacturers to block third-party operating systems, which allows them to lock you into using their own products, including their app stores, letting them restrict your choices and pick your pocket. 17/ But in exchange, UEFI is said to deliver a far more important benefit: a provably benevolent God, one who will never lie to your operating system about whether it is in the Matrix or in the real world, providing the foundational ground truth needed to find and block malicious software. 18/ But despite its long tenure, Cosmicstrand was only just discovered. That's because of the fundamental flaw inherent in designing a computer that its owners can't fully inspect or alter: if you design a component that is supposed to be immune from owner override, then anyone who compromises that component *can't be detected or countered by the computer's owner*. 21/ This is the core of a two-decade-old debate among security people, and it's one that the "benevolent God" faction has consistently had the upper hand in. They're the "curated computing" advocates who insist that preventing you from choosing an alternative app store or side-loading a program is for your own good - because if it's possible for you to override the manufacturer's wishes, then malicious software may impersonate you to do so, or you might be tricked into doing so. 22/ This benevolent dictatorship model only works so long as the dictator is both perfectly benevolent and perfectly competent. We know the dictators aren't always benevolent. Apple won't invade your privacy to sell you things, but they'll take away ever Chinese user's privacy to retain their ability to manufacture devices in China: https://www.nytimes.com/2021/05/17/technology/apple-china-censorship-data.html 23/ @pluralistic do you know the Mitchell and Webb sketch about the computer Colosson whose improbable emergency shutdown trigger is seeing a human holding up a photo of a duck? |
Trusting trust is still the most important issue in information security. Sure, you can run a virus-checker, but that virus checker has to ask your operating system to tell it about what files are on the drive, what data is in memory, and what processes are being executed. What if the OS is compromised?
8/