What’s the difference between these two laptops?
Besides 7 years and a few generations of processor technology, surprisingly little. The Dell was born in 2004, the Mac was born in 2011. There’s USB 2.0, Gigabit Ethernet, and 802.11abg on the Dell thanks to eBayed expansion cards, so the connectivity options are roughly equivalent. There’s better power management for sure on the latest Mac models, more/multiple cores, multithreading, more Instructions Per Cycle (IPC) at the chip level, and built-in video decoding, among various other technical improvements.
But does the laptop on the right actually help me get work done faster than the laptop on the left?
And therein lies a major source of Intel’s and Microsoft’s year-on-year growth problems. Their chips, and in fact, the whole desktop PC revolution, have produced machines that far outstrip anyone’s average needs. The acute cases, playing games, folding proteins, making and breaking cryptography, and so on, will always benefit from improved speed. But the average web surfer / Google Docs editor / web designer / blogger, probably could make due with a machine from 2004.
In fact, one might be willing to say that the old computer is actually better for getting work done, because it consumes such a huge fraction of CPU performance when playing YouTube videos. There’s no room for the faux, 10-way “efficient multitasking” people believe they are capable of doing these days.
Which means I can essentially single-task or dual-task (write some code, refresh the browser), and keep my attention focused on fewer tasks. I’m constrained in a useful way by the hardware, and I can use those constraints to my advantage.
The software stack remains the same (one of the great achievements of the Windows era): Google Chrome, PuTTY, iTunes, maybe a Firefox window open in there. All of this runs just fine on a Pentium 4-M at 2.0Ghz and 1GB of RAM (probably about the performance of the Nexus 7, a 4 + 1-core ARM processor w/1GB of RAM), and the CPU is still over 90% idle. When Intel announces that it’s building chips with ten times the performance, that just means the processor is ten times more idle on average, big whoop.
What I’m saying is: if all the average user needs is the compute-equivalent of a laptop from 2004, then that’s effectively what a current smartphone or tablet is.
But then what I’m also asking is: Why can’t I do as much work on a smartphone from 2013, as I can on a laptop from 2004?
I believe that there are a number of huge shortcomings in the phone and tablet world, which still cause people to unnecessarily purchase laptops or desktops:
- The output options are terrible. While it’s nice to have a Retina display with what is essentially mid-1990s, laser-printer dots-per-inch resolution, you can still only fit so much text into a small space before people can’t read it anyway. It doesn’t matter how sharp it is, the screen is just too damn small. What would be nicer for anyone buying these expensive, output-only devices, would be the ability to hook any tablet up to a 24″ HD monitor and then throw down some serious work or watch some serious Netflix.
- Why can’t I, for the love of God, plug a USB memory stick into one of these things? It’s just a computer, computers have USB ports, ergo, I ought to be able to plug a memory stick into a tablet. Whose artificial constraints and massively egocentric desire to organize all of the world’s information is keeping me from using a tablet to quietly organize mine in the comfort and literal walled-garden of my own home?
- There’s no Microsoft Excel on a tablet, and Google Sheets is obviously not a useful replacement. But that is a moot point, because…
- …the input options for tablets are terrible, there are no sensible input options for anyone wanting to do more than consume information. Content creation, especially when it comes to inputting data, is impossible. Keyboards-integrated-into-rubbery-crappy-covers will not help here. So tablets are still only useful for 10-foot view user interfaces, and things that don’t require much precision. Sure, Autodesk and other big-league CAD companies have built glorified 3D model viewers for tablets, but note how many of the app descriptions include the word “view” but not the words “edit” or “create”. No serious designer would work without keyboard shortcuts and a mouse or Wacom tablet.
- The walled-garden approach to user-interface design and data management in the cloud is terrible. There are multiple idioms associated with saving and loading data, multiple processes, there is no uniformity at the operating-system level. By killing off the single Open dialog coupled to hierarchical storage and user-defined storage schemas in the form of endlessly buried directories, it’s very hard for a user to get a sense of the scale and context of their data. People have their own ways of navigating and keying in on their data, but the one-size-fits-all cloud model rubs abrasively right up against that. It seems to me that eliminating the Open dialog was more an act of dogma than of use-testing, and that it’s almost heresy (and in fact, essentially coded into OS X (via App Sandboxing), iOS, and Android’s DNA now) to wish for the unencumbered dialog to return. Cordoning data off into multiple silos is akin to saving data on floppies or that old stack of burned CDs you might have laying about somewhere: it’s not really fun to have to search through them all to find what you were looking for.
Unfortunately, by refusing to consider the above points, no one’s currently thinking about using these phones and tablets for what they really are: full-fledged desktop/laptop replacements, particularly when combined with things like Nvidia’s GRID technologies for virtualized 3D scene rendering.
The market for computing devices is now very strange indeed. Phablet manufacturers sell input/output-hobbled devices for more than full-fledged machines and even command huge price premiums for them. Meanwhile, the old desktop machine in the corner is still more than powerful enough to survive another processor generation, but it’s the old hotness and no one really wants to use it. And no one is making real moves to make the smartphone the center of computing in the home, even though that is where most of the useful performance/Watt research and development is going.
So here’s a simple request for the kind of freedom to use things in a way the manufacturers never anticipated: I’d like to have text editing on a big monitor work flawlessly out-of-the-box with my next smartphone, using HDMI output and a Bluetooth keyboard. (Why aren’t we even there yet? It took several versions of Android and several ROM flashes before I could even get it working, and even now the Control keys aren’t quite working.)
Add in Chrome / Firefox on the smartphone and you have enough of a setup for a competent web developer to do some damage, and perform most of the write, refresh, test development cycle. And be able to do this anywhere via WiFi or HSDPA/LTE connections (a fallback data connection apparently being necessary if you ever want to do work at this place, snarkiness aside).
There’s so much more usability that could be enabled if the input, output, and storage subsystems of the major smartphone OSes were ever made to be unconstrained.
Such a shift to smartphone primacy is almost presently possible and I can only wonder whether in a few years’ time (2 – 3 phone generations) this scenario will become the norm, but I’m not holding my breath. The companies involved in the current round of empire building (including some of the companies involved in the last) are seemingly too concerned with controlling every part of the experience to allow something better to reach us all.