I had not fully subscribed to this theory, but it seemed more likely that I was romanticizing something that was actually Not Great that it was that nearly every modern software and hardware platform had inexplicably gotten shittier.
I am now prepaired to say, with conviction, that every modern hardware and software platform has gotten shittier, and that it's not inexplicable. I'm going to try to explain how I came to this conclusion, and give some potential explainations.
First, let me lay out a little bit about my technical literacy and my profession, this might help explain some of the assertions that I'm going to make. I started using computers, and the internet, in 1995. Our family computer, my first computer, ran Windows 3.11 (for workgroups). Later, in the late 90s, I was given an Atari 400 and reams of books and magazines on basic, followed shortly by an Apple II GS and dozens of disks of software.
I loved computers. I loved making weird little programs, and silly applications and games. I'd build things in GW Basic or Freebasic, and distribute it to my friends on floppy disks. Even in the latter half of the 00s, I was passing half broken games around on floppy disks (or collections on CD-Rs, when I could talk someone in to buying some for me.) Computers were, by and large, ubiquitous in my life. Nearly everyone had an old one they didn't want, and a new one they didn't understand.
I collected cast offs from neighbors, from thrift stores, from office upgrades. I rebuilt them, installed useful or fun software on them, and sold them or gave them away. All of my friends had computers of their own, because I had access to these machines, and I cared enough to outfit them with the appropriate software.
(It must be said, at this point, that 'useful' and 'appropriate' are relative terms. In 2009 I gave a good friend a computer that had been built for Windows 98. It was running Puppy Linux from a CD, and saving data to a USB flash drive over USB 1.1. It did word processing, email, and basic web browsing. It had a whopping 64MB of RAM, and was covered in glitter, googley eyes, and carpet samples. But it was free, and it wasn't useless, and that was important.)
Some of these shortcomings are legitimate bugs. Some of them are bafflingly short sighted or poorly considered architectural decisions. Just as many are cases of a divergence between the needs of the user and the abilities of a program. Modern programs are often feature incomplete, poorly supported, and difficult or impossible to customize. Modern computers are often slow, and cranky. I'm responsible for handling the fallout of this unfortunate situation.
I've seen how revolutionary a computer can be, if it is designed with the needs of the user in mind, and how disastrous the same can be when it is not. I've seen computers used to empower people, and used to oppress. I've seen computers be Good, and the consequences of when they are not.
So that's who I am, and my experience with computers so far. Those are my credentials, and my qualifications.
The Computer Chronicles was a TV show that ran from the early 80s through the early 00s. Over it's nearly 20 year run, The Computer Chronicles covered nearly every facet of the newly developing Computer industry. It was hosted by people with Opinions.
The guests were, frequently, people who were proud of the things they made, or the software they represented.
Watching the developer of CP/M and DR DOS talk to a mainframe engineer who worked at IBM in the 50s about the future of computers as seen from the 1980s was eye opening.
On the one hand, this show serves as an excellent introduction to, or reminder of, the capabilities of computers 35 years ago. It helps us see how far we've come in terms of miniaturization, while also demonstrating again that, in many ways, there is nothing new under the sun.
Before the advent of the internet, reporters were writing their stories on laptops and sending them in over phone lines, 25 years before the release of the iphone HP released a computer with a touchscreen, three years before microsoft released he first version of windows Apple and Visicorp demontrated GUIs wih features that Windows wouldn't be able to approach for another 9+ years.
And, of course, I'm reminded again of Douglas Engelbart's 1968 "Mother of all Demos", in which he demonstrated the mouse, the GUI, instant messaging, networked gaming, and basically every other important development of the following 50 years.
It took 5 years for Xerox to refine and miniturize Engelbart's ideas to the point that they thought they could market them, and another 10 years before Apple refined and further miniturizaed the same ideas, and brought us the Mac.
Nothing is ever new.
There were others working around the same time on similar ideas, or at least from a similar philosophy. Working to make computers, if not intuitive, at least comprehensible. I think this is a noble goal.
The computer is often thought of as a tool, but it is more like a tool shed, in which we store a collection of tools, a source of power, and a workspace.
That is to say, in the 60s and 70s, computers were weak and slow and computer users were also computer programmers. A small, tight knit circle of developers and computer scientists were responsible for the bulk of the progress made in that time, and the idea of designing tools for non-technical users was never considered.
Computers became more affordable, slowly. Affordable computers became more powerful, quickly. Within 10 years, non-technical users were interacting with computers on a daily basis. It was against the beginnings of this backdrop that the phrase I mentioned earlier was coined. "Human Literate Computers" or "Human Centered Computing."
Ease of Use was the holy grail for a lot of computer companies. A computer that was so easy to use that they could sell it to grandma. But, to me at least, Human Literate and Easy to Use are distinct ideas. Many modern applications are Easy to Use. Netflix is Easy to Use. Facebook is, for all it's faults, pretty easy to use. The iPhone, the iPad, and ChromeOS are super easy to use.
Well, they are easy to use as long as you use them in the prescribed way. As long as you let them tell you what you want to do, instead of the other way around.
That, IMO, is the distinction.
I think that many of the steps towards demystifying the computer of the 80s and 90s did good work, but ultimately, the computer industry left the whole idea behind, in favor of making some tasks Very Easy while making other tasks Practically Impossible, and turning everything into a surveillance device.
When I was a kid I was brought up with computers that showed you how they worked.
You booted in to a command prompt or a programming language, or you could get to one, if you wanted to.
I got to play with GW Basic and qBasic and, a little, with hypercard.
I got to take apart software and put it back together and make things that made people happy.
I often wonder why Hypercard had to die.
It was because Jobs wanted the Computer to be an Appliance. A thing only used in prescribed ways.
Letting people build their own tools means letting people control their own destiny.
If I can make what I want, or if someone else can make what they want, and then I can take it apart and improve it, why would I pay for an upgrade? Why would I pay you to build something that doesn't meet my needs?
Hypercard, if your unfamiliar, is powerpoint + instructions.
Here's a great introduction/example: http://www.loper-os.org/?p=568
The author walks you through building a calculator app in about 5 minutes, step by step.
Warning: There's a bit of ableist language tossed around in the last paragraph. Skip it, there's nothing worth reading there anyway.
You use the same kinds of tools you would use to build a slideshow, but you couple them with links, multimedia, and scripting.
Want a visual interface for your database of client data? Great! slap together a roladex card, and drop in a search function.
Go from concept to presentation ready in an hour or two (or less, if you've done this before!)
My nephew has an ipad.
He asked his dad how to write games. His dad didn't know. His dad asked me how to write games on an iPad. I told him not to bother.
My nephew asked me how to learn to write games.
I gave him a raspberry pi and a copy of pico 8.
Now he writes computer games.
He couldn't do that on his iPad.
@ajroach42 Raspberry Pi == the new C64/Beeb.
Lots of IT careers will be started with the Pi, just as they were with the C64/Beeb.
Python, by contrast is a good general purpose programming language that is syntactically easy to read (like BASIC, unlike Perl) and not so low level that you have to worry about pointers (as in C) directly.
@22 I understand his argument on the graphics front.
1. The lack of a GUI keeps things down to the brass tacks. If you give a new programmer a GUI they'll play with the GUI and not much else.
2. While JS dominates for client-side scripting in the browser, I think its going to have competition soon. I expect browsers will add Python interpreters soon-ish as MS Office recently did, if I'm not mistaken.
As for standalone executable portability, yeah, that's a weak point. Learn Python, then learn C.
@profoundlynerdy If I may continue to impose on your time—
1. I shudder to think how many kids we keep away from math and engineering by making them think that they have to be good at grade school arithmetic and calculation first. Similarly, I am ecstatic that visual (Processing.js, Cinder, OpenFramekworks, Scratch) and audible (Sonic Pi) programming draw so many in. As a learner and a teacher, I do not believe beginners should be made to prove themselves with stodgy, boring "basics".
(JS also rocks server-side. I find myself writing CLI apps/scripts just as often in Node as Python.)
@22 Yeah, there is a lot to learn... other technologies. I hate DOM. Hahaha
Browsers are neat! The inspector is a great tool.
Browsers have problems of their own, but those problems are solvable, and divorced from the OS and bundled software issues I was specifically talking about here.
We also might be talking about different age groups. I work with my nephew, and I do Code.Org\\hour of code with 2nd and 3rd graders.
To them, the inspector is pretty scary, but Code Combat or Pico-8 can go a long way.
@ajroach42 Super-duper thank you for your thoughtful remarks. You've convinced me to keep a Raspberry π+display+io with Pico-8 and Sonic Pi prepared. Yes, I'm mostly familiar with the high schoolers or older crowd (see http://nullprogram.com/blog/2016/09/02/ for a friend's story), and younger kids with only a shared tablet is a whole other level of complexity, I now see. Thank you! @profoundlynerdy
I taught myself BASIC to a limited degree in the 90s on a Commodore 128D and Apple 2 clone. I was pushed away from programming by people who expected I needed to be a math whiz.
Even as a dyslexia sufferer with math issues as a result, I'm still able to "grok" Python just fine. So much for the nay sayers.
@profoundlynerdy We are super-happy to have you coding with us <3!
John Taylor #Gatto in "Underground history of American education" railed against that cruel confounding of skills:
"#Pepys could only add and subtract right up to the time of his appointment to the Admiralty, but then quickly learned to multiply and divide to spare himself embarrassment. … You can learn what you need, even the technical stuff, at the moment you need it or shortly before."
@profoundlynerdy You know, this is probably where we get all these stupid people asking math/puzzle questions on coding interviews—this absurd 1990s notion that skill in grade school math (not really math, more like calculation) is correlated with skill in coding. There's no valid reason to ask for Gauss' sum formula tricks to see if you can code—if there was, you might as well as about KD-trees or FFTs. If you like people who like math, fine, but don't pretend you're interviewing for coders.
@22 I couldn't agree more.
Your'e much better off asking algorithm questions that are actually relevant:
* What is the purpose of a linked list?
* How might I optimize this to use less memory? Insert [example of a giant left-hand grab of SQL data that uses 2GB data]. Expected answer: stack/queue one 2 MB record at a time [example code].
Yadda Yadda Yadda...
JS is hard to read. It's hard to write. I've been doing web development for 10 years, and I still regularly stuff up little JS stuff in dumb ways (partially because I don't actually touch JS often, and partially because it's syntactically dense.)
This Mastodon instance is for people interested in technology. Discussions aren't limited to technology, because tech folks shouldn't be limited to technology either! We adhere to an adapted version of the TootCat Code of Conduct and follow the Toot Café list of blocked instances. Ash is the admin and is supported by Fuzzface, Brian!, and Daniel Glus as moderators. Hosting costs are largely covered by our generous supporters on Patreon – thanks for all the help!