8-bit. 16KiB of RAM. BASIC as the programming language. 640x256 resolution in 8 colours.
I could make that thing sing in an hour. It was hard to get it to do much, but then the difficulty was the fun thing.
By the time we got to the early 2000s and I could buy something with more RAM, CPU and storage than I could ever reasonably max out for the problems I was interested in at the time, I lost something.
Working within constraints teaches you something, I think. Doing more with less makes you appreciate the "more" you eventually end up with. You develop intuitions and instincts and whole skillsets that others never had to develop. You get an advantage.
I don't think we should be going back to 8-bit days any time soon, but in the context of this post, I want novices to try and build software on an A18 chip, I want learners to be curious enough to build a small word game (Hangman will do at first, but the A18 will let them push way, way past that into the limits of something that starts to feel hard all of a sudden), to develop the intuition of writing code on a system that isn't quite big enough for their ideas. It'll make them thirsty for more, and better at using it when they get it.
It absolutely does. But every system has constraints; even when provided with massive resources, humans tend to try things that exceed those resources, as evidenced by Parkinson's Law of data https://en.wikipedia.org/wiki/Parkinson%27s_law
The game Elite did something extremely evil and clever: it was actually able to switch between modes partway through each frame, so that it could display higher-resolution wireframe graphics in the upper part of the screen and lower-resolution more-colourful stuff for the radar/status display further down.
“Tomorrow's World: Nellie the School Computer 15 February 1969 - BBC”
I'd install Photoshop and Illustrator on my shitty computer I put together from spare parts my dad didn't have the use of anymore from his business computers. It was horribly slow, but I kinda made it work slowly.
The thing is that I think this is what made me think a bit differently, since everything was slowed down and took more time than I would want it to, I had to make deliberate decisions on what to add/edit. I still work the same way today to pa point, but that's because I'm both faster, more experienced and the computers have gotten more performant (and because I can afford better devices sure).
When I look at my half-brother and his teenage generation I wonder if they can still have such an experience. The personal devices have gotten better and faster, most things are really convenient and you sometimes even don't have to think a lot to do something also because they're cheap to do... they probably won't have the experience of "grinding it out" just for the sake of producing something they like...maybe sports is the closest...no idea, but have been thinking about this quite a lot recently...
when you're young, time is infinite, money is scarce.
Older, and time seems to take over. The limitations are - when can you free up the time? Is relaxing allowed?
I have a typical yuppie software job with decent pay, so generally I will buy the right tools for a job now instead of trying to make due with whatever I can scrap together. I'm not that busy of a person, but I certainly have more obligations than I did when I was sixteen, and now sometimes it really is worth it to spend an extra grand on something than it is to spend a week hacking together something from my existing stuff.
Still, I look back at the hours I spent making terrible YouTube videos with my terrible computer really fondly. I was proud of myself for making things work, I was proud of the little workarounds I found.
I think it's the same reason I love reading about classic computing (80's-90's era). Computers in the 80's were objectively terrible compared to anything we have now, and people still figured out how to squeeze every little bit of juice possible to make really awesome games and programs. The Commodore 64 and Amiga demos are fun to play around with because people will figure out the coolest tricks to make these computers do things that they have no business doing. I mean, the fact that Bad Apple has been "ported" to pretty much everything is something I cannot stop being fascinated by. [1] [2] [3] [4]
[1] https://youtu.be/2vPe452cegU
[2] https://youtu.be/qRdGhHEoj3o