The Real Story of Hacking Together the Commodore C128 179
szczys writes "Bil Herd was the designer and hardware lead for the Commodore C128. He reminisces about the herculean effort his team took on in order to bring the hardware to market in just five months. At the time the company had the resources to roll their own silicon (that's right, custom chips!) but this also meant that for three of those five months they didn't actually have the integrated circuits the computer was based on."
Mind blowing (Score:5, Informative)
The people I knew with 128s back then all used the 64 mode but used the 128 as an excuse to buy a better monitor. I never knew anyone using the CP/M mode.
Re:Mind blowing (Score:5, Interesting)
I went from a Vic20 to C128 instead of a C64. I was amazed that I could use CPM and a very advanced basic. The power of this machine enabled me and a good friend to build a robot in college made of nothing but old car parts, DC motors, relays, and plates with holes drilled in them for encoders. That directly led to my first job as an automation engineer.
The C128 also was the last computer that fueled my dreams. I went to college to become a computer engineer so I could build what I called the "compatibility machine". This machine could execute all the major 8-bit computer software (they all had Z80's or 6502) without the user intervening or worrying what version of software they purchased. The C128 showed me it could be possible!
By the time I left school the writing was on the wall that Mac / IBM style PCs would rule the world. It didn't stop me from getting an Amiga, but it was pretty clear that CBM was on the way out.
Re: (Score:2)
Re: (Score:3)
The closest I got to robotics back then was the Radio Electronics interface board to the Armatron... I never got the Armatron...
We used the user port to drive a board with 5 volt relays that, in turn, were used to turn on and off the DC motors (re-purposed windshield wiper motors). For input, I used the joystick ports since BASIC 7 had features to react to button presses, etc. and all the I/O was essentially just switches. I could POKE on the gripper motor and have the system react when the gripper closed "fire button" was hit before turning it off.
Reading and reacting to the encoders required a machine language routing to keep u
Re: (Score:2)
Re: (Score:2)
You don't have any pictures?
Someplace in a deep dark place with the rest of my college material there lives a picture. This thread makes me want to go dig around again! Probably sitting next to a banner saying "Commodore 128" made with Print Shop on a Citizen 120D and some really cool drawings I made on GEOS geoPaint.
I remember we named it MAXX but I cannot remember why anymore.
I have a soft place for the 8 bit machines and actually have a small collection: Osborne executive, Atari 800, Atari 400, Commodore 4+, several C64, C128, VI
Re: (Score:2)
Re: (Score:2)
OMG. You just reminded me about my first (sort of) "robot" -- I connected an Erector Set motor's power lugs to the switched power traces on the cassette interface of my c64 using alligator clips, and attached a weak rubber band to pull it back. It was utterly useless, and did nothing besides pivot a rod back and forth, but it WAS technically a crude robot capable of moving atoms via software ;-)
Thank ${deity} I didn't fry the cassette port. That would have really sucked, and it's the kind of thing that does
Megahertz myth and the 6502 (Score:5, Informative)
THe 6502 was an amazing processor. the Apple II was also a 6502. Unlike it's near contemporaries, the 8086 and Z-80 (and 6800), the instruction set was reduced. It had only 2 data registers (A,B) and two 8 bit address registers ( X Y) and fewer complicated ways to branch. Instead it effectively memory mapped the registers by using instructions like, offset Y by A, treat that as an address and get the byte at that location. Because it could do all that in one clock cycle, This effectively gave it 256 memory mapped registers. It also didn't have separate input lines for perifprials, and instead memory mapped those.
Nearly every instruction took a microsecond. Thus while the clock rate was 1 Mhz, it was much faster than a 4 Mhz 8080 series chip since those could take multiple cycles to do one instruction. Few memory chips (mainly static memory) could keep pace with that clock rate so the memory would inject wait states that further slowed the instruction time. The 6502's leisurley microsecond time was well matched to meory speeds. Moreover, on the 6502 only half the clock cycle was used for the memory fetch. This left the other half free for other things to access memory on a regular basis.
The regularity of that free memory access period was super important. it meant you could do two things. First you could backside the video memory access onto that period. On the 8080s using main memory you could often see gltiches on video displays that would happens when the video access was overridden by the CPU access at irregular clock cycles. As a result most 8080 series based video systems used dedicated video card like a CGA or EGA. Hence we had all these ugly character based graphics with slow video access by I/O in the Intel computer world. In the 6502 world, we had main memory mapped graphics. This is why the C64/Amiga/Apple were so much better at games.
This regular clock rate on the main meory had a wonderful side effect. It meant you could use Dynamic memory which was faster, cheaper, denser, and MUCH MUCH lower power than static memory. With the irregular access rates of the 8080 refreshing a page of dynamic memory requird all sorts tricky circuitry that trried to opportunistically find bus idle times to increment the dynamic refresh address, occasionally having to halt the CPU to do an emergency refresh cycle before the millisecond window of memory lifetime expired. As a result, the 8080 seris computers like Cromenco, Imsai, altair and Northstar all had whopper power supplies and big boxes to supply the cooling and current the static memory needed.
So the C64s and Apples were much nicer machines. However they had a reputation of being gaming machines. At the time that didn't mean "high end" like it does now. It mean toys. the Big Iron micros were perceived as bussiness machines.
Oddly that was exactly backwards. But until Visicalc, the bussiness software tended to be written for the 8080 series.
I think it was this memory mapping style rather than formal I/O lines to dedicated cards for periphrials (keyboard decoders, video, etc..) that lead apple to strive for replacing chips with software. they software decoed the serial lines (rather than using USART chips) they soft sectored the floppy drives rather than using dedicated controller chips, etc... And that was what lead to making the macintosh possible: less hardware to fit in the box, lower cost chip count, lower power more efficient power supplies.
Eventually however the megahertz myth made the PCs seem like more powerful machines than the 68000 and powerPC.
Re:Megahertz myth and the 6502 (Score:4)
And as a descendent to that is was amazing what the Amiga did with the 68000 and its custom graphics and sound chips, as you mention at the very end. you never saw smooth scrolling and sprite movement on a PC. The Amiga and the C=64 both had arcade quality graphics locked to a 60hz interlaced or 1/2 vertical res (single field) refresh rate of a standard NTSC television signal. Since the whole thing was timed to that frequency, you never got tearing. The only downside was interlace flicker without a frame doubler, but not a lot of applications used interlaced mode.
EGA and VGA scrolling (Score:3)
you never saw smooth scrolling and sprite movement on a PC.
This was true of CGA, but after EGA and VGA became popular, John Carmack figured out how to use these newer cards' scroll registers and built Commander Keen in 1990.
Re: (Score:3)
The reason for the popularity of the 6502 came down to one factor - cost. An 8086, 68000, Z80, etc., would've run you about $200 or so, while MOS was selling the 6502 for... $20. And you got a databook too.
The 6800 from Motorola was supposed to be the "cheap" chip (compared to the 68000), but it was still pricey - enough so that a bunch of Motorola
Re: (Score:3)
Re: (Score:2)
THe 6502 was an amazing processor. the Apple II was also a 6502. Unlike it's near contemporaries, the 8086 and Z-80 (and 6800), the instruction set was reduced. It had only 2 data registers (A,B) and two 8 bit address registers ( X Y) and fewer complicated ways to branch. Instead it effectively memory mapped the registers by using instructions like, offset Y by A, treat that as an address and get the byte at that location. Because it could do all that in one clock cycle, This effectively gave it 256 memory mapped registers. It also didn't have separate input lines for perifprials, and instead memory mapped those.
Actually the 6502 only had one accumulator, the A register. The 6809 had A and B. It is correct that the 6502 had very nice addressing modes. Zero page addresses acted more like machine registers. One commonly used addressing mode was z-page indirect indexed by Y. Two consecutive locations on z-page acted like a 16 bit pointer and register. Either that could be incremented OR Y could be incremented. So a block move of 1 256 byte page was easy.
I don't think I *ever* used ($23,X) where X selects the z-page l
Re: (Score:2)
What's kind of sad is that technically, VGA *did* have some of the same low-level capabilities of the C64 (besides sprites, obviously). At least, if you had a VRAM-based card like the ET4000. They just weren't supported by the BIOS, so they were (almost) never used in commercial software. You had to know how the video subsystem was wired together, where the various control registers were mapped, and bitbang them directly by hijacking system timers and dead reckoning.
One of the more hardcore examples I remem
Re:Megahertz myth and the 6502 (Score:4, Interesting)
I don't have time to correct all the errors in the parent post. So very briefly:
Schwab
Wrong (Score:4, Insightful)
Since I designed, wirewrapped, and programmed embedded 6502 and 8080 system in that era I am well prepared to assess your claims. In a nut shell you are an arrogant tard and the original poster is figuratively accurate inexact.
Your post is really bizarre.
Ok, with you there.
Well, that's just a bunch of crap: http://www.obelisk.demon.co.uk/6502/reference.html [demon.co.uk] (look at the "Cycles" column.)
What the original poster was likely saying, since it becomes clear later in the article, was that all the 6502 instructions were divided up into alternating cycles of memory fetches and internal calculations with an exact period of 1 microsecond for those. The 8080 series would use 1,2,3,4 and more with wait states cycles for an instruction with no regular pattern (in terms of future predictable times) of when the bus would be busy.
So you are wrong, have a reading comprehension problem, and are an ass about it.
The Wikipedia article on the 6502 indicates that DRAM access times were on the order of 250ns - 450ns. In particular, 250ns access times are well-matched to 4 MHz clock rates; do the math. At 1 MHz, 250ns DRAM has time to go make a sandwich before it needs to supply the next memory cells.
Sigh, again you have a reading comprehension problem. The original author was discussing static memory. Moreover, the cycle time for memory access always involves some overhead. The time when the CPU reads the data bus needs to occur after the bus has settled which is not at the start of the memories data valid period. But most of all 250ns memory was rare and expensive. Most computers in that time period did use wait states. Why do you think processors even allowed wait states?
Again you are being an ass about this as well.
No. Then, as now, video display glitches were caused by updating video RAM directly outside of a VSync pulse. You could just as easily get video glitches on 6502s as on 808x machines.
that was an additional restriction on 8080 machines. But on 6502 machines one did not have to wait for the vertical sync to update the video memory. In fact that is EXACTLY what the original poster was pointing out, without trying to flaunt jargon like you.
This makes you look stupid now.
Which leads us to:
Patently false. Video memory on an 808x machine (even on CGA and EGA cards) was most certainly memory mapped.
yes it could be done. But then you had the problem of glitches or waiting for VSYNC (or if you liked to live dangerously, HSYNC). It wasn't pretty to build hardware or write code for. Your interaction with it didn't treat it like main memory but rather so
Re: (Score:2)
I'm pretty sure you believe that, because the only 6502-based machines you've worked with have very low resolution. (Both the C64 and Apple II were 40x24.) I have a 6502-based machine, at home, whose resolution is 64x32, and it does require you to pay attention to Vsync. It's easy to play tricks like that, when you cut corners.
6502-machines' 40-column displays are probably one of contributing factors to their rep
Re:Mind blowing (Score:5, Interesting)
The C65 should have been what made it to market, not the weird 128 with its obsolete the day it left the factory CP/M mode running at half the speed of its competitors.
Whatever the merits or demerits of the two machines is irrelevant; the C128 came out in 1985, whereas the C65 [wikipedia.org] wasn't developed circa 1990-91.
C64 diehards have an obsession with the C65 and Commodore's perceived mistake in abandoning it, but despite the latter's numerous crap decisions, I'm sorry to say that in this case they were absolutely right.
The C64 was still selling as a budget option circa 1991 (*) viable due to sheer momentum. The 16/32-bit Amiga was not only established as the successor, it had already taken over (in Europe, at least) and was already nearing *its* own commercial peak(!)
Trying to release a (sort of) new 8-bit format by that point, even a very good one, would have made absolutely no sense, flopped horribly and stood on the low-end Amiga models' toes, mudding the waters pointlessly.
They could have sold it as cheaply as the C64 (i.e. the high manufacturing costs of a new machine selling at the same price as a "wringing the last profit from established cash cow model"), but what would the point of that have been?
The C128 at least came to market when there was still *possibly* a gap in the market for a high-end 8-bit machine between the C64 and the new (but still very expensive) Amiga.
(*) Apparently C= were still making them when they went bankrupt circa mid-1994(!)
Re: (Score:2)
Re: (Score:2)
Commodore was right to abandon the C65 by 1991. Yes. I think we agree there, I'm just saying C= should have focused earlier and the C65 would have more sense in the marketplace in 1986. Granted, it wouldn't have been the 1991 C65, sure.
But if C= had taken its engineers away from all the useless c
Re: (Score:2)
Spreading out into bizarre orphan architectures like the C16, C116, Plus/4, B128, C264 and all the other useless cruft they came up with.
While they (like Tramiel's Atari Corp. did later on) probably did too many overlapping things at once, it's only fair to point out that the apparently pointless introduction of a new, C64-incompatible architecture for the C16, C116 and Plus/4 family did supposedly start out for sensible reasons. According to the WP article [wikipedia.org], Jack Tramiel was paranoid that (as they'd done in many other industries), the Japanese would swoop in and undercut everyone with ultra-cheap consumer-oriented machines. That's why the ch
Re: (Score:2)
Re: (Score:2)
The C16 family was a good idea gone bad. Ideally, they should have released the C16 as a compatible successor to the VIC20
AFAIK, in the US, the C64 itself had become the de facto successor to the Vic 20 anyway, purely because it was being sold so cheaply there.
I also understand that this meant C= weren't actually making much money on them, and this is why Tramiel was forced out (i.e. he won the 8-bit computer market, but it was mostly a pyrrhic victory.) But that wasn't the end-buyer's problem...
At any rate, I think that by the time the C16 came out in late-1984, compatibility with the Vic 20 wouldn't have been that big a
Re: (Score:2)
(*) Ironically, the Japanese took over the US market another way, by launching the NES and everyone buying them for gaming instead of home computers.
The C64's good years are quite noticeable, 84 to 87. From the crash of 84, when people who still wanted to do electronic gaming almost had to jump to more expensive than a game console 8-bit computers, till the ascendancy of the NES. Twas Zelda that put the nail in the coffin. It didn't hurt that the NES was cheaper than a C64 system, without the load times, and with mostly better graphics.
But also, many of those who had C64's during that period only used them for games and only knew enough Commodore BASI
Re: (Score:2)
The C64's good years are quite noticeable, 84 to 87. From the crash of 84, when people who still wanted to do electronic gaming almost had to jump to more expensive than a game console 8-bit computers, till the ascendancy of the NES.
It lasted quite a long time in the UK- albeit having to share the market with the massive selling ZX Spectrum. Over here, the NES wasn't particularly successful (at least not compared to the US.)
In fact, the NES was outsold here by the Sega Master System- possibly because that was well-marketed, whereas AFAICT Nintendo didn't much care about Europe- but neither console dominated the UK gaming market, which remained mainly home computer based during the late 8-bit and early-16 bit eras. It wasn't until the
Re: (Score:2)
Re: (Score:2)
Having said that, had the C128 been a better successor to the C64, then things could have turned out much different. A successful 8-bit series through the late '80s might have eliminated the need to keep a budget entry model in the Amiga lineup. If we had a C256 and C512, the A500+ and A600 might never have been released.
Honestly? I think that would have been a major mistake.
Having a vastly improved higher-end 8-bit machine would have been good in the mid-80s, but it would still have been utterly misguided to rely on it as a replacement for a mass-market 16/32-bit machine; they'd have been hammered at the end of the decade as people moved towards true 16-bit models.
To have an 8-bit machine remotely competitive with what the Amiga 500 was- or even the Atari ST- they'd have ended up having to redesign the whole thing anyw
Re: (Score:2)
I had a C64 for years and at one time was slaving it to an Apple ][ with a nifty little interface, which I still have in a box somewhere. It was a dream to hack and play games on, despite having a mainframe at work which could do things I could only dream of at home (such as load/save from/to a HDD). My brother bought a 128 but never did anything with it as he wasn't a coder and had no idea what I was doing. Eventually I'd move to an Amiga 500 and then to a 2000 (which I still have.)
Re: (Score:3)
Wordstar! The 1571 floppy can read/write Kaypro formatted discs, as well as some other CP/M formats, and Commodore's own GCR'd CP/M format. With software the 1571 can read/write practically any 5.25 format out there, including DOS.
IIRC I've read tha CP/M on the 128 was popular for BBS sysops since it was inexpensive.
Re: (Score:2)
Re: (Score:2)
Re: (Score:2)
Well, apparently you never used the Plus4, in comparison to which the C128 looked like a nicely crafted supercomputer.
Re: (Score:3)
No smooth scrolling on IIGS (Score:2)
Re: (Score:2)
The C64 is not quite as good as the NES for sprite games, the NES can do more sprites, and the NES has tile based backgrounds.
However for pure bitmap and custom character set games (RPG's), the C64 had certain advantages....at least till the slow 1541 mattered.
Re: (Score:2)
The Apple IIGS was a 16 bit machine, so it was really competing with the Amiga and Atari ST machines.Whether it was competitive is another story.
Re: (Score:2)
Was my first as well, and ditto, I tried CP/M a few times and that was much pretty much it, didn't really serve a purpose for my seven-year old self.
What was cool about the C128 mode was the extended basic though. I was way too young / not autistic enough for assembler back then, but the basic had features like a rudimentary sprite editor and easy access to joystick input. I was able to create some "games" with moderate ease - they were horrible, of course, but at least I didn't just spend my time playing g
Re: (Score:2)
Any nice things I can say about it would go over you yong'ins heads.You're all abstract now - frameworks and whatnot - Java and shit like that .... do you kids even know what a register is? I think not ..
Not all people are abstract today. [greenarraychips.com]
Re: (Score:3)
Re: (Score:2)
Re: (Score:2)
Re: (Score:2)
Re:Mind blowing (Score:5, Informative)
One grain of salt (Score:2, Informative)
From the Article: "Commodore C-128, the last mass production 8 bit computer and first home computer with 40 and 80 column displays"
C-128 was in 1985, the Acorn BBC had 20, 40 & 80 column modes (and a teletext mode) in 1981.
Re: (Score:2)
C-128 was in 1985, the Acorn BBC had 20, 40 & 80 column modes (and a teletext mode) in 1981.
Yes, this is correct. Technically, I guess it could depend how one interprets
[The] first home computer with 40 and 80 column displays, dual processors, three operating systems, 128k memory via MMU and one heck of a door stop.
Was the BBC truly a "home computer"? I'd say yes, though it overlapped the educational market too, but one could argue the point.
And perhaps it could have meant "(40 and 80 column displays) BOOLEAN-AND (dual processors) AND (three operating systems) AND (128k memory via MMU)".
That said, this is probably overanalysing. The BBC Micro wasn't that successful outside the UK, and the US tech industry (well, the US in general!) tends
Re: (Score:3)
The thing with the C128 is that you can use both displays at the same time, meaning you can have a 40 column display hooked up AND an 80 column display. Most people used dual-mode monitors but there was some software that you did some things in 40 column mode but then the software displayed special output in 80, or vice versa.
Learn electronics repair from Bil himself (Score:3, Informative)
Bil will be teaching a class at the Vintage Computer Festival East [vintage.org] next spring. He also lectured about the 128 and Commodore repair at the same event in 2012. Details are on c128.com.
Hrmmmmm (Score:3, Interesting)
He is claiming a lot of "firsts" that I would swear were in my Apple ][e prior to Winter '85...
Re: (Score:2)
Re: (Score:2)
Which ones? While 80 columns and 128K were options on the //e, 6he //e didn't come with 128k as default till 1987 with the Platinum //e. That was also the first //e with a numeric keypad by default. The 1571 also has a higher capacity than Apple's 5.25" drives.
So yes, the C128 did have some features as standard before the //e.
Re: (Score:2)
So what? (Score:2)
A lot of early personal computers have a similar story. Software is often written with breadboarded or nonexistent hardware.
What is unique about the idea of custom silicon LSI chips for a 1980's PC?
The original Atari 800 (a design later copied by Commodore for the VIC-20 and Commodore-64 computers) had three custom chips (ANTIC, CTIA, POKEY) which made up the majority of the machine's circuitry when designed in 1978. And the OS and other early programs were written without the benefits of that completed
It was worth having the 128... (Score:3)
...to play Ultima V in dual SID mode.
After several C=64s and the 128, I moved to the Amiga, which got me into the VFX business thanks to the Video Toaster and Lightwave.
Looking forward to reading this article. If it's good I'll stash a copy next to my "Rise and Fall of Commodore" book.
This quote is great (Score:5, Insightful)
a couple of weeks later the marketing department in a state of delusional denial put out a press release guaranteeing 100% compatibility with the C64. We debated asking them how they (the Marketing Department) were going to accomplish such a lofty goal but instead settled for getting down to work ourselves.
Re: (Score:2)
It was actually a pretty important selling point of the C128. Keep in mind that I (and many others) had a collection of *hundreds* of C64 games before we bought the 128 (thank you, early DRM crackers). I probably wouldn't have bought one if all it could play was C128 software (what little there was of it).
Re: (Score:2)
Re: (Score:3)
The Atari STe line had the same problem with games, very few took advantage of the improved graphics and digital sound available on the newer machine.
The STe was clearly designed to close the gap between the "vanilla" ST (and STFM) and the Amiga, which had come down in price by that point. It might have worked... had Atari directly replaced the STFM with the STE at the same price when it launched.
Problem was that- almost certainly due to Jack Tramiel's penny-pinching short-sightedness- they charged more for the STe and continued to sell it alongside the STFM. So anyone buying an ST because it was cheap would get the STFM, and anyone who had a bit more
Re: (Score:2)
My dad bought me this as my first computer. (Score:5, Interesting)
I was in 5th or 6th grade, and I woke up to a new computer in my room. The printer immediately broke and I noticed the desk was half up-side down. My dad had assembled it and the desk in the dark, during the night, while I was asleep (I'm a heavy sleeper). He was no technician, but I appreciated the effort. I traded c64 games with kids at school and stacks of 5.25 floppies via mail. Commodore games were fantastic; much better than NES. Junior year of High School, I finally had the initiative to figure out what my dad had done to the printer, and it turned out to be a simple problem that I fixed. I used 80 column mode to type and print essays for school for the next two years. Much praise to my old man. Granted, first year of college and he helped me acquire a 386 with Windows 3.0, which I had for three years, then built my own. I'll never forget my C=128. Thanks, dad!
Re:My dad bought me this as my first computer. (Score:4, Informative)
Just wanted to say that's a great story about your dad.
Too little too late (Score:4, Interesting)
I was a big fan, and a game developer for the C64. Those were the days that a machine could be fully understood by an untrained person with a knack for programming. When the C128 came out, I was interested, especially in the 80 column screen and CP/M software compilers. But there were too many limits on the machine (no hard drive easily added, no real OS, etc.) and it didn't feel like enough of an advancement over the C64. My grandfather did buy one, and I had some time with his, but that never really sparked much either. My next machine would be the Amiga, and as soon as that become somewhat affordable by a college student (the A500), I never looked back.
RM
About time (Score:4, Funny)
Excellent, my wife's been on me to upgrade my C64
Software Design only took three people to write .. (Score:2)
Not the last 8-bit design (Score:5, Informative)
Re: (Score:2)
Similar to the Amiga (Score:3)
While not originally designed under the auspices of Commodore, the Amiga was also designed with VLSI custom chips. The prototype did not have the chips available, either. Instead, their larger-scale prototypes were the size of a small room. The booth they used at a trade show to demonstrate the Amiga was designed to hide the fact that the walls of the booth itself consisted of the prototypes of the custom chips hidden behind curtains.
It would be cool if we could find a photograph of it.
RIP Jay Miner.
Re: (Score:3)
More on the period: At this time I had been a die-hard TRS-80 Color Computer aficionado but I did appreciate the advances in which the Commodore camp had triumphed. I eventually embraced, for better or worse, the Commodore Amiga line, from 16- to 32-bit both in AmigaOS and Amiga Unix.
Re: (Score:2)
I also heard it called Comma-toy and Commode-door, at the time.
Re: (Score:2)
I called it "Commode-odor". I was an Atari fan, but most of my friends had C=64s.
A few years later, though, I got an Amiga.
Re:U.S. Navy? (Score:5, Insightful)
I called it "Commode-odor". I was an Atari fan, but most of my friends had C=64s. A few years later, though, I got an Amiga.
Assuming you mean the 8-bit Atari 400 and 800 (and its compatible redesigns, the XL and XE series), I did pretty much the same thing- was an Atari fanboy, but ended up with an Amiga. When one knows a little more about the "Commodore" Amiga and "Atari", it all seems a bit silly.
The major irony is that the Amiga developers included a number of ex-Atari staff- most significantly Jay Miner- who had worked on the 400/800 and the VCS/2600 before that. It represented (some have argued) a continued thread of architectural design that the 400/800 had significantly improved upon from the VCS, and had the same state-of-the-art custom chipset approach as its predecessors. (Indeed, just as happened with the 400 and 800, the Amiga was originally meant to be a console, before it evolved into a computer).
Also worth noting that "Amiga" was originally an independent company and it was only later bought by Commodore (after some legal wrangling with Atari, who'd had some involvement with them).
Meanwhile, Jack Tramiel had left Commodore (after falling out with the management), bought Atari Inc's computer and console division (i.e. the one that brought us the VCS and 400/800), which formed his new Atari Corp. The latter was a very different company to Atari Inc. (very obviously a much more shoestring operation). The Atari ST was designed by a different team after Tramiel had sacked most of the old Atari Inc. engineers, and very much reflected the "new" Atari; affordable, but much more off-the-shelf parts.
Atari Corp continued selling the XL and XE (cost-reduced versions of the 400 and 800), but they didn't design it; they merely milked the profits from a design they'd inherited while they focused on *their* Atari ST.
So... which was really the "true" successor to the Atari 400 and 800? By any measure, it was the "Commodore" Amiga. Who cares who made it? I briefly owned an ST because I couldn't afford an Amiga, but I ended up selling it and buying the latter a year later.
Re:U.S. Navy? (Score:5, Interesting)
True. I was one of those guys initially. I was a CS major in 1985 and my computer experience consisted of mainframes, CP/M machines and IBM PCs. Anything else was a "toy". I had a new girlfriend who suggested I could do my CS homework at her house because her dad "had a computer". She didn't know what kind it was, but he was an engineer so I figured it was probably pretty nice. I took her up on her offer, figuring that the suggestion was nothing more than a ploy to get me to come over to her house. When we got to her house she took me to the room he used as his office and pointed. I literally guffawed. It was a Commodore 64. She was somewhat offended at my reaction and I quickly apologized. Over the next few weeks I was a frequent visitor to her house and I began playing with the C64. The more I worked with it, the more respect I had for the platform. I especially liked the serial interface and how components could be daisy-chained. Far from being a toy, the C64 had the capability to do some pretty advanced stuff. And it was a LOT less expensive than an IBM PC. Eventually, the girlfriend became my wife, and her dad gave me the Commodore after he moved on to a PC. The wife and I broke several years ago. I still have the C64.
Re:U.S. Navy? (Score:5, Funny)
This is such a Slashdot story :)
"A girl invited me to her house on several occasions. Each time, I spent more and more time being impressed with the Commodore 64"
Re:U.S. Navy? (Score:4, Informative)
Well, you know, there was other stuff going on, as her father and step mother were out of town that winter, but this is Slashdot, not Penthouse forum, so...
Hey, I eventually married the girl. ;-)
Re: (Score:2)
The wife and I broke several years ago. I still have the C64.
Ahh yes, the old wife mate vs the machine issue. Probably been an issue since the first caveman banged out the first wheel. I bet the wife of the first cave monkey to discover fire probably rode his ass over that too.
Don't worry, many of us have had that issue, having to chose between our computer and a mate. Many of us made the same choice you did. Good thing the internet came a long and is filled with porn!
Oh my god, its full of porn!!
Re: (Score:2)
But they were what the cool kids had for toys back then.
Cool "affluent" kids, the vast majority of kids back then didn't own computers. It is only on Slashdot where everyone assumes everyone was one of those spoiled suburban kids with a WarGames or Ferris Bueller style set up like they had. You'll see things like:
"When I was 15, my Quantumlink/Compuserve/Source bill was around 300 a month"
Re: (Score:3)
It past the spell choker.
Re: (Score:3, Insightful)
Want to do a crazy program you can't write on modern computers?
What?
Simply loop through a sequence of poking two random numbers, and incrementing a number that you print.
What?
Every time, the system will do different things.
What ?
If you did this on a modern computer, eventually it'd corrupt system files and the thing wouldn't boot.
WHAT?
It makes you wonder why modern OSes aren't hardened with the theory: No matter what the user does, allow the computer to boot up safely next time.
You're an idiot.
Re: (Score:2)
What he's suggesting is possible but what's far more likely is the application and/or computer would crash long before anything truly bad could happen. In Linux I think you could even script such a thing with ptrace, not like I'd ever bother.
Re:Mistake (Score:5, Informative)
Want to do a crazy program you can't write on modern computers?
What?
Yeah, can't is a blatant lie.
Yeah, that's trivial to do on a modern computer too. A trivial loadable kernel module in linux could do so, for example.
Simply loop through a sequence of poking two random numbers, and incrementing a number that you print.
What?
That is what it says, write a random value to a random memory location in a loop.
Every time, the system will do different things.
What ?
Of course it will. Sometimes you random memory location will be the memory mapped to the screen and a character will show up. Sometimes you'll change a return address on the stack and run some random code.
If you did this on a modern computer, eventually it'd corrupt system files and the thing wouldn't boot.
WHAT?
That's true, eventually you'll write over some file data just before it is flushed to disk and trash a file required for booting. Or screw with memory the file system is using and mess that up on the next write (though given the use of checksums that's pretty unlikely). The key is eventually since you'll have to run it a *lot* of times before it does something like that before crashing itself.
And of course not when running as a normal user process.
It makes you wonder why modern OSes aren't hardened with the theory: No matter what the user does, allow the computer to boot up safely next time.
You're an idiot.
Yes he is.
Computers that have the OS on ROM unsurprisingly aren't susceptible to making the system unbootable by screwing with boot files. The same is true of a modern computer hardwired to boot off of ROM as well though. And of course it makes upgrading that base OS essentially impossible (short of replacing the ROM, or actually using an EEPROM - and of course if software can do the upgrade then the random memory setting could also cause it to happen and screw up booting)
Re: (Score:2, Interesting)
Well, he's saying that you can't write a program that simply twiddles random memory regions because all modern OSs employ protected memory schemes to avoid exactly this. Modern computers are multi-tasking systems that do hundreds or thousands of things at once and doing so would be disasterous.
On an older computer without protected memory you can write such a program and the results are beyond bizzare. Most forget that these old systems are very very "Bare metal". They don't have the layers of abstraction,
Re: (Score:2)
"Computers that have the OS on ROM unsurprisingly aren't susceptible to making the system unbootable by screwing with boot files."
Ummmmm, BIOS. Every computer essentially has an OS in ROM as-is, or it wouldn't DO SHIT.
Might want to go brush up on your A+ skills, Johnnyboy.
Re: (Score:2)
The bios is not the OS being talked about, which should be obvious since if that is what you are considering then modern computers are exactly the same as older ones - just turn it off and on again and you'll be back at the bios boot screen just fine.
Re: (Score:2)
I can tell you don't hold an A+ repair certification, let alone failed high school critical thinking and reading comprehension. Here, let me repeat myself.
"Computers that have the OS on ROM unsurprisingly aren't susceptible to making the system unbootable by screwing with boot files."
EVERY COMPUTER HAS AN OS IN ROM. That's how you can load any other OS onto it in the first fucking place. You brick the BIOS by failing an update. Bad BIOS, system WILL NOT FUCKING BOOT AT ALL (excepting systems with dual BIOS
Re: (Score:2)
Except the subject is the higher up OS not the bios. Making the computer unbootable means that the bios fails to boot the actual OS the user wants to run in the context of this particular discussion. Which should be obvious really.
Yes you can brick the bios too - I explicitly mentioned that if your ROM isn't actually RO then if you get really unlucky you could hit the right sequence of pokes to screw with it. Did you get your repair certification without learning to read or something?
Though I guess that you
Re: (Score:2)
I wish. Revolutionizing an entire field of science would be pretty good for 3 minutes work after all.
Re: (Score:2)
Indeed.
The "reset-ability" of older systems (I myself learned on a dragon32, which is basically a trs-80 knockoff) was very reassuring. You could crash the thing, but press the black button on the side and it's like it never happened.
Of course bringing this to modern computers would probably be hard and have all kinds of other consequences.
Re: (Score:2)
You can certainly do it on modern systems. LOTS of systems are written this way. You write your OS to media that can't be changed after the fact (DVD) and boot from that. No changes to the os can happen without re-burning it. That's how my hardware firewall is setup. It stores logs and such to the hard drive but the entire OS and all config files are stored on a DVD. There's no hacking that and making a permanent change.
The wright ability of modern OS's is definitely a "Feature" and very important, but it d
Re: (Score:2)
OS in flash (Score:2)
Well, a ROM then was pretty small, but today, given how the BIOS has been redefined - w/ UEFI & all that, wouldn't it be possible again? Take a flash memory device that's 32GB, put an OS in it and make that the BIOS. For the rest of the stuff - the applications and all that, take a suitably sized SSD and put it on that. Anything portable would go on a USB drive.
Lock that OS BIOS, making it alterable only by the owner (in the same way that we currently alter BIOS) and all attacks that cripple an
Re: (Score:2)
If you take into account live CD's, it's actually been possible for a long time. Just burn your OS to CD or DVD, put it into a computer that lacks a hard drive, and you've basically got a system that you can easily restore back to it's original, working state by just hitting the reset button. Though the several minute (typical) boot time from the optical media is a lot slower than a C64. You could probably speed that up by putting the OS onto some kind of write-protected flash memory, though you may have
Re: (Score:3)
Re:Mistake (Score:5, Funny)
Re: (Score:2)
Re: (Score:2)
Comment removed (Score:4, Interesting)
You're giving away your age .. (Score:2)
You're giving away your age
Re: (Score:2)
Re: (Score:2)
Hey, self-love is love too right?
Re: (Score:2)