Gordon_Peterson_DP_history

I was responsible for development of the company's Disk Operating System
post-DOS 1.2 and (in Advanced Product Development) was the person responsible
for the ARC System (LAN) software (which I proposed and wrote).

There are a number of errors in your present pages and also in some of the
"extra information" page data.  I'll try to set many of those right with this
message, as well as adding a little more information that you apparently didn't
have before.

On the Datapoint 2200 page: You've got the chronology wrong about the birth of
the microprocessor. Computer Terminal Corporation started out (as you might
imagine) building "glass teletypes" and for a time was the largest third-party
manufacturer of CRT-based computer terminals.  One of the keys to their success
was a patent for an unusual way of "painting" alphanumeric data to a CRT
display... the "diddle scan".  Instead of a television-style "raster scan", the
(early) Datapoint displays instead painted each character, column by column,
then the next character column by column, and so forth, and then row by row to
the bottom of the screen.  This patent allowed them to use (relatively slow, but
much simpler and less expensive) CMOS shift-register memories to hold the data
on the screen and refresh it at the required rate.  They only needed to access
each position in that memory ONCE for each screen refresh, since they only had
to check each character cell ONCE per refresh, and then painted all 35 pixels of
that character (7 high by 5 wide) before continuing to the next character on the
same row. As a result of these terminal sales, Computer Terminal Corporation was
at the time the world's largest buyer of MOS memory.

One of their successful products was the 3360 terminal system (which was a
clustered system based on a rack-mounted HP minicomputer and the 3360 terminals)
which were a successful replacement product for the IBM 2260 clustered CRT
display system.

At the time, anyhow, all these terminals (as sold by different manufacturers)
were chaotic;  there were no real standards for things like cursor positioning,
handling of characters displayed past the end of the line, screen scrolling (the
2260 and 3360 didn't even support scrolling at all!), wrap (or not) at end of
line, baud rates, parity, how many bits long characters were, or indeed much of
anything.  Everyone's terminals were different, and the result for a third-party
company building "compatible" CRT terminals was a confusing maze of factory
strapping options that had to be specified when the terminals were ordered.

It seemed like it would be simpler and better for everyone concerned if these
terminals could be configured in SOFTWARE to implement all these kinds of
options, and a fellow named Vic Poor (who I think was at Frederick Electronics
at the time) decided that it shouldn't be all that difficult to implement the
simple instruction set required with few enough gates and parts that it could be
implemented on a single and relatively inexpensive silicon chip, and also could
use the same kinds of serial shift register memories that the company already
was buying in large quantities.  Vic, another fellow Jonathan Schmidt, and a
young man named Harry Pyle rapidly evolved the concept (Harry was notably
responsible for the use of a hardware stack as the subroutine call return
address mechanism).  The use of the shift register memory for programs was the
reason for use of LSB-MSB byte order... (which Intel uses to this day!) it
allowed readily propagating the carry to the higher order byte of addresses,
since the high-order byte was the next byte to come out of the shift register
that held the program.

[Note that this microprocessor was not "shopped around" to other companies.  It
was intended for CTC's "terminals" from the beginning.  Case Western Reserve
University was indeed significant in the history of CTC... I think that Harry
Pyle attended there (but as a student, not as a professor) and Mike Green (who
made a number of contributions to the company's software over the years,
including notably writing the STATH (String mATH) package that Databus used) was
likewise.  Harry wrote much (most, probably) of the original software for the
2200, including the CTOS, the initial Databus language compilers/interpreters,
and the first three releases of the original Disk Operating System (through DOS
1.2, when I took over DOS development).  By the time I arrived in May 1974,
there were maybe eight or ten programmers in Datapoint Software Development.

The machine (which became the 2200) was prototyped using TTL and MSI parts, and
in the end the TTL/MSI implementation was the production version throughout the
2200's lifetime.

A number of integrated circuit manufacturers were approached to make the part
(which was to become the first general-purpose 8-bit microprocessor on a single
chip) and nobody really wanted to make it (Intel, in particular, was convinced
that there was no market for a general-purpose microprocessor on a chip... Intel
was a memory company... but since Computer Terminal Corporation was the world's
largest buyer of MOS memories, it's fair to say CTC had Intel's attention since
they saw it as a way to cement CTC's memory business).  In the end, TI delivered
a prototype part (but it barely worked, the rumor was that there were but
millivolts of difference between signal and noise levels) and Intel's one was
significantly better (but was delivered very late).  CTC had meanwhile started
shipping the 2200 in the MSI/TTL implementation.  The Intel part (which became
the 8008) was enough of a hassle to interface, wasn't as fast as the MSI/TTL
version, and was pricey enough that in the end CTC decided to give Intel the
rights to the part in exchange for being relieved of the obligation to buy them
(of course, don't grieve for Intel, in the meanwhile they had become convinced
that maybe, just maybe, there WAS a market for a single-chip, general-purpose
microprocessor after all! :-) )

There is a thesis written by a student several years back which went into all
this early microprocessor history in great detail, including contacting and
getting interviews with most of the key players.  I'll try to locate my copy of
that he sent to me and make it available... it's probably the best single
historical document about those issues that has been produced to date.

Finally, I think the first company to actually implement and ship an 8008-based
system for small offices was a company in New York called Q1 as I recall.  For
whatever reason, they didn't succeed and that company closed after a few years.

Meanwhile, the Datapoint 2200 was amazingly innovative for a whole variety of
reasons.  In a day when computers universally used big and heavy
transformer-and-capacitor-based power supplies, the 2200 contained an exotic
switching regulator power supply (positioned across the back side of the unit). 
It was the first complete, general purpose business-oriented computer system in
a single, attractive (designed by Jack Frassanito, to give credit where credit
is due) desktop enclosure that was the size of an office electric typewriter,
and integrating a CRT display (12 rows by 80 columns), office-style keyboard,
and digital tape storage.  The traditional row of bit-oriented toggles and
blinking lights (found on most all other computers of the era) was replaced by
the five keyswitches at the right of the keyboard... (top to bottom:) RESTART,
STOP, RUN, KEYBOARD, and DISPLAY.  There were (incandescent bulb) indicator
lamps under the lowest four... STOP and RUN reflected the status of the
processor, and the KEYBOARD and DISPLAY lamps were software-programmable.

The two digital cassette tape drives on top of the unit each held about 150K
bytes and could read (and write?) both forwards and backwards, at two speeds
(programmable) plus a (faster) rewind speed.  At normal speed they could read or
write 350 characters per second (as Vic once pointed out to me, that looked
pretty good compared to punched paper tape speeds, which was the standard for
minicomputer temporary storage at the time!)

The 2200 was delivered with a Cassette-Tape-Operating-System (CTOS) with a tape
file structure and directory, keyboard commands, a cassette-based general
editor, several cassette-based compilers and an assembler... this machine looked
and felt a LOT more like a computer than it did like a "terminal"!!!!

When the RESTART key was pressed, the rear tape would (by a hardware bootstrap
process, NOT firmware!) rewind to the beginning, then start moving the tape
forward and read the first block of data from the tape into main memory
(starting at location zero), then stop the tape deck, and pass control to
location zero.  Normally that block would be a bootstrap which would proceed to
find and load the next file from the tape, which would be either CTOS or (on a
"load-and-go" tape) the editor, compiler, or whatever other program was on the
tape.

The 2200 came in two versions... Version 1 and Version 2.  Version 1 used serial
shift register memories for program storage (which had 'drum-like' timing
characteristics, which is where some of the confusion may have come from) and a
seven-level subroutine call stack.  Version 2 used random-access memories for
program storage and had a larger 15-level subroutine call stack.  Version 1 used
one to four memory cards, which had 2K bytes each, for a maximum of 8K.  Version
2 used one to four memory cards, which had 4K bytes each, for a maximum of 16K
bytes.  Version 2 supported a programmable timer interrupt which could pass
control to location zero every millisecond... the only interrupt supported by
the system.

The CTOS ran in 8K bytes of RAM (for the Version 1) and ran (better) in 16K
bytes.  The system would run, in small configurations, with 4K bytes of RAM
(there were several limited versions of the Databus programming language which
could run in 4K of RAM).

The DOS could run in 8kb and up configurations, although in practice most or all
2200-based disk systems were 16K machines.  The DOS included a simple
software-based debug facility which could (among other things) include a
real-time program counter display in the last column of the CRT display... so
debugging (even on 2200s) wasn't QUITE as bad as has been suggested.  :-)

As for mainframes of the era, the other members of the "seven dwarfs" were NCR
and RCA.  For a long time the "other" companies besides IBM were referred to
collectively as the "BUNCH"... Burroughs, Univac, NCR, Control Data, and
Honeywell.

It turns out anyhow that Datapoint never bought and used the 8080 either.  By
the time it was released, it was "too little and too late" since Datapoint was
already well along in developing the 5500 (which was MUCH faster and much more
capable, in general).  The 5500 came with 48kB of main memory (3 16Kb memory
boards) plus an additional 8Mb of RAM contained on the BIOS ROM card (containing
bootstrap code and a BIOS=based software "debugger" tool).

I had previously "usurped" the 8K of RAM on the BIOS card for use with my
virtual-machine Partition Supervisor, which allowed running multiple independent
copies of the DOS within the memory space of a 5500.  (The release version was
restricted artificially to just two partitions, one Datashare partition and a
DOS partition).  When it came time to develop the ARC System, I used the same 8K
of RAM (actually just 7.75K, since one 256-byte page of that 8K was used by the
debugger tool) for ARC... 4K for an emulation of the data buffers (normally in
the disk controller) and the remaining 3.75K for the ARC System client software,
redirector, enqueue/dequeue subsystem, etc.

When the ARC System was being developed, I/we initially wanted to call it a
"network" but that was not accepted by management, since "networks have the
reputation of being complicated and hard to manage".  My initial proposal
document referred to it as "DISPDOS" (for Dispersed DOS, "dispersed" being a key
part of the company marketing message at the time).  Other names that were
kicked around at the time included Datanet (not cool, since Honeywell used
"Datanet" as their trademark for their communications controllers, and they were
a big OEM customer of Datapoint that we didn't want to annoy), DOSNET, and even
(believe it or not) "Internet".  :-)  Eventually we settled on Attached Resource
Computer, or ARC, demonstrating that it was a system to attach networked
resources together to form a single, cohesive, basically seamless computing
system.

ARCnet and The ARC System was the first commercially available LAN, and ARCnet
(at 2.5 megabits) was faster than the two megabits which was the speed Ethernet
(still under development) used at the time (in its clumsy thick-wire
implementation, which was never very successful... even at its later 10Mbits
'release' speed.  The second, thin-coax-based version was more successful, but
Ethernet didn't really "get it" until they finally adopted ARCnet-style active
hubs and ARCnet-style "interconnected stars" cabling topology).  The ARC System
was proposed in summer 1976, I started coding shortly thereafter, and the system
started coming up in its original release mode (good enough for me to move my
development onto the new OS) in early spring 1977.  The first out-of-house
installation for the ARC System was at Chase Manhattan Bank in late September
1977, and the product was announced to the public in a press conference on
December 1, 1977 at the Plaza (Hotel) in NYC.

At the time, I commented privately that we were pushing a big rock over a very
tall cliff on that day, and that from that day forward the days of mainframes
for business computing were numbered.  Even my friends said, "Gordon, you're
crazy... companies will never give up their mainframes and run their processing
on networks of little computers!"... My reply was simply, "JUST WAIT AND WATCH!"
:-)

The 2200 never exceeded 16Kb... although that was augmented (slightly) by very
clever use of the "speed buffers" of the disk controller (4 256-byte pages in
the 2250-350 disk system, and 16 256-byte pages in later disk controllers) in
conjunction with Datashare's "virtual memory" approach, which interpreted
compact compiled pseudo-code directly out of the disk controller buffers.

The 2200 had very limited ROM... basically just a hardware character generator
used for refreshing the CRT screen display, but it was not (in the original
version) program-accessible.  A later "RAM display" CRT controller board for the
2200 added program-alterable character generator (which was initialized from the
ROM copy, but the ROM was still not program-accessible as such).

The 2200 through 6600 systems all used diddle-scan displays and thus did not
support any graphic modes (the character rows on the screen were separately
scanned and did not meet on the screen) although some limited "graphic"
capabilities could be achieved by dynamically loading the character generator
under program control.

The beep capability on those systems did not use a beeper "in the keyboard".
There was a small speaker on the (inner) front panel of the computer,
approximately in the 5-o'clock position under the "D" logo.  No speaker grille
was present (nor was it required).  The processor supported two interesting
instructions to control the speaker... a "Beep" instruction and a "Click"
instruction.  Other tones and sounds were later produced by combinations of
those two instructions, including "clicking" at a rapid rate which could
implement tones at other frequencies.

Actual dimensions of the 2200 (and 5500 and 6600) were 9-5/8" high, 18-1/2"
wide, and 19-5/8" deep.  The 2200 weighed 47 pounds (I'm not sure how many
memory cards that figure includes).

The 2200-family processors supported up to 8 serial terminals... 5500 and 6600
systems supported up to 24 terminals.  Initially, the single-user Databus
programming language had been upgraded to a timeshared/multiuser mode (called
Datashare) to make good use of a bunch of the company's 3360-type terminals
which had been coming back in off lease.  Datashare systems ended up being the
company's "bread and butter" product, an astonishingly capable and efficient
system which mated perfectly with the hardware's capabilities.  Initially it was
thought that the reason for Datashare's great success was because it reduced the
cost per terminal-hour, but I'm convinced that at least as important was the
fact that it allowed multiple users to interact very effectively against a
common shared base of data files.  Of course, the ARC System (where suddenly it
didn't make much difference where files were physically located within the
aggregate system) only enhanced that quality, and really was the enabling
springboard that set the company up for unprecedented growth and success from
1978 through the early 1980's.

The ONLY I/O connector on the 2200 (/5500/6600) system was the 50-pin I/O
connector on the back next to the power supply, which was basically the
processor's I/O bus.  Other peripherals (serial, printer, LAN, disks, etc) were
daisy-chained using (via external interfaces) that I/O bus.  All I/O was
programmed, byte-at-a-time through the accumulator.

The 2200 (and subsequent 5500 and 6600) never used a storage drum.  The disk
systems used by those systems included: 1A   E  Diablo 2.5Mb cartridge disk (1-4
drives) 1A   E  Wangco 2.5 fixed/2.5Mb removable disk cartridge drive (up to 2
drives supported, 4 disks) C  F Shugart single sided/single density floppy disk
drive (1-4 diskettes supported) B D   Wangco 10 fixed/10 removable disk
cartridge drive (up to 4 drives supported, 8 disks) B D   Memorex 660 25Mb
(2314/2319-type) mainframe-class removable pack drives (up to 8 drives
supported) B D   Telex (also 2314/2319-type) removable pack drives (up to 8
drives supported) B D   CDC 60Mb "MIDS" removable disk drive (up to 3 drives
supported)

DOS 1.x  --  2200 instruction set 8-16Kb 
DOS.A    --  2200 instruction set, 2.5Mb drives 
DOS.B    --  2200 instruction set, 10 and 25 Mb disk drives (supported MIDS drives too?? no idea).
DOS.C    --  2200 instruction set, 1 to 4 8" floppy drives
DOS.D    --  5500 instruction set, 10-25-60Mb disk drives
DOS.E    --  5500 instruction set, 2.5Mb drives
DOS.F    --  5500 instruction set, 1 to 4 floppy drives (DOS.F was never publicly released)

There were a variety of variants of the basic 2200/5500/6600 processors.
Notable were the cassetteless versions:
1100         2200 instruction set, no cassettes, uses ROM, boots from floppy drive
1150         5500 instruction set, no cassettes, uses ROM, boots from floppy drive
1170         6600 instruction set, no cassettes, uses ROM, boots from floppy drive
6010         6600 instruction set, "64K" RAM, no cassettes, uses advanced ROM, boots from ARCnet
6020         6600 instruction set, "128K" RAM, no cassettes, used advanced ROM, boots from ARCnet

Price of the 2200 varied depending on Version 1 or Version 2, and then one added
memory cards.  A Type 2 2200 with 16K bytes of memory, as an example, sold for a
little over $14,000.  The 2200-350 disk system (desk-mounted controller with one
Diablo 2.5Mb disk drive) originally listed at $9800 and an addon 2.5Mb drive (in
a floorstanding "tub") was about $8500 more.  These reflect 1974 (US!) prices.
(I'm sure I still have a lot of other old price sheets around here).


Now let's advance to the "additional information" pages...

Adrian Barnes... While there were a LOT of 2200s shipped, the 5500 and 6600
later substantially surpassed those, and (probably due to the success of the ARC
System, and the larger disk systems) those two processor systems were the
company's "workhorse machines" which most users eventually adopted.

The boot is always done from the REAR tape deck, and not the front one.  Data
tapes could occupy either drive but commonly would use the front deck, simply
because it was easier to reach and the rear drive usually held a boot tape.

The 1500 system was a quite different beast internally, and used a Zilog Z80
microprocessor (whose instruction set and I/O architecture were significantly
different from the other Datapoint CPUs).  It was intended as a low-cost,
entry-level system or for field offices, etc., and was remarkable because it was
a complete, general-purpose, diskette-based system with lots of processing and
communications possibilities at an entry-level price of just $6K.

The 1800 system (and the similar, diskette-less 3800) used a different
processor, designed to be a lower-cost (if significantly slower) machine.  The
3800 was designed to be a cheap[er], processor-based single-user workstation to
hang off ARC Systems.  The 1800 was essentially a 3800 with one or two external
floppies (and in fact also offered some other cartridge disk hard drives a
little later on).

Barney Fleetwood While I appreciate Barney's comments, his view was a remote
one.  The main reasons why we kept the company based on proprietary stuff was
simply because NOBODY ELSE MADE ANYTHING that was really very suitable.  When we
first built the ARC System, we were in what was fixing to become a serious fight
with SyFA/Computer Automation for our higher-end customers, and their machines
were faster than ours at the hardware level.  The ARC System was the
"unexpected" "disruptive technology" that hit SyFA like a ton of bricks... and
by the time they realized what a significant product the ARC System truly was to
our market, it was too late for them to seriously try to respond.

There were other reasons we didn't want to release low-level implementation
details.  One of the big ones was because in other areas, people (who had NO
business being concerned about really-low-level details) would argue endlessly
about ultimately pointless details like register architectures, machine
instruction sets, etc., when they in fact would NEVER EVER program the machine
at anything remotely approaching that level.  Likewise, they would get caught up
in low-level academic trivialities like "does CSMA/CD really work?" rather than
dealing with bigger, more important, higher level issues like maintainability,
extensibility, diagnosing and isolating faults, file and record locking
mechanisms, conceptualization of remote resources, password and remote access
security, and stuff like that.  We decided that the best way to eliminate all
kinds of ultimately pointless debate about the internal mechanisms was to simply
state that ARCnet "worked by a miracle" and let it go at that.  :-)

Another reason for keeping stuff proprietary was because Datapoint had always
'given away' all its software for little more than the price of the distribution
media.  (Some tentative company steps towards per-system software licensing
were, like SyFA, largely derailed by the ARC System, where suddenly where
software was installed became largely irrelevant).  Obviously, the development
cost of the software was built into the cost of the hardware (and that's NOT a
bad system, since it means that everyone gets to use the software, and it makes
the hardware more compelling).  A "cherry picker" company who designed and built
compatible hardware could have undercut our hardware prices, but that wouldn't
have supported ongoing software development... it's not a plan with much of a
future.

As far as "open architecture", you need to realize that there was NOBODY even
prepared to TALK about the stuff we were doing, and waiting for discussions and
industry consensus would have taken YEARS and probably resulted in a product
that (1) wouldn't have been suited to OUR equipment or customers, and (2) would
have been less efficient and innovative.  Now, I do agree that it would have
been (in retrospect) interesting to have implemented a version of the ARC System
that ran on top of Microsoft MS-DOS (and/or IBM PC-DOS), and would have probably
allowed Datapoint to usurp the growth of Novell which finally did the same
thing, Datapoint was still hoping to migrate their customer base to
still-more-evolved LANs incorporating revolutionary stuff (for the time) like
laser printers, interbuilding LightLinks (high-speed optical LAN interconnects),
and such.  Trying to support PC-class hardware would have diverted key
development personnel off to other projects just when they were needed on more
advanced stuff.

The ARC System packet protocols were very much keyed to the features integrally
provided by ARCnet, and ARCnet's feature set was quite different than the "dumb
network" transport envisioned by the then-popular OSI model.  Fault recovery,
for one thing, was VERY VERY different, and quite inconsistent with the OSI
approach.  The ARC System's network architecture also used a very unusual model
for distributing resources and computing throughout the network, including the
distribution of rather low-level operating system functions... this was NOT a
model designed for heterogeneous computing environments.  While there are other
approaches possible (and we've seen them move to the fore over the years since
then) I still consider that the approaches I used in The ARC System were the
right choices (for a whole variety of reasons) at the time.

Security reasons were another reason to not publicize the internal network
protocols.  While a lot of effort had been made to make the ARC System secure,
there was no valid and compelling reason to hand potential invaders a map,
either.  It simply made more sense to force them to start from further back if
they were going to try to force their way in.

The "diverted product" issue probably ought to be explained.  Salesmen were
urged to meet their sales quotas quarter-by-quarter, and the company was running
significant order backlogs often reaching six to ten months or more. Salesmen
urged customers to sign "tentative" orders, giving them credit for the "sale"
and knowing that before the factory was ready to ship it, they'd have a customer
for the equipment... either the original customer, or someone else who came in a
little later.  In (rare) cases, they'd have the equipment delivered to a
"self-storage" miniwarehouse if there was a gap of a few weeks between the time
the equipment was ready for delivery (to an original purchaser who changed their
mind somehow) and the time they'd found another eager customer for it. This
worked for the salespeople as long as there were big production order backlogs,
but as the factory ramped up production to meet the great demand they reached
the point where salespeople were not able to place the gear as fast as the
factory was shipping it, and that's when the "scandal" suddenly showed its head.
 This was NOT a corporate headquarters malfeasance... it was salespeople trying
to scamper to satisfy their customers, and be able to deliver customer equipment
when those customers needed it.  The company got slammed for something they
weren't really guilty of.

When arb Asher Edelman took over the company, he expected it was golden (and it
still was, sorta) but when he found that the treasure trove of rapid cash he
expected to find hidden inside wasn't there, he realized that instead of just
parting out and flipping the company, he'd have to actually run it (and that
wasn't his forte).  Worse, his personality really wasn't compatible with
Datapoint's San-Antonio-style entrepreneurial/innovative corporate culture... I
actually met with Asher at one point and hoped I could engage him, but sadly
there was no charisma or favorable personal chemistry there at all.

Peter Theune By the late 1980's ARCnet had become (for a time) THE standard
networking protocol used by Novell PC-based LANs (Novell called it RX-NET but it
was exactly the same thing).  Ethernet finally did get released (and in a 10Mbit
version instead of the original 2Mbit version that Xerox and Bob Metcalf had
developed) but it went through a dismal thick-wire version (requiring ugly and
inconvenient 'vampire taps'), than a not-much-nicer "Cheapernet" coax-type
version (linear bus, where you had to take the whole network down to add a new
node anywhere other than at the ends, and where it was a nightmare to diagnose
and isolate faults... and ANY point on the network could take the whole mess
down with something no more complicated than a paper clip) before Ethernet
finally realized the inherent "rightness" of ARCnet's "interconnected stars"
cabling topology based on active hubs.  Once they adopted that key component of
the ARCnet design, Ethernet finally became (relatively) practical... even though
ARCnet still used its bandwidth far more efficiently and effectively than
Ethernet did.  (ARCnet systems typically got about twice the effective
throughput for a given bit rate than Ethernet did, and didn't collapse under
heavy loading situations).

The "small" CRT on the CPUs was an issue on 2200/5500/6600 CPUs but usually
those were used as relatively dedicated systems... Datashare central systems or
ARC System file servers, and for THAT use the 12-line screen was not a big
issue.  Most user apps ran on Datashare terminals (24x80) or 3800-class machines
(also 24x80).

I don't know what Peter was talking about with the "somewhat cumbersome"
networking tools.  In most ways, the tools offered by Datapoint for maintaining
and operating the network were more flexible and more powerful than those
offered by just about anyone in the networking business at the time.

I know that some users were using COBOL, but that was never a core product at
Datapoint.  It was provided (like BASIC) for those customers who wanted to (or
insisted on) using it, but the great majority of our key customers were
Datashare users, and that was the software applications environment that nearly
everything was designed around.

Geoffrey Vasiliauskas The 2200 screen was 12 lines by 80 characters... a very
"normal" number at the time, especially since the industry-standard IBM punched
card (which was still king back then) was exactly the same capacity. The speaker
could make a CLICK in addition to a BEEP.  One could CLICK repeatedly to make
beeps of other frequencies, under software control. If Geoffrey had four
floppies, he didn't probably have a 2200 (ALL real "2200s" had cassette decks). 
He probably had an 1100.

The white disk cartridges he's referring to are either the IBM-2315-type 2.5Mb
cartridge disks (slide in from the front) or top-loading 10Mb cartridges (the
one atop the 2200 on this site's main 2200 page is one of the 10Mb toploading
disk cartridges).  He must have a small motorcycle.  :-)

Geoffrey IS right that these machines were VERY well-built... they were quality
in a way that one rarely or never sees anymore.  The heatsink in the back of the
2200 (on which the switching power supply was built) was heavy-duty CAST
aluminum.  The whole system (once the cover was popped off) hinged up for easy
access to the individual circuit cards.  The cassette drives were assembled by
Datapoint onto a heavy, machined, aluminum baseplate.

Datapoint did not invent virtual memory, but Datashare did use virtual memory to
astonishingly good effect.  Under Datashare, there were separate code and data
spaces.  Compiled code was never modified (so never had to be paged back to
disk) and was exceedingly compact... a typical high-level instruction might
produce only three or five bytes of pseudocode.  Data space WAS resident in main
memory, where it was always available without paging.  During the time that
screen or keyboard I/O was being done (often MUCH of the time), there was often
NO user application code at all in the system, other than the list of data items
being displayed or entered.  When paging IN pseudocode, it was interpreted
DIRECTLY out of the disk controller's speed buffers, eliminating the need (on
2200 systems for example) to bring in more than just the few pseudoinstructions
actually needed from that sector.  (Datapoint disk drives used 256-byte disk
sectors).

Geoffrey is wrong about the size of the computer unit.  The "CRT" was in fact
the WHOLE computer, power supply, display, and the rest.  The desk contained the
disk cartridge drive, disk controller, and the power supply for those parts.  I
realize that a lot of people were confused about what was where, and by the time
the 6600 was coming to the end of its life as a product some folks within
Datapoint argued that keeping everything in the "CRT" was no longer meaningful
to our customers.  The result of eliminating that constraint was the (much
bigger) 8800 CPU and that had serious implications to the factory regarding
testability of the large circuit boards which resulted.

The power supplies in the desk or tub powered the disk units and controllers,
but not the CRTs or printers (except the Diablo-based daisywheel "servo
printer").  Other printers always contained their own internal power supplies.

The CPU unit had enough power in addition to internal needs to supply several
communications interfaces but large and heavily configured systems could exceed
the internal power capacity and so many external devices (such as RIM network
interfaces) used auxiliary power supplies within the interface device involved.

Programming languages available for Datapoint systems included seven versions of
cassette Databus, disk and cassette assemblers (including a relocating assembler
and linkage editor), BASIC, RPG II, COBOL, and a number of versions of the
company's workhorse applications programming language, Datashare.

The cassette-based OS "CTOS" was amazing (given the system it ran on) but
ultimately few users were content to forego hard drives (or at least floppy
diskettes).  Cassettes in the later years were mostly used for software
distribution and as a robust and portable data transfer medium.

Craig Brown The 8600 never supported cassettes.  Cassettes were only supported
on the 2200, 5500, and 6600 models.  I don't recall if there was ever a model of
the 1100 which had cassette drives (although it wouldn't have been hard to have
offered that, since in virtually all respects the 1100 was just a rebadged
2200).

Most of those systems almost never ran CTOS... the disk systems were far more
useful as products.  CTOS was used for DOS generation cassettes when setting up
a hard drive from scratch, and other diagnostic software, although in later
years the packs were usually built and software installed onto the hard drives
at the factory.  LGO (load-and-go) cassettes, where the application usurped the
CTOS itself, saved time and were often more popular for users.

The 1500 (Z-80 based) was not at all a "dumb terminal" although it was always
something of the odd man out since it didn't run quite the same instruction set
as most other Datapoint systems.

The Freedom Printer was not a daisywheel printer.  The Freedom Printer was a
dot-matrix printer (and basically the only printer that Datapoint ever built
from the ground up).  Datapoint offered a variety of daisywheel printers,
however, and the most popular used the Diablo "servo printer" mechanism.

The 3600 was a dumb terminal, designed to be a cheaper replacement for the 3300
and 3360 dumb terminals.  It was the first Datapoint "dumb terminal" with a
direct-view CRT (i.e. without the fancier tinted plexiglass over the screen). 
That terminal, at least, was hardly "a huge step above the old 2200".  At a list
price of under $2K, it was a very cost-effective terminal at the time.

The systems where the "monitor, keyboard and disk drive were built into a desk"
still used the complete computer system which was contained in the "CRT".  The
desk provided space for the disk drive and controller, and (on the back, and in
the tub) power strips and mounting locations for other communications, multiport
serial, LAN, and other interfaces as well as space to dress the I/O cables more
neatly.

Datapoint for a time was a "Fortune 500" company with (for a while) more than
10,000 employees worldwide.  The company was THE early leader in local area
networking.

The "two" raiders Craig refers to was just one... Asher Edelman.

Datapoint had built (believing that IBM-quality customer service was part of the
key to IBM's success) a VERY good (but expensive) customer service and hardware
maintenance operation.  While exceedingly good (and this gave customers
confidence when using Datapoint equipment for mission-critical applications) it
wasn't cheap, and this combined with the fact that Datapoint leased much of its
equipment in use by its customers meant that customers willing to forego the
"expensive" maintenance contracts were motivated to replace the Datapoint gear
with cheaper PC-class hardware (which, even if not provided with similar service
capabilities, at least could usually be replaced if necessary with a 'cheap' new
unit).  Instead of trying to cut costs within the maintenance operation, the
decision was made to treat it as a profit center and "cash cow" and eventually
to spin it off as a separate company.

Mark Massey The 4630 was indeed a 6600 series processor with 10/10 (10Mb fixed,
10Mb removable) Wangco disk cartridge drives.  Most of these systems used boot
ROMs to allow booting from the hard drive, and many no longer had the cassette
drives (which were costly to manufacture).

A 4630 could in fact run up to 24 terminals at a time, even if they were running
24 separate application programs.  It would run really amazingly well (presuming
well-written code) for such uses.  But Mark is right... Datashare was an
astonishingly effective and powerful tool, and VERY well-matched to the
Datapoint hardware.  The language is still being used on an everyday basis
(although using PC-based implementations, and generally not with 'dumb' serial
terminals) today at no small number of business customers... and there are still
applications areas where its features are even today "the best solution".  I
still support several of my consulting customers happily running such
Databus-based applications systems.

Joe Dumsha. Asher Edelman took over Datapoint, and one part of what he did to
the company was to spin off Intelogic Trace (although I think he remained part
of the team controlling Intelogic Trace, too).

"Opening their architecture" is always easier said than done.  Apple Computer
has flirted with the idea several times and has retreated each time.  Part of
the reason is that there are simply many innovative things you can do when you
make BOTH the hardware AND the software that are difficult or practically
impossible to do when you do only one side... the ARC System is a good example.

Kobi Haron I don't believe that the 5500 ever supported 512K of main memory...
56K is more like it.  :-)  The 5500 and 6600 still used a 16-bit address space,
although one could map a larger physical memory within the smaller address
space.  (Even if Kobi had a 512K system cobbled together on a homebrew basis,
I'd be surprised if it was still on a 5500 instead of a 6600).

It's true that Databus didn't support arrays, although it did offer a workaround
in terms of a list of variables where one could load or store a value into a
variable from the list selected by a numeric subscript.  While it's true that
this isn't QUITE an array, it also offered some unique properties... you didn't
HAVE to have the same sizes for the elements of the data list, or even for that
matter the same type.  You could remap the individual elements in different
lists and orders in different places in the program, if desired.  Like many
things in the language, it was DIFFERENT, and programmers who were willing to
think "out of the box" and develop innovative approaches based on the unique
properties of the language were usually the happiest Datapoint programmers.  :-)

The need to roll out to sort or compile could be largely mitigated by using
LAN-attached processors or (for non-networked systems) my virtual-machine-like
Partition Supervisor, which allowed running a DOS partition concurrently with a
Datashare partition.  But as Kobi points out, most users finally decided that it
was simply easier and more flexible to just add another processor to the LAN...
an approach that more and more users have adopted as LANs have become (as they
have today) ubiquitous.