Hacker Newsnew | past | comments | ask | show | jobs | submitlogin

> it was firewalled and available only to large corporations and those with personal relations to big tech execs.

The previous call-out to IBM seems relevant: before PCs, this exact statement would've been true for (mini)computers and mainframes.



> > it was firewalled and available only to large corporations and those with personal relations to big tech execs.

> The previous call-out to IBM seems relevant: before PCs, this exact statement would've been true for (mini)computers and mainframes.

Pre-PCs, IBM mainframes actually were quite open – up until the mid-1970s, IBM released its mainframe operating systems into the public domain. On the software side, the IBM S/360 was actually a lot more open than the IBM PC was – OS/360 was public domain with publicly available source code and even design documents (logic manuals), PC-DOS was copyrighted proprietary software whose source code and design documents were only publicly released decades after it had ceased to be commercially relevant.

As we move through the 1970s, IBM became less and less open. The core OS remained in the public domain, but new features were increasingly only available in copyrighted add-ons – but IBM still shipped its customers source code, design documents, etc, for those add-ons. Finally, in 1983, IBM announced that the public domain core was being replaced by a new copyrighted version, for which it would withhold source code access from customers ("object code only", or "OCO" for short).

The main way in which IBM mainframes in the 1950s-1970s were "firewalled" was simply by being fiendishly expensive – most people's houses cost significantly less.

It is true that IBM did engage in anti-competitive business practices, but those were primarily non-technological in nature – contractual terms, pricing, etc – the kind of techniques which Thomas J. Watson Sr had mastered as an NCR sales executive in the lead-up to World War I. In fact, a big contributor to IBM becoming "less open" was the US Justice Department's 1969 anti-trust lawsuit, which led to IBM unbundling software and services from hardware–and its software culture became progressively more closed as software came to be seen as a product in its own right.


> The main way in which IBM mainframes in the 1950s-1970s were "firewalled" was simply by being fiendishly expensive – most people's houses cost significantly less.

This was the primary aspect I was referring to, in the same way that training a ChatGPT-like NN can be (or could become) prohibitively expensive.

But your comments about openness are relevant on an entirely different axis.


> This was the primary aspect I was referring to, in the same way that training a ChatGPT-like NN can be (or could become) prohibitively expensive.

It is fundamentally different though - let's say it costs US$5 million to train a ChatGPT-like system. Someone only has to pay that once, and open source the results, and then everyone else gets it for free. US$5 million is a lot of money for the average person, but a drop in the bucket as far as corporations/governments/universities/research labs/etc go. By contrast, IBM's 1964 S/360 announcement priced the top-of-the-line model at US$5.5 million – which is over US$50 million in today's money – and that only bought you one mainframe, a second one would cost about as much as the first. A mainframe is hardware, but ChatGPT is software. ChatGPT's runtime (post-training) hardware requirements are hefty, but (on a per user basis) still cost less than a car does.


The problem is that AI research is moving incredibly fast. You might train a LLN today for $5M but a year from now the competition will have implemented an absolutely killer feature that needs $10M worth of training


AI research isn't particularly expensive. US$10 million to train a new model? Other fields have R&D budgets measured in the billions. I bet if you were a senior researcher at OpenAI, and you decided to quit and start a competing firm, there'd be a whole line of investors wanting to give you a lot more than US$10 million.

And you don't need to be coming first in the technology race to make money. A lot of people would be willing to pay for something ChatGPT-level with less restrictions on use. And then next year OpenAI will come out with something even more advanced, and they'll ask themselves "do I want a 2023-level solution which I'm free to use as I like, or a 2024-level solution with all these strings attached?", and many of them will decide the former is superior to the latter.

Maybe GPT-10 will cost US$10 billion to train? Anything could happen. Even if it does, the US government will ban China from using it, and then Beijing will spend US$10 billion to clone it. Even 10 billion isn't that much money if we are talking about nation-states pursuing their national interests, like not being left behind in the AI arms race. And then maybe China will outcompete OpenAI by offering an equivalent product but with far less limitations on how you use it.


Agree completely. The state of the art is probably going to always be closed and proprietary, but, especially with hardware becoming more and more powerful, training a custom model is not going to be beyond the budget and capabilities of even small organizations.




Guidelines | FAQ | Lists | API | Security | Legal | Apply to YC | Contact

Search: