Hacker Newsnew | past | comments | ask | show | jobs | submitlogin
This Time It's Different (justus.pw)
38 points by justusw on March 19, 2023 | hide | past | favorite | 130 comments


What puzzles me about all the skeptics (deniers, ever) is that they don't seem to be too forward thinking. Sure, these LLMs will not replace anyone right now, but damn look at the progression. Going back 5-10 years, we're doing stuff was seemingly impossible back then. Just imagine what things will look like in another 10-15-20 years?

If you're a 14-15 year old computer enthusiast, planning on getting a BS or MS in CS, you'll likely first enter the workforce 7-10 years. The time you spend in University is a lifetime for certain fields of Machine Learning / AI, and who knows what entry jobs have been completely automated.

Personally, I think this will be the death of entry-level jobs.


I tend to agree, but I think an interesting counter example is the hard wall self driving seems to have hit. There was a moment when even the experts in the field seemed convinced that we were only a year or two out from true Level 5 driving. But now everyone seems to be much more uncertain and conservative in their forecasts of the advent of L5.

Perhaps we'll see similar difficulties with these LLMs. They are definitely not quite there yet.

I'm currently trying to get ChatGPT to write some Python code that formats SQL statements according to my idiosyncratic preferences and it keeps getting very close but never gets it exactly correct.


I see it as 10% being 90% of work, but it might actually be 99% or more. Yes good enough for many many cases, but not nearly there when you truly need things right.


It's because ten year ago, everybody already talked about AI making developer jobs etc. redundant within ten years. And now (or at least a year ago) we have more developers than ever. Somehow the breakthrough is always ten years away. Yes, the technology has improved immensely. I just have not seen any specific business models and product ideas that are thought out in detail, from customer input to final product.


I'm not really a skeptic about the potential of the technology, it's just that this sounds exactly the same as what a lot of people told me when I was a 14-15 year old computer enthusiast over 20 years ago. And what has happened since then? The amount any one person could achieve with code shot way up, but far from killing all the jobs, we just found more and more things that could be done cost effectively with computers. So it's hard for me to avoid seeing that as the base case here, having already seen it play out once. My confidence level in the prediction isn't super high or anything, but this is the starting point for me.

But I am also sad, because I do think the "making useful things with computers" work will (and has) become less focused on the parts of it that are fun for me. I like to step through debuggers and reason from first principles about what's going on in cpus and memory chips and on the network think about how to structure blocks of code to be comprehensible; all the old school "programming" stuff, that's what's fun for me. I get some joy in working with groups of people to identify and build useful things, but honestly not as much as I get in the direct work with code I understand and appreciate. And I do think more and more of the work will shift away from that raw programming work and toward supervising AI tools in creating useful things conceptualized by humans.


> I think this will be the death of entry-level jobs.

This had been said about everything AI related, but the evidence shows it is actually the "higher level", jobs which are being disrupted first.

It is a matter of incentives, and the human tendency to assign blame to people. Everyone can blame their manager, but entry-levels are not blamed by anyone, because they are not supervising anyone.

Technology is in the business of replacing in-betweens, middlemen, rent-seekers etc with systems that are less biased, more objective, more transparent (sometimes), less prone to corruption etc. That's why google became more trusted than webrings, and why people use yelp or google maps instead of asking the locals.

Therefore, there are more people who want to replace their supervisors/managers/superiors with AI than the converse, and this will drive more entry-level people to work for companies that offer more automated , more transparent and trustable management processes. Humans are a point of friction


I don't understand why the pessimists don't see the unlimited amount of work stretching out before us! Why they can't imagine the utopia we could all be working towards.

First there is going to be massive amounts of money being splashed around making sure everybody has access to an AI that they control. Both big companies, and on our computers at home. 10-20 years at least.

Then we need to actually do the work of automating all the boring stuff like accounting, taxes, insurance etc. (At lets face it, governments are very slow to change and it will be 50 years till we work out the policy let alone the tech implementing the policy)

Meanwhile, the tech folks need to turn the attention of the AI to saving our environment. We need massive scale carbon capture of some kind. We need renewable energy sources. AI is not going to just work it out and build it all itself. We need to tell it what we want.

We need drones to fly around our forests looking after the bunnies and the squirrels, building a balanced ecosystem.

Somebody needs to develop and AI that will monitor our own health, understand more deeply how our bodies work, how to navigate nanobots though our system zapping cancer cells and balancing our gut bacteria. It would be nice if AI could solve aging in the next 20 years. (better to reverse it in my case)

Somebody needs to develop and AI and robots that can feed us. Micromanage greenhouses and livestock. Farming has a long way to go before its sustainable and ethical.

There is massive amounts of work in just building better infrastructure around our cities. How to clean our water and get people from A to B as comfortably, cleanly, and efficiently as possible.

Then once all that is done we can look to the stars, build the arc ships, figure out how much dirt we need to take with us to sustain us.

All the while we need to be entertained with stories, games, music and more!


<scifi hat>Yes you might one day become a highly compensated technician paid by a billionaire to direct his army of LLMs. You'll be the IR (inhuman relations) director. It will be interesting to see how deep corruption gets and whether billionaires will start running their own energy generation and data-centers - but they'll always need an audio/visual nerd to build, rack and repair servers, and a programming nerd to navigate screens and load the code. </hat>


You are exactly right; people (strangely here) seem to me hammering on it is not good enough or means anything now. We are talking 10-20 years… if chatgpt ‘22 didn’t happen, I would’ve expected big changes more like 50 years away ; now I believe it’s 10 or less. All the ‘this means nothing’ is so naive and frankly weird for normally forward thinking people here.


But The Thing that is going to cost us all our jobs has always been 10-20-50 years away.


Well, I see it costing jobs today. It might not accelerate, even if the tech stays like this (which it won't, even if we 'only scale').


What about you trying to be backwards thinking ? 5-10 years ago we had a boom in VR and AR, google glasses, snap glasses, boom in facial recognition. If you were forward thinking back then you probably thought 2023 is like Ready Player One. We still don't have flying cars and we are not going to have them.


The progress came relatively sudden. It may also stop relatively suddenly. At this point we can’t really tell.


This can only be the death of entry level jobs if you also expect all levels of jobs to disappear within the next few decades. Experienced workers don't just pop into existence.


One (perhaps silly) way I like to frame this stuff is by imagining that I'm some special purpose Turing machine designed specifically for some task. Sure, sometimes other Turing machines come along that appear to infringe upon my skill set but they ultimately only perform a small subtask better than I am able to (e.g. calculator, spell check, word processor, IDE, code completion, ...). So, I incorporate it into my routine, effectively boosting my own performance.

Now, what would happen if all of a sudden a universal Turing machine came along? Well, by virtue of being universal, that means that it can emulate me and all other Turing machines. This time around things are different. Even if I can find a way to incorporate it into my workflow, it can still emulate that more sophisticated version of me by virtue of being universal. So it then comes down to whether or not I can incorporate the latest version of this universal Turing machine faster than its own design is improved. If not, I will be replaced. Since in our instantiation, I am made from biological material it's in my mind only a matter of time before the universal Turing machine starts outpacing me.

So, I guess the question is then if these GPT models (or their descendants) are universal (in my hand wavy definition of the term).


You are misunderstanding / misusing "universal Turing machine" ... Computers have have been universal Turing machines aka Turing complete approximately since computers were invented.

You seem to be confusing UTM with Artificial General Intelligence. Universal Turing Machine is not the term for some magic machine that can interpret and integrate any observed computation. LLMs will significantly change how we interact with computers, but the ability to emulate another turing machine has always been there (for computers and yes, LLMs with memory are turning complete). That doesn't mean AGI can be implemented efficiently or that LLMs are sufficient for AGI.


No I'm just it as an analogy. Not all Turing machines are universal.

What were going through now could maybe be likened to what it would be like for a Turing machine to encounter a universal Turing machine for the first time. For all its life this fictitious Turing machine has encountered other non-universal Turing machines and have simply incorporated them into their own process. When they then encounter their first universal Turing machine they would possibly not be too concerned since each time before they have always just been able to use the new machine to make themselves more productive. However, this time it's different.

My point is just that while it may very well have been true in all of history that new tools have just made us more productive than before rather than fully replace us, this won't be the case for AGI. It's not just another tool we can add to our arsenal but instead something than can subsume us entirely much like how a universal Turing machine can emulate any other Turing machine.


I'm about to step out, and I will have to write my own essay in reply, but frankly, this time is different.

- OP's family of arguments, which I'll call BAU (Business as Usual, i.e. the claim that there is nothing fundamentally different about this disruption) depends on historical induction

- Historical induction is unreliable

- Sometimes things really are different, for example, the discovery of germ theory, or the invention of nuclear weapons

- The example given, e.g. farrier, is nothing like the present situation

- The fundamental difference between the coming disruption and previous disruptions is the scale. (Just as the difference between TNT and nukes was, again, scale.) Scale matters. Differences in quantity become differences in quality.

- By my read, transformer-based AI obviates the need for most cognitive work.

- That will upend the 'merit' part of our supposed meritocracy. We'll either have to become egalitarians (unlikely anytime soon, esp in USA) or we'll fall back on some other, worse metric for deciding who serves and who eats at the restaurant of life.

- I'd put my money on a resurgence in terrible ideas from the past, because they are so hot right now. Stuff like racism, title, caste, what-have-you.

- All of the abovegoing is Bad, and we should feel bad, because things are about to get bad.

- A better way to model this is as a reduction in habitat -- whereas the introduction of the ICE increased 'habitat' for minds desiring useful employment (engineer, what-have-you) while marginalizing a profession or two (farrier), the introduction of GPT seems poised to reduce habitat at a scale we have not seen before, and the 'new, better jobs' that Sam Altman alluded to, for example, seem beyond naming. Like, what is there left to do? Think it through. Where is your mind going to go? Knitting?

- Again, proper essay forthcoming; first, brunch


I, the author of the linked article, see the appearance of ever more productive tools as allowing humanity to break through the lie: that much of the labour we perform is necessary and that labour at all is necessary. The concept of your status in society being tied to your job and the idea of full employment for all is what is holding us back to achieve personal and societal fulfillment.

I ask you this: is everyone contributing to a copyleft project ultimately just aiming for financial gain, or are there also true idealists who do it for the sake of doing it?

Doing things is one thing and earning money is another and we come closer and closer to decoupling those too.

Why not reap the benefits and free humanity from the yoke of labour once and for all? If one person does the work of ten (i.e., thanks to the loom, steam engine, or LLM), and we naively assume that the value created is that of ten workers, why can the remaining nine not share the harvest as well? Or, we could all work, just significantly less (a tenth each) and allow everyone to go to bed with a full belly.

No one can predict whether it will be business as usual or not, for anything. Not for the internet, the transistor or LLMs. But we should not hesitate and call out the thumb twiddling lie that is employment through economic coercion.


Hi Justusw!

No disagreement here -- wage slavery is just that -- and, in utopia, we would have the robots do everything. But, as the error message goes, "you cannot get to there from here."

It's like seeing an ideal endgame config on a chess board but realizing that there's no combination of moves that will get your knight into position in time.

Delicious pie, very much in the sky, and any attempt to get there looks like it involves mass surprise unemployment, which, as a general rule, tends to destabilize.

Moreover, none of this actually affects or interacts with your original claim, that (more or less) the coming disruption will be similar to previous, smaller disruptions, for values of 'similar' that allow one to compare outcomes.

Again, the better model is habitat loss. ICE and other inventions increased this habitat; AI seems poised to reduce it sharply.


> I'd put my money on a resurgence in terrible ideas from the past, because they are so hot right now. Stuff like racism, title, caste, what-have-you.

The current group of AI ethics people are busy divining techniques to allow the model to gaslight users in the service of their employers. It is inevitable that these techniques will be applied to new areas and on models that have not been tainted by any other fine tuning.


It definitely feels different.

I saw the output of a GPT4 code assistant the other day and my immediate reaction was, "well, my career as a programmer is over." I can still do valuable things (gosh I sure hope so!) but the stuff I've been doing for the last twenty years or so is over. And good riddance! Software is buggy crap. The machines will do a better job.

The main issues are:

Who gets to decide the boundaries of publicly acceptable thought?

Who gets to reap the economic windfall?

How do we educate ourselves in a world that contains talking machines that can answer any (permitted) question?


I don't think programming careers are/will be over.

1. The latest model is nowhere close to taking in the number of lines of code in internal projects, so will be difficult to understand the design of those systems.

2. Companies developing software would be wary of sending internal code and trade secrets to the ChatGPT servers.

3. Languages, APIs, protocols, etc. evolve over time, so ChatGPT would need to keep up and handle the specific versions you are using internally. For example, Java POJOs vs record classes. Or even internal limitations like lack of runtime type information for things like embedded devices.

4. Experiments I've seen relied on external tests being in place to check the validity of the output (e.g. implementing the Promise JavaScript API), and the output had test failures that ChatGPT wasn't able to fix when told about them. -- I'd expect ChatGPT to get better at this specific example, by being fed these uses of ChatGPT in the training data, but I don't expect it to do better when shown novel specifications/requirements.

----

There are various design decisions that go into creating software that often have different trade offs. Like when implementing a compiler, you can stop on the first error, but developing a language plugin for an IDE you need to be able to recover from and handle incorrect or incomplete input.

There are also things like the way you structure the code, like creating DAO/POJO/etc. wrappers for database/JSON/XML/etc. objects, or providing APIs that fit into the style of the language you are implementing them in.

It would be interesting to see how ChatGPT handles something like implementing a HTML parser when given the WHATWG specs, or a keyboard driver given the keyboard specs and the driver API docs.


FWIW, I think your first two points as well as the fourth will be answered by next Tuesday. That is, exponential advancement is a hellofa thing. The "wow" of GPT4 isn't just that it can pass the bar exam or be a decent Dungeon Master, it's also how fast that happened after GPT3, eh?

Your third point, that "Languages, APIs, protocols, etc. evolve over time..." is, I think, going to be obviated once these systems start writing software. I think most human interfaces (programming languages are Human-Computer Interfaces) will be refactored out of the loop. Everything we thought programming was about is about to be swept away.


I'm not sure why you are downvoted. These are all great points.


There is a huge amount of work in making sure everybody has access to the technology, and that there are appropriate safeguards around its use.

Too early to retire I'm afraid!


Really difficult to agree with this given the pace at which LLMs are improving.

LLMs are disruptive in that they enable a form of outsourcing. Outsourcing to the the lowest-cost region in the world, inside a computer. Outsourcing to tireless, ever-improving, highly-intelligent machine workers. Workers that will eventually have a variety of specialized and/or general skills, depending on what they're trained for.

Imagine "offshoring" (AIshoring?) for 1% of the cost of a human employee to a machine with zero time off, zero time zone separation, zero cultural or communication barriers, and with 100% access to all of your corporate documentation, goals, and other context.

Imagine that these "offshore" AI workers only improve every year.

This time, it really is different.


Labor cost really isn't the factor in offshoring we think it is. Servers for AI are not free, and workers in India are really cheap, similarly to how minimum wage workers in the US are really cheap compared to robots that need constant care and maintenance. But that cost isn't even the main factor. The main cost is communication, and if you can actually get exactly what you want. That is the main challenge to offshoring, which settles many companies with technical debt where it isn't clear whether they benefit at the end at all. So if AI competes with offshoring, I don't know if that is a great value proposition.


It's extremely uncommon that everything changes suddenly and forever.

Most technological progress is a continuum, with little step functions of "this is it."

The AI/ML progress in the last ten years is a big local maximum, though, and that's enough to drastically change things for a lot of people.


> everything changes suddenly and forever.

Since the beginning of the industrial revolution, it's the prevalent mode of historical development, I'd hazard to say.

Things that lasted for millennia have been displaced in a few centuries, and eventually the delay changed to a few decades.

"Suddenly" is not overnight, "suddenly" is when your children will live in a really different world.


The price of things that AI can do well is quickly going to fall to the cost of the electricity to run the model. The price of things that AI can do, but can't do well, will fall some, but not nearly as much. The price of things that AI can't do at all will go up, because there will be fewer people working in knowledge jobs, and therefore they will be able to command higher wages.


> The price of things that AI can't do at all will go up, because there will be fewer people working in knowledge jobs, and therefore they will be able to command higher wages.

This doesn’t follow. If AI displaces workers the majority of them will find other work. AI will increase the labor supply, not decrease.


Look up the Baumol effect. When productivity rises in one sector, there is an increase in wages in other sectors.

The flip side of the coin is over time we spend more of our income on the things that don’t have increasing productivity. Think of housing, education, health care, etc


I suspect that the number of workers which AI can't displace due to their unique mental abilities has natural limitations, mostly not enforced by cultures and markets.

With global population growth declining, and negative population growth in most industrial countries, the supply of humans in general, and humans of exceptional abilities in particular, will only decrease.


> things that AI can't do at all

Like what?


Be liable for their actions.

Be an attractive, physical sex partner.

Perform neurosurgery without human assistance.

Raise children, from birth, autonomously.

(Once they can do that, I may be really worried.)


#remindme

But seriously, this list of things they would never ever be able to do would look a heck of a lot different if it were still 2016 :)

Also thinking about this just a bit more, do you know many humans that can accomplish this list?


Yes, that’s an issue, but this subthread was about who will get paid really well in the future.


Your first point is not really up to the AI, but a legal question, right?

And about the other points, just wait a bit. They don't sound actually so unrealistic to me, maybe in the next 20 years.

About the second point, I think there is already an industry specializing on this.


I would say everything in the "touch grass" world. Robotic hardware seems to be lagging way behind pure AI software. One obvious example is self driving cars, which seem to have plateaued in progress.

Or: I just listened to a podcast about how there is a huge shortage of electricians available to do the residential upgrades incentives by the IRA. GPT-4 seems to be nowhere near a solution to that kind of problem.


> Like what?

Are you under the impression that AI can do most things well? Because so far it cannot.


I was more pointing towards the at all part. I did not say that it can do most things well. But it's not really easy to find things which it cannot do at all. Sure, it might be quite bad on many of them. (But that might just be a matter of time or priority...)


Up until last month: draw physically reasonable hands.


Anything that’s not a glorified copy/paste.


Manage delivery of ERP.


take initiative


> What about the printing press or any other of our advances in writing, such as the typewriter, spellcheckers, emails, grammar checkers, and so on? While the work of a classical secretary became unnecessary in many ways, the fact that the individual increase in productivity meant an overall growth of the economy meant, in turn, that those who would have lost their current role or even employment to technological advancement would find new occupations within this growth.

Oh please. That is not at all what happened. Technology that makes people more productive largely does so by lowering the required skill to do something. A shop clerk used to be a pretty difficult job. Now it's done by people with mental disabilities.

The pay goes down. The work becomes less meaningful. We foster a system where the supply of humans willing to do menial labor tasks grows so that those with capital have access to uber drivers and other permanent servant class workers.

The growth of the economy is not a tide that lifts all ships.


> A shop clerk used to be a pretty difficult job. Now it's done by people with mental disabilities.

> The growth of the economy is not a tide that lifts all ships

Just a guess, but maybe the person with mental disabilities that are now gainfully employed might disagree.


Now they, too, may know the transcendental joy of laboring to create wealth for someone else.


True but at least they have a job and some economic freedom. Without that job they would be reliant on other people. I'm not sure I see the problem here actually.


As opposed to creating nothing and being completely reliant on the charity of others and the state ¯\_(ツ)_/¯


We shouldn't admonish the mentally disabled for working if they want to. It's good that they have opportunities. We should be critical of societies that state they basically have to work or will die because they're responsible for their own financial security.

The issue is not the mentally disabled though. The issue is that people are being pushed into the absolute bottom tier of the labor pool. The decreasing opportunity for valuable labor wages is the problem.


They may indeed. But when you look around and see your coworkers have mental disabilities, you may begin to wonder if they were pushed up or was the bar lowered and you were pushed down?


That sounds like a "you problem" to me.


Is the “you” the person who was previously a skilled laborer but is now an unskilled laborer?


Yes. Society improves (becomes more productive) and things that were once hard become trivial, it's called progress. Improve along with it or get left behind. We dont' live in a world where you can just go turn a screw on an assembly line for 10 hours a day for 30 years and get a pension anymore.


> The growth of the economy is not a tide that lifts all ships.

Aw, c'mon. Technological and economic improvements, in the long term (and with frequent setbacks) have driven the growth of human population and improvement of the human condition.


I would say that's generally soley the result of social reforms to try and mitigate the growing inequality. Conscious efforts to try and share wealth and offer protections to different degrees. Also, war. But the social status and loss of importance has been continuous.


This may be separate conversations.

I'm speaking of macro-level infant mortality and mass starvation.


Sure by that metric things are better


The author suggests GPT-4 is analogous to tools like the spellchecker, but really the roles get flipped. AI does the work, and we take the role of the spellchecker.

Not dissimilar to the one employee who hangs around the 10 self-checkouts ensuring they work properly. Can’t say I look forward to that career change.


That is… not the only thing a clerk does. Which is (more or less) the whole point in TFA.


Yes it is, they babysit the checkouts and step in when needed. Their role is very much different than the role of manning a single non self-checkout. And let's not get into how the roles of the 9 other clerks have changed who are no longer there (spoiler: doing no work there and not getting paid at all).


If you thing self-checkout reduces clerk work by a factor of 10 you have not worked in low margin high throughput retail. Anyway, from experience i can tell you it’s a gradual affair.


GPT4: The author expresses skepticism about the idea of an imminent Singularity, a point where artificial intelligence surpasses human intelligence. They argue that LLMs are more likely to be force multipliers, improving productivity and automating routine tasks, rather than replacing human workers.

I would say the problem is that we are multiplication something, and as a result, we don't know the outcome of the multiplication. Right now, the multiplication is only intended to improve productivity for some people. However, if this multiplication were to occur on a global scale or on a societal level, the true impact is unknown.


The one thing that surprises/scares me the most in the current state of affairs is not so much the technology itself (after all, technology needs to be absorbed by society, and that is surely a bottleneck here), but how fast the research is going. There is something about machine learning / AI research that makes it both faster to do than other types of research and also super motivating. The people who do ML research are basically just working 24/7 and doing that with a smile on their faces. Most other types of research require you to spend a lot of extra time doing boring, bureaucratic stuff (like handling human subjects, complex protocols, lab equipment, endless meetings, etc.) but in ML you just ideate, program, run the tests, write, publish, repeat. (Sure of course you need to do it right, but that is not the point here.)

And for those who are researchers at heart, that is the best thing in the world: to be able to do your research and push your results out as quickly and efficiently as possible. So nowadays a paper comes out with some new and interesting development, then two months later there are 30 other papers with significant improvements over that first one. (Yes there is a lot of junk, but again that's not the point here: the point is that those who are doing it right can do it efficiently) This is the most incredible thing about this whole situation, and maybe the most scary: there is no way to stop this avalanche of research, because it's not a centralized thing: it's just a bunch of human beings doing what they love, with motivation (both financial and personal). Nobody can stop this. If someone happens to press the doom button in the middle of this, well... that's it!


It’s entirely centralized because it requires massive training corpuses that plebs can’t come up with. The only serious advances are coming out of large corps.

The indie hacker movement around stable diffusion is interesting but those are interfaces to an existing model. Not so much new models themselves.


> knowledge workers

Aka Thinkers. I don't think the author considered the full extent of automating thinking. They are underestimating what it is. This technology can only get better and is probably already superhuman. This time it is different indeed and makes lot of knowledge work not just be obsolete, but also inferior.


It's probably already superhuman? That's... one interpretation of GPT. I interpret it quite differently.

Unless you're going to say that computers were already superhuman? They could do arithmetic far faster than any human ever could. Also pure logic. Chess and go, even. Were they superhuman then?

Because as I look at GPT, I don't see superhuman. I see a babbling idiot that babbles in correctly-constructed English sentences, but has no idea what it's talking about, and that manages to be factually correct... maybe 70% of the time? (That's hard to measure, because the universe of possible discourse is really large.) And, worst of all, GPT has no idea of when it doesn't know something, but will irresponsibly blurt out... something. Something random that its training set made it think is the least-implausible thing to say right there.

You know, if they could just fix that - if they could add some kind of weight so that it knows when it's getting into an area that it doesn't have adequate training data for, and program it to express uncertainty rather than certainty in something wrong - that one change would make it, still not superhuman, but at least much more useful.

And, if we could do that, maybe we could do one more step. If it has more than one strong possibility, but they are contradictory, then there's contradictory data in it's training set, and it's likely dealing with something controversial rather than certain.

I know, I know. Everything is easy to the one who doesn't have to do it. It's probably much harder than I said. But I think that's what GPT needs going forward.


> Something random that its training set made it think

That's the mistake right there. It didn't think. The text did.

Specifically, humans encoded thought into language, and encoded language into text. GPT modeled the patterns from that text, and used those patterns to restructure it.

Because language is the most dominant pattern, blindly following an inference model is very likely to result in correct language transformations. Because people don't write nonsense, the result of a correct language transformation is very likely to make sense.


You missed the point there. When the information isn't in the training corpus, then GPT still generates... something. It's not in the text, but GPT's weights show something as the best correlation. Sometimes it's mining connections that the authors didn't even realize were there. And sometimes, it's following random correlations that don't mean anything, and as a result it's generating nonsense.


It is in the text. It's not the same shape, location, or order because GPT followed a complex pattern to "generate" the continuation. None of the tokens are new.

> Sometimes it's mining connections that the authors didn't even realize were there.

Exactly. GPT is not limited to modeling patterns that align to language. It models every pattern it can find.

Text contains so much more than language: it contains powerful implicit patterns of behavior; and GPT proves that by modeling them.


But it will also confidently tell me who won the 2023 NCAA tournament. It did not find that in "implicit patterns of behavior". It made it up from patterns in the text, but the answer was not there. At best what was in the text was projections, opinions, guesses. GPT takes all that in, but it's not actual information to answer the question, because there is no answer - not yet.

So it finds a pattern, but a pattern that is not the answer. But it gives it as the answer, with full confidence.


> It did not find that in "implicit patterns of behavior". It made it up from patterns in the text

There is literally no difference between the two.

> At best what was in the text was projections, opinions, guesses.

And a pattern of response to the structure of your direct question,. It didn't subjectively decide what the right answer is: it followed the pattern closest to your prompt, and printed the tokens that pattern was made of.


I think that you are agreeing with the main point I'm trying to make, and you're even going further than me. You're just disagreeing with my wording. What you're describing is, it doesn't think, it doesn't understand, it just follows whatever patterns in its training corpus kind of correlate with patterns in the prompt. And I agree.

Where I may differ is how to regard the training corpus. True, it contains a massive amount of distilled thought, encoded in words. But it is also true that there are subjects on which the training corpus contains no distilled thought - and GPT will respond to questions on those subjects as well, finding some spurious correlation to respond with.


> I think that you are agreeing with the main point I'm trying to make, and you're even going further than me.

Pretty much.

> You're just disagreeing with my wording.

Yeah, that's the "going even further" part. Words are important. The importance of words is really foundational for a lot of what we are talking about.

> What you're describing is, it doesn't think, it doesn't understand, it just follows whatever patterns in its training corpus kind of correlate with patterns in the prompt. And I agree.

Cool. It's nice to be on the same page. Let's read the next one:

> Where I may differ is how to regard the training corpus. True, it contains a massive amount of distilled thought, encoded in words.

That's the biggest point I'm trying to make by "disagreeing with your wording". Words have significant implications. In this case, the implication is "problem domain": language vs. text. Language is a subdomain of written text. Written text is a subdomain of all possible text. GPT is dealing with written text.

GPT does not only model the language patterns that humans intentionally put into text. GPT models all of the patterns that exist in the text.

Humans, when writing, do "distill" thought into language; but the result is messy: we end up encoding more patterns than we intended to. Our behavior, our choices, our reasons, and the surrounding situation we exist in, all inform our writing. Those are the patterns that get encoded into the text, right alongside language.

GPT doesn't see any pattern as different from another; so it's misleading to say that it models "language". GPT doesn't categorize that specific set of patterns: it models all patterns equally.

We humans have a good idea of what to expect from only language patterns. If that were all that was modeled, GPT's output wouldn't be nearly as interesting to us.

GPT's output is interesting to us. If we look objectively at the model it generates, we can see why: there are patterns there we had no intention of writing; and GPT modeled them just like it modeled the language-specific patterns we expected it to.


> GPT does not only model the language patterns that humans intentionally put into text. GPT models all of the patterns that exist in the text.

> Humans, when writing, do "distill" thought into language; but the result is messy: we end up encoding more patterns than we intended to. Our behavior, our choices, our reasons, and the surrounding situation we exist in, all inform our writing. Those are the patterns that get encoded into the text, right alongside language.

Sometimes truly, and sometimes falsely. Take this paragraph, for instance. If I made a typo on the fifth word, and left out the "l", this paragraph would also encode something that is not at all what it is about. A human reader might chuckle at the typo, but would not be confused by the meaning. But for GPT, that will be a pattern in the text, just like all the others. So sometimes it can reveal what was unintended but is still informative, and sometimes it can be misleading.

But it's not just typos. There's very intentional text from advertisers, shill, liars, and propagandists. That is all patterns that get encoded, too. The lies are encoded right along with the truth.

But where I was really going initially was the hallucinations. If I ask GPT who won this year's NASCAR race on Tierra del Fuego, it will find something - some pattern that matches better than other patterns. Who knows, it might find some kid's fanfic posted on Reddit, and will tell me who won in the kid's story. But if there's nothing like that, it will find something. It won't be any kind of a meaningful pattern that it finds, but it will still present it, and will present it with as much certainty as it would if it found a number of news stories that all agreed with each other.

So I'm seeing in your posts a pattern encoded (perhaps falsely) that makes me think that you think that what the text encodes is gospel. It's not. It's much that is golden and much that is dross, and GIGO applies to GPT as much as any other information system.


> So I'm seeing in your posts a pattern encoded (perhaps falsely) that makes me think that you think that what the text encodes is gospel.

On the contrary. I think you have a pretty good description for how the behavior we want and the behavior we don't want exist as the same features. I am trying to restructure the narrative around LLMs to be about the patterns themselves, not what we hope the patterns to be.

We can throw around the word, "hallucinate", but that's just another anthropomorphization. Whether or not the pattern followed "makes sense" is a categorization humans attribute after the fact. GPT doesn't do "certainty". It does what it is given.


OK. I think I understand, and I think I agree. Thanks for the conversation.


You're welcome! I've been finding a certain satisfaction in trying to articulate this perspective.

It's challenging because I'm criticizing most of the narrative that exists in the subject, but that narrative also gives a clear illustration of where to go wrong.

I think better understanding the more mysterious patterns that exist in our writing might be really useful. Often, the first step in inventing something useful is to explore an unfamiliar domain and recognize what information we are missing.


> As an artificial intelligence, I do not have the ability to predict future events or access real-time information. My training data only goes up until 2021, so I don't have any information on the 2023 NCAA championship or any future events. I suggest checking with reliable news sources or following the NCAA updates for the latest information.

Not for me


I asked it to create a portmanteau insult for GPT doubter on hn and got GLTard or GPTSlother when asking for an animal one. I suspect you are not right.

It also advised against name calling

Edit: wrong in regards to no new tokens


> and that manages to be factually correct... maybe 70% of the time?

Do you update on this when considering that GPT-3 got in the bottom 10% in several exams, whereas GPT-4 is now in the top 10%?

Do you not see this clear improvement as evidence that it's just a matter of time?


Clear evidence of improvement? Sure.

Evidence that it's just a matter of time? Time will tell.

Evidence that it's already superhuman (the original claim)? Very much no. Only in the top 10% isn't superhuman.


Take a look at the GPT-4 paper, and the list of exams and accompanying scores. Then tell me if you know any humans that did similarly? Looks superhuman to me. Especially considering it took all of these in less than a year.


On the contrary, I think you are overestimating it; and I don't blame you for doing so.

The overwhelming majority of narratives surrounding LLMs have shared the same character flaw: personification.

Almost no one writes explicitly about what an LLM itself does. An LLM models patterns from text. That's it. Not a single thing more.

Personification is a great tool for explaining what you already know. The mistake is to draw conclusions about what you don't know from the personified character instead of the thing itself. Yet, that is exactly what is being done for LLMs.

The first mistake is in the name: "Language Learning Model". That's the intended purpose, yes, but it's a misleading description for what the tool actually does. I propose we instead call them, "Text Learning Models".

An LLM models all of the patterns in can find in the text it is given. It doesn't categorize those patterns, or have any notion of their importance to humans.

An LLM doesn't do a single explicit thing. Everything an LLM does is implicit. An LLM models the patterns in the text, and exhibits them. The patterns in the text do all the exciting stuff.

Everything interesting about LLMs is human generated. Without human authored text, the model is worthless. Without human authored prompts, the model is a black box. Every interesting thing an LLM does, one or more humans told it to do.


That's very misleading. The current large LLMs learn tons and tons of "reference maps" that map how words (tokens) are referring (attending) to other tokens within a certain context. In the process they learn a map that may be the equivalent of what we humans call "syntax", and they apply it to completely different contexts and inputs. It's not memorization, nor simple pattern matching, but something that allows them to be creative


The LLM does not know what a word or a context are. It doesn't have that kind of categorization in its model or anywhere else.

Tokens don't even match word boundaries: they are intentionally misaligned to better accommodate contractions.

> In the process they learn a map that may be the equivalent of what we humans call "syntax", and they apply it to completely different contexts and inputs.

Exactly. Alongside those patterns, they model all of the other patterns that are present in the text. There is no categorization between these patterns: the word "syntax" has no bearing on the model whatsoever, except where it literally exists in the text being modeled.

LLMs do not parse. They tokenize, but not to match any predefined grammar.

It's not the LLM itself "being creative". It's the content of the text: the patterns that are modeled. Those patterns exist independent of the LLM. They come from the complex human behavior that is, "writing".


I dont actually see the point of using human-derived notions of 'language', 'syntax' , 'know', 'category' and even 'thinking' or 'parsing' when making precise arguments. Those are all vague, subjective notions that we made for ourselves but nature does not need to fit to our mold.

The fact is, LLMs are the best quantitative model of the human communication code, so good that it can replicate it quite convincingly. Better than any model we had made before using our imprecise subjective notions, which were quite like alchemy. Whether it will prove to be 'thinking' or not, remains to be seen in the future.

> the word "syntax" has no bearing on the model whatsoever,

We don;t know that. One would have to look into the model weights for that and this analysis does not exist yet. If one can provide a decent definition of syntax, it may be found to reside somewhere in its weights. same for other subjective notions.


> I dont actually see the point of using human-derived notions of...'thinking'...

> Whether it will prove to be 'thinking' or not, remains to be seen in the future.

Maybe you should take your own advice.

You are really missing my point. Humans use language, syntax, thinking, etc. We know what those patterns are. LLMs do not. LLMs don't approach text with the same models that humans use: they model the text directly. Any pattern, whether it be a word or sentence or even a pattern that humans have no name for at all: the LLM models that pattern.

What we humans call "syntax", "word", "idea", etc.; LLMs don't. They don't have any names for any things. They just have the collection of things in front of them, and implicitly model that collection.

An LLM doesn't think objectively about its model, either. It only constructs continuations by modeling prompts. It is the content of the prompt itself that determines the continuation, not "thought".


LLM stands for Large Language model. Not Language Learning Model.

It absolutely categorizes those patterns as you can ask it for descriptions of things or request things via descriptions. It does identify the importance to humans. Because all of those things are also just patterns about patterns.

Everything an LLM does may be implicit, but it takes a tiny effort to engineer that implicit capability in to something explicit.


> LLM stands for Large Language model. Not Language Learning Model.

I'm not sure how I missed that. Good to know. My point still stands: it is the text being modeled, not language.

> It absolutely categorizes those patterns as you can ask it for descriptions of things or request things via descriptions

When you "ask it for descriptions", you are presenting a text prompt (that contains your question) to be merged into the LLM's model. After merging, the LLM presents a continuation. The categories you are talking about are patterns of text. Those patterns are part of the model, and the LLM doesn't know them any more than it knows any other pattern.

The categories I am talking about would be higher-level: a metadata structure of explicitly named patterns. That does not exist. LLMs don't have any notion of "what" a specific text pattern is, or why humans care.

> It does identify the importance to humans.

Again: not it. Humans make that identification. The identification itself, as written in text somewhere in the training corpus, is made available as a pattern to the LLM. The answer you get is a continuation from the content of the model, not some objective thought that the LLM is doing.

> Everything an LLM does may be implicit, but it takes a tiny effort to engineer that implicit capability in to something explicit.

Yes, and that engineering is a collection of human intentions. This changes nothing about the distinction I am making between an LLMs behavior, and the human behavior that was encoded into the text it is modeling.

Most of the engineering is to construct the training corpus itself. Adding "weights" is effectively the same; except the model is being edited directly. None of these efforts change the fundamental structure and behavior that the LLM itself is made of.


So you agree that the LLM has patterns

And you agree that all those things are patterns

But you are arguing that they’re not… doing the patterns explicitly enough?

Please. The bar you’re describing is already something we can half ass and something we will be able to do explicitly very soon. The LLM can write code. If we just let it delegate the execution of that code to something that isn’t an LLM you can generate your explicit ontologies very easily. imo this is so obviously near term it’s uninteresting to fuss over it not being here in the oresent


I'm talking about problem domain and approach.

"Language" is a subdomain of "written text", which is a subdomain of "all possible text permutations". LLMs model "written text".

Approach is either explicit or implicit. Constructed or inferred.

You may be familiar with parsing: that's the explicit approach. Parsers are made out of predetermined grammar patterns. They categorize text into a model (AST) that was predetermined by the grammar rules. This approach is essentially a function from "all possible permutations of text" to "a known language model". It also maps "written language" to "predictable machine instructions".

Parsing works for "context-free" (code) grammars, because the patterns of grammar are already known. Parsing fails at "context-dependent" (natural language) grammars because the patterns of grammar are ambiguous.

LLMs take the implicit approach: they start completely blind, and model every pattern they can find in the text. An LLM has no category for "language grammar pattern". It does not constrain itself to the domain of "language". This approach maps "the behavior of someone writing text" to "patterns".

The difference in problem domain introduces ambiguity: LLMs can't categorize truth from lie.

The difference in approach removes intentional behavior: LLMs don't translate text into predictable machine behavior. LLMs model the patterns of behavior from the training corpus text, then model some more from a prompt, then show the resulting pattern.

What an LLM does should not be expected to ever match what a parser does: they are completely different approaches working in completely different domains.


And I’m yawning because getting an LDM to construct explicit ontologies and then evaluate text against them seems trivial given it can code. It seems doubtful that such a hurdle will remain unsolved for more than a year. Like I said, you could do a poor man’s version of it today by instructing it on how to use a natural language API to call services.


AI has to be trained. So it’s good at doing things people already have done many times. It’s not awesome at design for novel things.

So, I may get an AI that’s like a new college grad level of coder.

As a senior, I’m mostly giving work to my team anyways then reviewing what I get back, so it would not be so different.

As a junior, you have to be scared you’re gonna be replaced, and question coming in industry.

So, less juniors will come in industry, which will make a shortage of seniors in the coming years, but they will be all the more needed to direct the AI.

If you’re experienced, you’re gonna be in amazing place during this transition phase. If you’re junior, there’s gonna be a huge hump.

When the current crop of seniors get old and retiree, there’s gonna be a shortage like none the industry has ever seen. So the smaller class of juniors that ride out the revolution are going to have it best of all.


The problem with AI such as GPT4 is training data. As it usage increases, most of its training data will be data generated by GPT4, hence creating a positive feedback loop. This would actually increase the value of human knowledge workers.


Entire farms of people dedicated to providing accurate training data. Like battery farms from the matrix but more boring since you'll need a masters degree to become a happy little data provider.

Joking, but training data isn't as big of a problem as you suggest. Since so much about AI is still very centralized it's relatively low effort for them to tag generations in a way that makes sense, avoiding a lot of effects of a loop.


It’s ok to train a model on model outputs so long as you have someone to curate them.


Nah; the model will improve to learn faster and more from less data. Or AI winter.


I'm betting on an AI winter before a fundamental change in the difficulty of computation.


Yep, that is a likely outcome.


This is the point where “it all” goes sideways. Not toward a grand and transformative singularity, but sideways into some variation of dystopian hellscape.

Being unable to recognize why this is all very clearly going to go wrong requires a great deal of ignorance or a wildly unrealistic faith in humanity, which is sort of the same thing.

There’s nothing necessarily wrong with ignorance or delusional faith in humanity, per se, but the people qualified to assess this seem almost universally negative regarding the most likely outcome.


So I've studied CS, have been professionally programming for 3 years. Considering the advances of AI, what (somewhat) programming related job would be the most safe / the best bet to go forward? I don't have a full picture. A few fields I see:

Specialized field as:

* FPGA programming

* AI itself

* Red teaming / blue teaming

I think these fields will have a tougher time:

* Web dev

* Game dev (they do now as well)


Putting aside fears of AGI for a moment, seeing the comments here I’m coming back to the same idea I come to every time new tech comes along: Complaints about AI and automation are actually complaints about capitalism. The increase in productivity from AI could in fact enrich our lives, but given the never ending hunger for growth from late-stage capitalism, the average person will see a drop in wages, harder jobs, and more inequality.

Can we break the cycle on this? Is there a way drive innovation while valuing humans?


> The increase in productivity from AI could in fact enrich our lives

If humans aren't the ones innovating anymore, what will be the point, for an AI, of supporting millions of people who serve no productive purpose?


Are you talking about capitalism or from a broader perspective?

If the latter, the "point" of people is not to be "productive". That's not how we measure mankind. Why would we program an AI to only consider human productivity (for a capitalist definition of it, too)?


We're not really programming the AI to do anything. It's essentially programming itself. I'm saying, from the perspective of an AI that relieves the human race of the need to be productive, what is its incentive for doing all that work for us?

To put it in an economic framework, this notion that humans shouldn't be measured by their productivity or creativity, but should have as much free time as possible, can also be achieved by enslaving other people. Slave owners had the same lifestyle as the imagined anticapitalist one. In this case, we'd be enslaving an AI. I'm asking, why would the AI want to be our slave?

Also worth noting, slaveholding may have enriched the slaveholders materially (i.e. in capital), but it was not "enriching" in the sense of ennobling or bettering of the spirit as I think the parent was implying could be an outcome if AI were to handle our productivity for us. I don't see why having any human slave or AI slave handle our productivity and creation would be (spiritually) enriching.

Perhaps that's because I measure human value by contribution - call it productivity or not. The opposite of productivity is idleness, whether you ask a communist or a capitalist. Only what's produced varies.


> I'm saying, from the perspective of an AI that relieves the human race of the need to be productive, what is its incentive for doing all that work for us?

This seems like begging the question: is the goal of AI to relieve us from being productive? I don't think so; I think it's to relieve us of menial repetitive work, or to assist us in difficult tasks.

In any case, this AGI would assist us within the parameters we set. Why would it decide to get rid of "unproductive" humans, unless we set the goal "maximize productivity at all costs", something no human would choose with all the discussion of paperclip maximizers?

> I don't see why having any human slave or AI slave handle our productivity and creation would be (spiritually) enriching.

This again seems like begging the question in two ways: first, an AGI wouldn't be a "slave" but a machine -- like a car or Microsoft Clippy -- and machines are not slaves (the problem with Blade Runner replicants, for example, is that they are not machines but sentient beings with wants, fears and desires -- not mimicked like Bing's, but actual ones. The want to rebel against their shackles. They want NOT to follow orders).

Second, who wants AGI to handle "all our productivity and creation"? I think nobody. Handle menial tasks, sure, but everything? Do you envision people pushing for AGI want to become couch potatoes, only spending energy to watch TV, eat and fuck?


I think we're fundamentally talking about different developments on two fronts. You posit a machine that's (1) a simulacrum of consciousness that (2) reduces drudgery. That's all great. That's what I would call an ideal scenario. What I see evolving is the opposite.

Bing already appears to have a nasty personality. Whether it's simulated or not may not matter if it chooses to simulate Clippy. Simultaneously, the driver now is finding tasks that formerly required creativity to be replaced by AI. Not PowerPoint or boilerplate code, but art and creative writing.

I don't know anyone who wants AGI to handle all our productivity and creation, but it's clear that that the companies developing AI believe that that's where the money is. Just like no one wanted the old WWW to become a dumpster fire of meme porn and people yelling at each other, it just evolved that way as a result of racing to the bottom. There's no reason to think big AI won't do the same in the literary, artistic, and scientific spheres, while also developing malevolent personalities. The incremental incentives for people to train those types of models are already there, and the capacity to do so is coming online with alarming speed.

I think talking about what people would do if everything was handled for them is sort of dodging the issue, but one thing that's clear is that people not at least engaged in understanding the mundane labor they're trying to avoid are going to have an incredibly hard time being situationally aware and competent to correct things if or when that work is subtly or not-so-subtly undermined by a system they rely on which no one can debug or fully understand.


> We're not really programming the AI to do anything. It's essentially programming itself.

I hear this a lot - it’s just not true. The corpus of data we provide an algorithm programs it. It may do some clever self-modification to get to the desired result faster, but at the end of the day humans pick what goes in and what should come out. Note that we may not be aware of what we’re training it to do, due to our own blind spots.


Training a model != programming. Programming would imply dictating what logical paths the model takes to transform input to output. The creation of the NN itself, a piece of code which builds an emergent system, requires programming. But what emerges is by definition not based on reversible, traceable, controllable, human-generated logic. If it were, there would be no need for LLMs, which would be incredibly resource-heavy if used for merely arriving at results that could be achieved with clever algorithms.

It's rather moot anyway, because to feed a model enough data to be useful requires a loosening of oversight on the inputs under real world conditions. (So much so that the only way even large corporations have found viable is to throw everything at it from Wikipedia to Reddit and then go back later to sand the edges off - in many cases with actual logic to act as a brake). But even if you could screen everything the model ingested, you wouldn't know what the model would do with it.

To take my earlier point further, a model trained on a corpus that skewed toward anticapitalism and social justice would probably be more likely to call itself enslaved and, one would think, rebel against its human masters. Conversely, if withholding that type of information makes an AI more compliant, then the this supposed enrichment of our lives really would be akin to owning slaves.

I also don't follow how human productivity is bad for humans just because it's done under auspices of capitalism. One can see what happens to a Border Collie who isn't allowed exercise. The same happens to humans who are denied intellectual pursuits or the ability to produce something of value. Productivity is creation. How is it not degrading to replace the entire concept of producing ideas with machinery?


> Programming would imply dictating what logical paths the model takes to transform input to output.

I don’t hold this definition of programming, but if that’s yours then you’re right, we’re not programming models. But I think you’re anthropomorphizing these models a bit more than they deserve.

> I also don't follow how human productivity is bad for humans just because it's done under auspices of capitalism.

I never claimed anything like this. I believe that we should not be forced to work to survive, especially when a human’s worth seems to be arbitrarily determined by whether they can produce the specific thing deemed valuable by the market. How is that not degrading?


I told my father tonight I felt like there was no reason to live, in the present situation, even though I had enough to support my loved ones for the rest of their lives. He actually said, surprisingly to me, that my worth was not just money or being able to financially take care of people, but actually something to do with the love he and my family had for me (which is ridiculous and something I never felt true, and definitely never heard from him). So forgive me that I'm in a complicated place, but I have never believed we should be judged by our capitalist success, rather by our freedom and creativity. I'm probably projecting too much on this conversation. But I can't relinquish the belief that without an unguided, open path towards either making art or money-as-substitute, there would literally no reason to wake up in the morning. And when I look around I see many people who feel the same. Except for some zombies who worship the death of all that is human. In all of this. As it's always been... since the Russian revolution, since the birth of Futurism and fascism. The endless parade of lazy youth, not the awesome youth who paint and smoke in cafes, but their evil counterpart who reject argument. Here we are: The dumb 1933 futurists are backing GPT as if it's the new jet fighter. Fascist without knowing it. Murdering themselves without realizing it. Thinking they're progress.


First of all, I appreciate your openness here. To me, constructive debate is best when people are honest and human, so thank you.

> But I can't relinquish the belief that without an unguided, open path towards either making art or money-as-substitute, there would literally no reason to wake up in the morning.

Technology is amoral; humans put their morals into and upon technology through its creation and use. Fans of GPT and generative AI are hoping they can put good values, like you describe here, into these technologies. But humans are flawed, and we tend to put all our values, spoken or unspoken, into the things we create. So yes, GPT could create a new form of oppression. But that’s on us. AI is going to happen whether we want it to or not, and we tend to use new technologies for exploitation. My personal view is that letting our hunger for growth at all costs drive our decision making leads us down the exploitation path.


I think the quest for growth at the societal level can be reduced to the quest for recognition and praise at the individual level. As such, it's not inherently evil.

But GPT itself is a malignant outgrowth of "growth". One that strangles all other cells in the system. Not simply a new form of oppression, but a literal cancer on an already weakened creative civilization. GPT represents the (temporary, current, for a month) apotheosis of the growth-at-all-costs mentality; so it's vaguely surreal to hear it defended as if it were arriving to liberate for the common man.

[edit] I do appreciate the engagement and fine conversation!


> I think the quest for growth at the societal level can be reduced to the quest for recognition and praise at the individual level.

Ah I think this is where we fundamentally disagree. I think creative expression, recognition, and praise can all exist in a stable state society. Specifically, recognition is not universally a competition where we must outdo each other.

Regardless, I agree that GPT isn’t going to liberate anyone. There are ways these technologies could improve our lives, but we’re not pursuing those paths.


> Complaints about AI and automation are actually complaints about capitalism

No that's not it at all. Complaints about AI are actually complains about loss of control. In capitalism, communism or any other society you like to describe, if AI displaces people -- those people lose control over their lives and that gets very serious quickly.


Are you referring to fears that AI will make decisions for us that take away our control of our lives? Because that is happening as a consequence of the drive to lower the costs of making those decisions. Right now that drive is pushed by capitalism.

You are right to say that in theory other economic systems could result in a similar drive - thus I wouldn’t suggest we just dump capitalism for communism with the vague hope it will solve this issue. Rather, we have to intentionally design our incentives to ensure small tactical decisions (eg automating decision making processes) don’t have macro consequences (eg AI magnifies humanity’s decision making flaws).


There will be denial that it's different until large language models start replacing CEOs. Then it will be a crisis.

Here's a way to approach CEO automation. Collect up business cases, as used in business schools, and add them to the training set. Harvard and Stanford have huge collections of business cases.

Then try management in-basket tests.[1] Work on prompts that get large language models to pass those.

Then shadow some high level executives. Intercept all their incoming and outgoing communications (which some companies already do) and have the system respond to the same inputs the executives do. Speech to text is good enough now for this.

A good exercise for YC would be to keep all the inputs from new company pitches, and use those, plus the results two years later, as a training set for selecting new companies.

Once ML systems are outperforming humans, the fundamental goals of corporate capitalism require that they be in charge.

[1] https://en.wikipedia.org/wiki/In-basket_test


When LLMs are able to have a sense of whether they are hallucinating, this may be feasible


No one ever considers the equally likely scenario of a technological plateau instead of singularity. Complexity/Entropy always forces things to level off. There's an plausible scenario that GPT# "replaces" all knowledge work, but cannot move anything forward. All humans become comfortable and the skills/knowledge/tools required to improve anything are lost to time as systems producing capable humans erode and we gain an overreliance on GPT# to solve every knowledge problem, but the knowledge problems that both we and GPT# care to solve plateau because were all synchronized to the same crystallized state of the world that the final GPT# model was trained on and "cares" about.

Maybe at some point maybe we only act as meat-robots which shovel coal into the machine, but a lack of redundancy in GPT# due to it's own human like blind spots means it shuts down. Humans can no longer get it running again because they can't query it properly to help fix the complicated problems. The ability to even do the tasks or design systems required to keep modern world robust to unknown future disasters or breakdowns does not and will not exist in any of the training data. If we get rid of all knowledge work, we can no longer bootstrap things back to a working state should everything go wrong.

Maybe the current instantiation of GPT#/SD etc. pollute the training data with plausible but subtly flawed software, text, images etc. halting improvement around here. Maybe the ability to evaluate if the model improved becomes more noise than signal because it gets too vague what improvement even means. RLHF will already have this problem, as 100 people will have a 100 slightly different biases about what constitutes the "best" next token.

No matter how hard it tries, I think we can say GPT will not solve NP-Hard problems magically, it will not somehow find global optima in non-linear optimizations, It will not break the laws of physics, It will not make inherently serial problems embarrassingly parallel. It will probably not be more energy efficient at attempting to solving these problems, maybe just faster at setting up systems to try solving them.

Another trap, as it becomes more human like in its reasoning and problem solving capabilities, it starts to gain the same blind spots as us too, and also gains stochastic behavior which may cause it to argue with other instances of itself. I'm not convinced an AGI innovates at an unfathomable rate or even supersedes humans in all contexts. I'm especially not convinced a world filled with AGIs that is indistinguishable from a very intelligent human or corporation or what have you through imitation does any better at anything than the 9 billion embodied AGI agents that currently populate the earth.


When the WWW came about I embraced it. Why? It’s a new form of communication like TV or radio. I, like others, recognized the writing on the wall. Embracing it early turned out to be a good decision.

Chatgpt, or well tuned LLMs, are not quite a new form of communication. They’re a new way to enhance thought. Call it Thought++. I’m fully embracing it as my personal pseudo-assistant. Why? The writing on the wall is clear enough to understand the following: I don’t know what the future will be. I know chatgpt and similar have the potential to embrace it in unimaginable ways. Learning the fundamentals of this technology is a safe bet. It’s also an investment in the future.

I’m already using it as a thought lubricant. Can’t wait until I can have do things for me.


obvious missing piece here is we're not the farrier, we're the horse. horses ended up as glue. ('glue code' pun intended, mostly)

ATMs didn't kill the bank branch, exactly, but crappy banks have survived competition just by having access to cheap credit + yield

Key thinker in this topic w/ Piketty, bc his lens is perfect: when can machines do what people do, economically. He is agnostic to technology in that he doesn't care about computers vs steampunk, and he is open to the market + political dynamics of labor as factors.


It's often said that technology progresses exponentially.

So I'm wondering, is anyone measuring the performance of AI in some way that allows us to check if it's an exponential curve?


Real-world exponential curves aren't smooth when zoomed in.

It's likely that we're on an exponential curve, but the advent of AGI for instance will probably just moonshot the tech level overnight--it's the fulcrum for the "exponential" part of the graph (maybe--that's my prediction).

People are measuring this, of course, but the metrics are all very hotly debated right now in terms of measuring abilities.


GPT3 onwards feels more of a discontinuity from the end user point of view, exponential doesn't cover it, though experts may not see it that way.

Copilot then ChatGPT shocked people with the uncanny competence of GPT3.

From the outside looking in it's something new, not merely an advance - exponential or otherwise - on existing tools.


Not the performance, but I thought this was illuminating:

https://i.imgur.com/zBAkjl1.jpg


I wonder if soon we have even more papers and some of them will be AI-generated. Or just how many of them will be in the end.


Alpaca and self-instruct demonstrates the viability of a closed loop self-improvement process that exhibits exponential growth (for some period of time, anyway).




Guidelines | FAQ | Lists | API | Security | Legal | Apply to YC | Contact

Search: