Edit: Benedict Evans no longer works for a16z. However I think it must be a reference to Why Software Is Eating the World.
We need AI to solve climate change, but the energy costs will contribute to climate change. We need a cure for cancer, and an end to COVID and other dangerous viruses, a way for a personal AI that does tasks and earns the user money for basic income or something.
It's a chat bot. Solving climate change is way above its paygrade.
something something RNA Moderna something -- yeah, okay, in niche fields.
what, specifically, is AI doing to fix climate change? outside of creating a shit-ton of carbon burning through electricity, that is.
And probably also note that the proposals of the climate-justified-vandals are too unrealistic
There's nothing more nonsense than a solar-nuclear antagonism. Build more solar. Build more nuclear, preferably SMRs, and do more research on ways of making it safer and cost-effective.
Fill the dry sunny areas with solar and batteries. Fill the windy areas with wind turbines and treat anyone opposing this with the same contempt as people vandalizing paintings "for the climate"
Every car should be at least a hybrid, because there's nothing more clunky than plugging an engine that deals badly with varying RPM and torque directly to a variable load.
The problem is that people’s don’t use it for what it’s good for, but they try to make it code and play chess, because they can’t accept that it can’t think (yet).
The environmental impact of solar and wind turbines is huge, much greater than coal or nuclear.
(By "environmental impact" I mean "natural habitat destruction", not "my property values went down because the air is icky".)
See, this is the kind of BS boomers come up with to justify the current idiocy and that other boomers will eat hook, line and sinker without any critical thought
Pray what kind of "environmental destruction" putting solar panels in the desert will cause that shadows actually digging the soil for coal?
The fact that a coal plant emits more radiation than a nuclear plant should be sufficient to establish how much of a BS this fact it
But I know, some people think training AI in space is cheaper, some will believe anything they're paid enough to believe
(I obviously am in favour of nuclear plants and your statement might be correct in part there)
Because we do not live in the desert, and moving electricity is terrible (loss is insane)
Habitats of desert animals will be destroyed and they will go exinct. (But humans don't live in the desert and desert animals hold no cultural significance, so who cares, amirite?)
> The fact that a coal plant emits more radiation than a nuclear plant
Irrelevant. Chernobyl was the best thing to happen to Europe's biodiversity in centuries.
If you take its future promise as true (I do not, but for sake of argument lets pretend it is as powerful as you're saying), then it's impact is on a completely different scale and ability than even industrialisation. Its silly to retreat to "no one shoes horses anymore" platitudes as if this tool isn't widly more capable (again, if we take its future promise as written) and disruptive.
Agriculture workers, became industry workers. (Mechanized agriculture).
Industry workers, became service and white collar workers. (Automatization).
I don't see a sector for white collar workers to shift to now. Even more service workers?
There is room in the serf and capitalist classes though. Hopefully we end up in the later rather than former.
But there is a need for political reform to open up the later I guess.
3 day work week and so on the lower the lowering pressure on wages. Progressive corp. tax? But I don't see that happening in the US.
How does this work? Is AI eating Mobile now? Is there a new world to be eaten every decade or so?
Ah, wait. It's just meaningless hyperbole.
Yes
It's a meaningless soundbite.
2: Mobile replaced the PC as the main way people use the internet and do their day-to-day computing. The consumer Internet runs on smartphone apps, not PCs. In 2013 a lot of people didn't understand that that was happening, so it was worth saying.
Though some people want to save it.
I prefer eating.
Back in the 2000s-2010s the hyperbole was that software was eating the world. And it did, almost everything on the modern world depends on software to function.
Will AI eat the world? We will only know in hindsight, right now it's too much in flux but it will be clear in another 20 years if the hyperbole is true or not.
Indeed, you would see that if you'd read even the first half-dozen 5 slides ;)
My time is not free, sorry.
I remember in the late 90s, me and some friends discovered the internet and went "Woah! Letters, books, shops, work, study, gaming ... pretty much everything will take place in this virtual realm soon!".
We were right. Except for the "soon".
When the first internet café opened, we started spending a lot of our time there, meeting lots of interesting people. Everything outside the internet already seemed anachronistic to us.
What I didn't expect back then was that even replacing letters with email (which already worked!) would take another painstaking ... 20 years!
Today, my work is already massively transformed via neural networks. Pretty much everything I do starts with an interaction with a neural network. Usually a question to an LLM. And then LLMs or other networks are involved all the way up to finishing the task. Thinking about how I worked before these large neural networks came up makes me shiver. How cumbersome. How anachronistic.
But when I talk to people outside the tech sphere, most are not using LLMs or other neural networks at all.
Everything humans do will be deeply transformed. Even more so than via the internet. Will it take 20 years again? Or is technological progress accelerating and this time it will take only 10 years? So far, my feeling is that it will be a bit faster, but not twice as fast.
But maybe it will be even more surprising. One future I can imagine is that people don't even need to adapt. Maybe we will see software simply do all aspects of a job. One can already imagine it for driving cars or making movies. It is harder to imagine it for managing a company, for example. But who knows. LLMs are already surprisingly good at creating chains of thoughts. Maybe we'll find ourselves in an unexpected future in a few years already, where human involvement in anything is just a burden to the process.
I can’t even type a note to myself in gmail without “polish!” Popping in to tell me that I’m shit at writing a good note-to-self and it needs polishing. (Wherein it typically reforms to , “dear customer service, …”)
Every Google search is using llm now right? At least I can’t figure out how to stop it from happening.
It's an append only system and the happy path for note taking is a simple keybind pop up that only displays 30 characters of input on screen. I cannot go back and edit the notes later, but I can open a larger app to write new ones that reference(/supercede) the originals, which allow me to refine the note
The problem is reliability,you can't bulid of unstable foundations.
We would need some kind of library of "verified sentences" in LLMs, like facts verified by experts for them to be useful for a tons of stuff.
Right now if you are using LLMs for anything else that is not verifiable in a objective way (like programming, a program works or does not [let's not even talk efficency or security]) or fluff (generating fancy text for non crucial reasons) you are using a tool wrong.
You can build valuable, reliable systems on top of unreliable foundations. That's how humanity has progressed over the centuries.
First of all humans or websites have reputations, with GPT you just hit refresh and you are talking to a entirely different entity and everything they said is gone.
I feel like there's a difference.
In practice, I've found that the risk of LLMs hallucinating against well chosen context in low enough that I rarely worry about it.
When reality is you get a bit smarter assistant than ever before, still not very trustworthy but already can be saving some time if used right. Good luck hearing this ever from folks riding that hype wave.
I'm sure if we look, we can find a five year old HN post that starts out: "My work is already massively transformed via crypto. Pretty much everything I do starts with an interaction with a blockchain."
Pretty much nothing I do starts this way.
Look LLM's are interesting. I sure spend a lot less time writing basic one off scripts because of them. The "extra step" of tossing emails to an LLM is just proof reading with less tedium.
LLMs gave every one an intern that does middling work quickly, never complains and doesn't get coffee.
We need them to be cheap (to run) and localy/owned hardware (for security and copy right reasons).
At first glance that might seem expensive, but then consider how insane it is that you can ask your laptop arbitrary questions and have it respond with really cogent answers, on almost any topic you can think of, without relying on a massive rack of gpu machines behind an api. It uses barely more power than an old incandescent bulb while doing it!
Even just getting cursor or boltai and you have unlimited llm cloud access for the price of a few coffees.
Just an example was in a mild debate I was having regarding cars, the other person posed a question about how fast a Golf Type R could get in a specific distance and Chat GPT spat out a number that the other person accepted as fact, but I already knew it was too high.
What ChatGPT done was taken the posted 0 to 60 time and extrapolated a linear distance vs velocity formula. Which was impressive granted, but wrong; velocity over distance is logarithmic at best.
It's a great tool, but I think a lot of people are just taking what it spits out without slowing down to question if the output makes sense or not.
That doesn't make me some fancy scientist level programmer (I'm definitely not) that I often find my attempts at using AI falling into that category...a lot of the time it's just due to niche platforms and libraries and things that are specific to our shop or the regulatory environment or a thousand other issues of that nature. I imagine that similar issues are incredibly widespread for basically anybody that is not doing greenfield work that is somewhat isolated and at young companies and isn't spending tens of millions to do custom training on their specific environments.
The whole "everything web, most everything open source, ship ship ship new code" style work environments you tend to find among young start ups are not as common as I think they seem if you guage your view of technology jobs based off of hacker news. Given that most of the training of the most powerful models is basically scraping the web, it's not at all surprising that they are seriously lacking in other areas. And I'm not sure to what extent they can seriously be expected to improve there...besides the obvious issue of uploading internal documentation to give an external LLM better prompting...the thing has still got to be able to use public training data to make predictions about internal libraries and whatnot that may very well be old or anachronistic or batshit crazy, because the difference in volume of data between say your internal software and everything posted publicly on the internet is massive.
Crypto was useful for speculative investment and occasionally for working around money transfer restrictions.
I continue to try applying small models to tasks life producing structured data from unstructured (using lmstudio's schema parameter) and it's really neat that I can extract data from work emails, but every once in a while it will hallucinate a new phone number when the one I needed was in the prompt, so it still hasn't met the kind of reliability I want in order to actually automate stuff where dollars are on the line
I don't for one second really suspect it will be the case (not for the usual technical criticisms although I'm skeptical there as well, but more that I don't think it would be socially sustainable for an extended period of time)...but let's for a moment take your last paragraph at face value and in good faith.
I mean...what exactly is it that you are advocating for or accepting? Even if we get some kind of very generous UBI, there's something about human nature that makes me suspect the consequences of this would be an almost guaranteed miserable existence for pretty much everyone.
Even in the best case scenario, where the results of this transformation are kept under control and distributed in a reasonable manner and the whole thing doesn't cause a social and political meltdown...what is everybody going to do? There's some amount of wisdom in the old saying that "idle hands are the devil's plaything".
Thats the real issue I am the most concerned about and that seems to be the least often addressed by big AI boosters and detractors (I realize both of these camps often have ulterior motives). I suspect many are feeling some amount of concern like that...why is this (I would argue most fundamental) question about the impact of AI never talked about?
I don't want to hear anything about some big terminator style fight against AI or about how wonderful and unpredictable the inevitable future of WALL-E style luxury gay space communism is going to be...none of those discussion points get to the heart of what makes many people so uncomfortable with the concept...and I think the people believing in some version of that second scenario being at all socially plausible is what gives me the most pause. It makes the terminator scenario almost seem like a preferred outcome if we were given a binary choice...in reality I think most would prefer neither and would agree with me in saying that we aren't even discussing the right issues w.r.t an "AI gets much better" potential future.
Superhuman AI seems to be a building block of it. And deep transformation of how we work will come with it.
So I raised the question of the timeline.
That's all.
I think it's partly these hangups about the situation that I have that give me the tendency to assume a post like yours is necessarily "advocating" as well, so my apologies if that was not your intent.
I guess the "deep transformation of how we work"
part is what I don't really understand.
At the moment, it looks like we'll either become completely obsolete in the context of work, or we'll morph into cyborgs with vastly greater knowledge and processing capabilities than we have today.I suppose I'll leave it at that.
If those don't terrify you then I guess we just don't really have similar temperament as people and that's totally okay with me. I did enjoy discussing it with you.
To this day, and even though I am deeply involved with AI, I do not regret that decision—but I am constantly reminded that the expectations practitioners have are _way_ off what most people will actually want.
Some true, but not totally. The role of email has many differences from old-age letters. And letters are not only replaced with email, but also more modern communication manners. And the more modern communication manners not only replace letters, but telephone etc.
And the letter case is specified case. There are many old-age things which are not replaced totally.
The same is for AI. It will replace something, but not all.
ChatGPT got to 100m users much faster than anything else because it's riding on all the infrastructure we already built in the last 20 years. To a consumer, it's 'just' a website, and you don't have to wait for telcos to build broadband networks or get everyone to buy a $600 smartphone.
But, most people go to the website and say 'well, that's very cool, but I don't know what I'd use it for'. It's very useful for coding and marketing, and a few general purposes, but it isn't - YET - very helpful for most of the things that most people do all day. A lot of the presentation is wondering about this.
A quick Google search for "most common job" came back with
Cashier
A cashier works in a retail environment and
processes transactions for a customer's purchase.
I wouldn't be surprised if robots can do that on their own in 10 years.What they can't do is call the police when the hobo gets too wild, can't fix the inevitable bug in the process(by doing some 4th level menu bypass) and other random stuff that might pop up.
And when the robot can do all that humans are no longer viable as economic entities and will be out competed.
That's the beauty of human interaction, it can't be massively truncated down to just even finger pointing.
AI is a trash compactor run through an HR department. It's already neutered and censored to hell. If you want to get the official, approved version of something, ask an AI. If you think that's fun, you've probably forgotten what it was like to be young.
Here's a useful question: Suppose the LLM hallucination problem is not solved in the next 10 years. What happens to the AI boom?
this is one of the "no-one knows" questions
In LLR you fit a massive chunk of data, you do an interpolation and you use the result with care, or, you fit the data, you do an extrapolation and you live dangerously. So its all good fun, but how to you eat the universe with it?
You might try to become the Matlab or SAS or Excel that brings Large Linear Regression to the world. Godspeed, a giant Python has already swallowed C++ and they will be coming be after you.
You may try to eat all the world's data so that nobody but you can do Large Linear Regressions. Good luck with that pharaonic quest as well.
Or you might try to print silicon that does LLR quickly, hoping that nobody else can master that dark art. But silicon is as plentiful as sand and keeping dark secrets dark is not easy when you want to eat the world.
Thats about it. No melodrama, no medieval moats, no megazillion dollars. LLR will be everywhere and yet nobody will care.
The bandwagon will move to the next big tech thing. Quantum Coin maybe?
Assuming the deteriorating reality of our actual condition does not catch up with this clown circus. Because the end condition of the world being serially eaten by the tech bros is... the world being dead.
In most cases, probably giving OpenAI a bunch of money.
For whatever reason, the full stack hasn't been commoditized yet to a degree where you could self-host easily. For example, I can put the paid or free version of GitLab on my servers and get repo management, issue tracking, CI/CD, Wiki and a bunch of other stuff. It covers most use cases and works out of the box, even if not always in the ways I want.
As for AI... there's OpenAPI and GitHub Copilot, even JetBrains has their AI solutions. You pay for access to the back end component and there's IDE plugins that integrate with that, even custom IDE's or editors like Cursor. But what if you want an editor/plugin that talks to models running on your own servers? Sure, you can get models off of HuggingFace and hook them up to run locally on a machine that has the hardware to take advantage of them... but then what? What about integrating with merge requests in the aforementioned GitLab instance? Obviously it's all possible, but somehow I haven't seen many solutions that offer you something similar to GitLab but for AI.
Even GitLab's own solution talks to their servers: https://about.gitlab.com/solutions/code-suggestions/
> Code Suggestions is available to self-managed GitLab instances via a secure connection to GitLab.com.
I'm guessing CodeGPT is probably a piece of that puzzle, or maybe the Tabnine enterprise setup.
So I think what you're currently imagining won't happen until GPU prices go down massively