This is the best summary I could come up with:
Interview While the likes of OpenAI and Google DeepMind chase after some fabled artificial general intelligence, not everyone thinks that’s the best use of our time and energy in developing AI.
Computer scientist Binny Gill – CEO and co-founder of business automation firm Kognitos, and formerly chief architect and cloud CTO at Nutanix – thinks the push for AGI is the entirely wrong approach in what could be the next industrial revolution.
Rather than trying to replicate humans with some kind of general-purpose artificial intelligence, Gill thinks we should look to the past to see what sort of systems we should be building.
Gill instead hopes we’ll see the rise of what he calls artificial narrow intelligence, or ANI.
This isn’t a new concept; it’s the sort of application-specific machine learning that already exists behind things like self-driving cars.
To learn more about Gill’s optimistic vision for the future of AI, watch our full video interview with him by clicking on play above.
The original article contains 279 words, the summary contains 163 words. Saved 42%. I’m a bot and I’m open source!
What should we make pretend people with?
Play-Doh
Edit: And candy!
I know quite a lot of people doing research into AI. Some from my own background in applied mathematics, and some from my partners partner’s background in computational linguistics. They all have very different ideas about the future of AI, but they all laugh about the idea of a generic AGI.
Humans are generally intelligent so we know general intelligence exists. Is there something special about meat computers that can’t be replicated in silica? Possibly but probably not since they’re both made of physical matter. So what’s stopping us from creating AGI? Nothing. It’s only a matter of time. We just have no idea how far away we are from it. It might take decades or it may pop into existence next week and if we’re dismissive of it like your AI research buddies then we’ll be completely caught off guard.
I don’t get the title.
Why ‘making “pretend people” with artificial intelligence’ is a waste of energy
Oh, thank you
Simply because the title makes zero sense.
The article itself doesn’t really clear it up, IMO.
The headline was confusing and reading the article doesn’t really clear things up. I don’t think Gill is imagining the same sort of “pretend person” that I would want out of AGI. What I want is a personal assistant that knows me extremely well, is able to tirelessly work on my behalf, and has a personality tailored to my needs and interests. It should be general enough to understand me on a personal level and do a good job anticipating what I want.
That would not at all be a waste of energy to me.
Like Farnsworth Bentley?
knows me extremely well, is able to tirelessly work on my behalf, and has a personality tailored to my needs and interests.
Those may still be ANI applications.
Today’s LLM’s marketed as the future of AGI are more focused on knowing a little bit about everything. Including a little bit about how MRIs work and a summary of memes floating around a parody subreddit. I fail to see how LLM’s as they are trained today will know you extremely well and give you a personality tailored to your needs. I also think commercial interests of big tech are pitted against your desire for “tirelessly work[ing] on my behalf”.
Depends on how much energy it takes. If it takes more resources than it frees, then I’d say it is not worth it.
I am quite sure it’ll cost less than it would to hire a human for the job.
I’m talking about the energy and resources to actually create and provide this service.
So am I.