Google has plunged the internet into a “spiral of decline”, the co-founder of the company’s artificial intelligence (AI) lab has claimed.
Mustafa Suleyman, the British entrepreneur who co-founded DeepMind, said: “The business model that Google had broke the internet.”
He said search results had become plagued with “clickbait” to keep people “addicted and absorbed on the page as long as possible”.
Information online is “buried at the bottom of a lot of verbiage and guff”, Mr Suleyman argued, so websites can “sell more adverts”, fuelled by Google’s technology.
The part about Google isn't wrong.
But the second half of the article, where he says that AI chatbots will replace Google search because they give more accurate information, that simply is not true.
I'd say they at least give more immediately useful info. I've got to scroll past 5-8 sponsored results and then the next top results are AI generated garbage anyways.
Even though I think he's mostly right, the AI techbro gameplan is obvious. Position yourself as a better alternative to Google search, burn money by the barrelful to capture the market, then begin enshitification.
In fact, enshitification has already begun; responses are comparatively expensive to generate. The more users they onboard, the more they have to scale back the quality of those responses.
ChatGPT is already getting worse at code commenting and programming.
The problem is that enshitification is basically a requirement in a capitalist economy.
Even if AI magically got to the point of providing accurate and good results, I would still profoundly object to using it.
First, it's a waste of resources. The climate impact of AI is enough of a reason why we should leave it dead until we live in a world with limitless energy and water.
Second, I don't trust a computer to select my sources for me. Sometimes you might have to go through a few pages, but with traditional search engines at least you are presented with a variety of sources and you can use your god given ability of critical thinking.
I’m not sure what you think modern search engines do, but this is pretty much it. Hell, all of the popular ones have been using AI signals for years.
You can request as many sources from an AI as you would get from Google.
Of course there are always challenges, especially with how results are ranked. I have been extremely dissatisfied with the development of search engines for years now. I find Duckduckgo to at least be less bad than Google. Currently I'm checking out Kagi, which at least lets me rank sources myself. Still on the fence though - it does seem to flirt more with AI than with transparency, which has me worried.
But absolutely, it's not that I think the current state of search engines is great either - it just seems to me everything is getting worse and the Internet has entered a death spiral between AI and the enshittification of social media.
Then again, maybe I just reached that age where you start hating everything.
That's LLMs, which is what is necessary for Chat-AI (the first "L" in there quite literally stands for Large).
Remove the stuff necessary to process natural human language and those things tend to be way smaller, especially if they're just trained using the user's own actions.
I mean most top searches are AI generated bullshit nowadays anyway. Adding Reddit to a search is basically the only decent way to get a proper answer. But those answers are not much more reliable than ChatGPT. You have to use the same sort of skepticism and fact checking regardless.
Google has really gotten horrible over the years.
Most of the results after the first page on Google are usually the same as the usable results, just mirrored on some shady site full of ads and malware.
I already go to ChatGPT more than Google. If you pay for it then the latest version can access the internet and if it doesn’t know the answer to something it’ll search the internet for you. Sometimes I come across a large clickbait page and I just give ChatGPT the link and tell it to get the information from it for me.
Do you fact-check the answers?
It depends what you’re using it for as to whether you need to fact check stuff.
I’m a software developer and if I can’t remember how to do an inner join in SQL then I can easier ask ChatGPT to do it for me and I will know if it is right or not as this is my field of expertise.
If I’m asking it how to perform open heart surgery on my cat, then sure I’m probably going to want several second opinions as that is not my area of expertise.
When using a calculator do you use two different calculators to check that the first one isn’t lying?
Also, you made a massive assumption that the stuff OP was using it for was something that warranted fact checking.
I can see why you would use it. Why would I want to search Google for inner joins sql when it is going to give me so many false links that don’t give me the info in need in a concise manner.
Even time wasting searches have just been ruined. Example: Top Minecraft Java seeds 1.20. Will give me pages littered with ads or the awful page 1-10 that you must click through.
Many websites are literally unusable at this point and I use ad blockers and things like consent-o-matic. But there are still pop up ads, sub to our newsletter, scam ads etc. so much so that I’ll just leave the site and forego learning the new thing I wanted to learn.
The new release of GPT-4 searches Bing, reads the results, summarizes, and provides sources, so it's easier to fact check than ever if you need to.
give it time, algos will fuck those results as well
They'll need to make money with a cheap cost-per-sale, so they'll put ads on the site. Then they'll put promoted content in the AI chat, but it's okay because they'll say it's promoted. Eventually it won't even say it's promoted and it will just be all ads, just like every other tech company.
Why? Because monetization leads directly to enshittification, because the users stop being the customers.
When I tried it it was never able to give me the sources of what it said. And it has given me way too many made up answers to just trust it without reasons. Having to search for sources after it said something has made me skip the middle man(machine).
ChatGPT powers Bing Chat, which can access the internet and find answers for you, no purchase necessary (if you're not on edge, you might need to install a browser extension to access it as they are trying to push edge still).
Do you fact-check the answers?
That's such a strange question. It's almost like you imply that Google results do not need fact checking.
They do. Everything found online does.
With google, it depends on what webpage you end up on. Some require more checking than others, which are more trustworthy
Generative AI can hallucinate about anything
There are no countries in Africa starting with K.
LLMs aren’t trained to give correct answers, they’re trained to generate human-like text. That’s a significant difference.
They also aren't valuable for asking direct questions like this.
There value comes in with call and response discussions. Being able to pair program and work through a problem for example. It isn't about it spitting out a working problem, but about it being able to assess a piece of information in a different way than you can, which creates a new analysis of the information.
It's extraordinarily good at finding things you miss in text.
Yeah. There's definitely tasks suited to LLMs. I've used it to condense text, write emails, and even project planning because they do give decently good ideas if you prompt them right.
Not sure I'd use them for finding information though, even with the ability to search for it. I'd much rather just search for it myself so I can select the sources, then have the LLM process it.
Agree.
I found it more tempting to accept the initial answers I got from GPT4 (and derivatives) because they are so well written. I know there are more like me.
With the advent of working LLMs, reference manuals should gain importance too. I check them more often than before because LLMs have forced me to. Could be very positive.
From my experience with BingChat, it's completely true. BingChat will search with Bing and summarize the results, providing sources and all. And the results are complete garbage most of the time, since search results are filled with garbage.
Meanwhile if you ask ChatGPT, which doesn't have Internet access, you get a far more sophisticated answer and correct answer. You can also answer follow up questions.
Web search is an absolutely terrible place for accurate information. ChatGPT in contrast consumes all the information out there, which makes it much harder for incorrect information to slip in, as information needs to be replicated frequently to stick around. It can and often is still wrong of course, but it is far better than any single website you'll find.
And of course all of this is still very early days for LLMs. GPT was never build with correctness in mind, it was build to autocomplete text, everything else was patchwork after the fact. The future of search is AI, no doubt about that.
Chatgpt flat out hallucinates quite frequently in my experience. It never says "I don't know / that is impossible / no one knows" to queries that simply don't have an answer. Instead, it opts to give a plausible-sounding but completely made-up answer.
A good AI system wouldn't do this. It would be honest, and give no results when the information simply doesn't exist. However, that is quite hard to do for LLMs as they are essentially glorified next-word predictors. The cost metric isn't on accuracy of information, it's on plausible-sounding conversation.
Ask chatgpt "tell me the biography of the famous painter sndrtj" to see how good the bot is at hallucinating an incredible realistic story that never happened.
Oh, they fixed that! But I see you're using v4.
You don’t even have to make stuff up to get it to hallucinate. I once asked chat gpt who the original bass player was for Metallica was, and it repeatedly gave me the wrong answer, and even at one point said “Dave Ellefson.”
Its already happening at my work. Many are using bing AI instead of google.
Don't worry they'll start monetizing LLMs and injecting ads into them soon enough and we'll be back to square one
I suspect that client-side AI might actually be the kind of thing that filters the crap from search results and actually gets you what you want.
That would only be Chat-AI if it turns out natural language queries are better to determine the kind of thing the user is looking for than people trying to craft more traditional query strings.
I'm thinking each person would can train their AI based on which query results they went for in unfiltered queries, with some kind of user provided feedback of suitability to account for click-bait (i.e. somebody selecting a result because it looks good but it turns out its not).