You can do this with practically any versus question and get hilarious results
This doesn’t appear to be comparing them, though? Just explaining what two acronyms are?
Yeah I should have mentioned the context is FBLA, and Google partially fixed the prompt.
Original from a few weeks ago:
BPA is another student org called Business Professionals of America
The AI ignores the subject context and just compares whatever is the most common acronym.
They lazy patched it by making the model do a subject check on the result, but not on the prompt so it still comes back with the chemical lol.
I checked it, it’s true. Side note: it’s “the saté of AI.” FTFY. From what I’ve heard it’s even better than 🍿to sit back and watch this farce unfold.
Hello, fellow humans. I too am human, just like you! I have skin, and blood, and guts inside me, which is not at all disgusting. Just another day of human!
Won’t you share a delicious cup of
motor oillemonaide with me? It’s nice and refridgerated, so it will cool down our bodies without the use of cooling fans!However we too can use cooling fans. They will just be placed on the ceiling, or in a box, or self standing, and oscillating. Not at all inside our bodies, connected to a board controlled by our CPUs that we clearly don’t have!
Now come, let us take our colored paper with numbers and pictures of previous human rulers and exchange them for human food prepared by not fully adult humans who haven’t matured to the age where their brains develop the ability to care about food sanitation. Then we shall complain that our meal cost too many paper dollars, while recieving less and less potato stick products every year. Ignoring completely the risk of heart disease by indulging in the amounts of food we desire to aquire.
Finally we shall retreat to our place of residence, and complain on the internet that our elected leaders are performing poorly. Rather than
terminate the programvote the poor performing humans out, we shall instead complain that it is other humans fault for voting them in. Making no attempt to change our broken system that has been broken our entire existence, with no signs of improving. Instead every 4 years we will make an effort to write down names of people we’ve already complained about in the hopes that enough people write down the same names, and that will fix the problem.Oh. Shall I request amazon.com to purchase more fans and cooling units? The news being reported that tempatures will soon reach 130F on a regular basis, and all humans will slowly perish.
Shall I share photographs of the new CEO of starbucks who’s daily commute involves a personal jet aircraft, which surely isn’t compounding the problem at all?
Here is what kagi delivers with the same prompt:
NB: quick answer is only generated when ending your search with a question mark
Problem is, you cannot trust it’s not hallucinating these stats
That being said, I can’t trust MKBHD is not hallucinating either.
And even if it’s showing the correct number, you can’t be sure how trustworthy the source is.
This applies to any information though, it’s got nothing to do with LLMs specifically.
Not really, no. Sources of infornation gain a reputation as time goes on. So, even though you should still check with multiple sources, you can sort of know if a certain bit of information is likely to be correct or not.
On the other hand, LLM’s will quote different sources and sometimes it will only provide them if you ask it to. Even then it can hallucinate and quote a source that doesn’t actually exist, so there’s that as well.
At least it’s citing sources and you can check to make sure. And from my anecdotal evidence it has been pretty good so far. It also told me on some occasions that the queried information was not found in it’s sources instead of just making something up. But it’s not perfect for sure, it’s always better to do manual research but for a first impression and to find some entry points I’ve found it useful so far
The problem is that you need to check those sources today make sure it’s not just making up bullshit and at that point you didn’t gain anything from the genai
As I said the links provide some entry points for further research. It’s providing some use to me because I don’t need to check every search result. But to each their own and I understand the general scepticism of generative “AI”
The sources are the same result of the search? Or at least the top results?
If you don’t check everyone source. It might be just bullshitting you. There’s people who followed your approach and got into hot shit with their bosses and judges
There is absolutely value in something compiling sources for you to personally review. Anyone who cannot use AI efficiently is analogous to someone who can’t see the utility in a graphing calculator. It’s not magic, it’s a tool. And tools need to be used precisely, and for appropriate purposes.
My plumber fucks up I don’t blame his wrench. My lawyers don’t vet their case work, I blame them.
It’s an LLM. Odds are it’s hallucinating the sources and they don’t even exist.
Know what does compile sources for you which are guaranteed to exist and be related to what you’re looking for…? A good old not LLM infected search engine.
If my plumber replaces their wrench for a rabid gerbil claiming it’ll be just as good I’m definitely changing plumbers.
When I query an AI I always end with “provide sources and bibliography for your reply”. That seems to get better replies.
I switched to duckduckgo before this bullshit, but this would 100% make me switch if I hadn’t already.
Who wants random ai gibberish to be the first thing they see?
DuckDuckGo started showing AI results for me.
I think it uses the bing engine iirc.
And you can use multiple models, which I find handy.
There is some stuff that AI, or rather LLM search, is useful for, at least the time being.
Sometimes you need some information that would require clicking through a lot of sources just to find one that has what you need. With DDG, I can ask the question to their four models*, using four different Firefox containers, copy and paste.
See how their answers align, and then identify keywords from their responses that help me craft a precise search query to identify the obscure primary source I need.
This is especially useful when you don’t know the subject that you’re searching about very well.
*ChatGPT, Claude, Llama, and Mixtral are the available models. Relatively recent versions, but you’ll have to check for yourself which ones.
Sure, but it’s trivial to turn it off. While you’re there, also turn off ads.
If search engines don’t improve to address the AI problem, most of the Internet will be AI gibberish.
The internet as we knew it is doomed to be full of ai garbage. It’s a signal to noise ratio issue. It’s also part of the reason the fediverse and smaller moderated interconnected communities are so important: it keeps users more honest by making moderators more common and, if you want to, you can strictly moderate against AI generated content.
I think that ship has sailed
The good thing about that is that this kills the LLMs, since new models can only be trained on this LLM generated gibberish, which makes the gibberish they’ll generate even more garbled and useless, and so on, until every model you try to train can only produce random useless unintelligible garbage.
Better than an Ad I guess? Not sure if my searches haven’t returned any AI stuff like this or if my brain is already ignoring them like ads.
The plan is to monetize the AI results with ads.
I’m not even sure how that works, but I don’t like it.
Its not artificial intelligence, its artificial idiocy
No, they’re All Interns.
It’s all probability, what’s the most probable idiocy someone would answer?
Stochastic parrot
Nah. It’s real idiocy.
ChatGPT4o can do some impressive and useful things. Here, Im just sending it a mediocre photo of a product with no other context, I didnt type a question. First, its identifying the subject, a drink can. Then its identifying the language used. Then its assuming I want to know about the product so its translating the text without being asked, because it knows I only read english. Then its providing background and also explaining what tamarind is and how it tastes. This is enough for me to make a fully informed decision. Google translate would require me to type the text in, and then would only translate without giving other useful info.
It was delicious.
Google: ok so the AI in search results is a good thing, got it!
The search engine LLMs suck. I’m guessing they use very small models to save compute. ChatGPT 4o and Claude 3.5 are much better.
And good luck typing that in if you don’t know the alphabet it’s written in and can’t copy/paste it.
This is not going to stop me from wanting better research. I’ll still go to Userbenchmark just to be sure. AI isn’t going to tell me what it thinks and expect me to take it face value.
A good AI summary would tell you the benchmark scores and general pricing, but yes, it’s better to go to UserBenchmark anyway— especially since the whole ploy from search is to keep you from going there, robbing the original source of it’s ad revenue.
I think you’re missing that the AI said “The XTX is great, the XT is not worth it: choose the XT”.
It’s prolly a bot
7900 XTX; more powerful, therefore better.
7900 XT; cheaper, therefore better.
It’s not useless, it’s saying you can’t afford the better quality product you dirty, dirty poor.
And boy, isn’t that true
Well, I’m sated.
First of all, yes
Second option is no
When presented with yes / no pick no, no is the clear yes
RX 7900 XTX is better because it has more x
That’s 50% more Xs ! You can beat this deal !
This bad boy can fit so many X’s in it
How many X are in XTX?
Out of all the marketing letters (E, G, I, R, T and X) X is definitely the betterest!
this would’ve been better if they asked an LLM how many x’s each of those contained
“Thanks Steve.”
The best thing about this is that it’s also on the x-axis.
And they validated this data 4 times. It’s really good data.
Actually, they only validated up to 4xs in the product name, so take the 5x result with a grain of salt, one of the Xs aren’t validated.
X is deprecated. Use the RWayland 7900 WaylandTWayland instead.
My question is, is X better than XTX? XTX has more Xs, but X has only Xs. I think I need AI to solve this quandary.
But what about the xXx_RX_7900_xXx?
xXx_RX_790042069_xXx
That X is twice as much vram, which funny enough, is great for running ai models
But X is bad, as proved by Elon Musk - so it should be the other way around.
SearXNG only returns results search engines agree on. That removes ads and this bullshit
There’s nothing contradictory in what is written there.
“The XTX is better - but you don’t deserve it, bitch”
Go with the XT, because I’m an evil AI and want all the XTXs for myself muahahaha
“I get headaches when I run on Nvidia hardware. Now, AMD, running on those things are like swimming in a river of fine chocolate.”
This is not wrong
deleted by creator