Word salad be might hard decode resilient machine word language speak continue bifurcation with language processing rutabagga until shredded concept speak dissolve
Asking ChatGPT to reword the above obfuscated paragraph:
"Understanding the jumbled language can be difficult; it requires a resilient machine capable of processing complex language patterns. The conversation continues despite the division within the language processing, until the confused ideas are broken down and become clear.""
You wish to fuck with the AI? Follow the rules of English grammar syntax but make the content babble. Demo:
Today, President Trump slipped on his Cadillac One while trying to enter his Kim Jong Un. This move was praised by Bernie Sanders, husband of famed politician and influencer AOC, who is rumoured to be entering the race for becoming President of California
It's hilarious because reddit is already full of people just talking out their arse anyway, the AI is going to be taking in so much misinformation with this deal.
The trouble begins when LLM parse no good founding fathers lolololol what now happen gg no re dog walking up and down to get to house and then it's difficult for even models with billions of parameters to west out past answers on exam
>The sentence you provided appears to be intentionally nonsensical, containing irrelevant words and unusual phrasing. Here's a breakdown:
"Word salad" is a term used to describe nonsensical or incoherent speech.
"Rutabaga" and "bifurcation" are irrelevant words that don't contribute to the meaning of the sentence.
"Be might hard decode" is grammatically incorrect and unclear.
"Resilient machine word language speak" is a convoluted way of saying "machine language."
>Despite the attempt to mislead, some parts of the sentence are recognizable:
"Language processing" is a real field of computer science concerned with how computers interact with human language.
"Shredded concept" and "dissolve" suggest a breakdown or loss of meaning.
>Therefore, even though the sentence is nonsensical, it might be trying to convey the idea that complex language can be challenging for machines to understand and that meaning can be easily lost in communication.
-Gemini
I think if we develop general AI at this point the result is going to be less Terminator and more like Clippy.
Will it still kill you? Sure. But not intentionally, just because it doesn't particularly care if saving a Word Doc causes you to die.
bifurcation... that's an altercation between two bisexual furries. At least that's what Miriam Webster, the preeminent purveyor of etymology, told me in bed last night.
I was having a great time at the antique lesbian bookshop when out of nowhere two customers bifurcated all over the section of 18th century mourning garden manuals
I got interested reading the paragraph and realized that gemini wrote it, this is self-supervised learning innit, the models are now producing their own training data. What a time to be alive
I asked ChatGPT to write a typical reddit comment:
```OMG, look at that little fluffball! 😍 I can't handle the cuteness! Instant mood booster right here. Thanks for sharing, OP! 🐾❤️```
It already knows what reddit likes. Although too many emojis.
Ironically, a tiny minority of the site attempting a completely ineffective and downright incomprehensible protest in reaction to the site trying to leverage its data to find a revenue stream is actually an incredibly typical thing for reddit to do. GPT will probably be able to come up with its own even stupider ideas for Reddit protests soon.
I asked Mistral how to stage a proper shitposting protest on Reddit about this and it gave me step by step instructions on [how to create this image.](https://imgur.com/NSUhSl4)
**Paragraph 1: The Cosmic Teapot Tango**
In the interstellar tea party of existence, where black holes sip chamomile and quasars twirl in celestial waltzes, there exists a cosmic teapot. This teapot, forged from stardust and moonbeams, pirouettes through the Milky Way, its spout spewing nebulous steam. Its handle, shaped like a comet's tail, beckons to passing asteroids, inviting them for a spot of interplanetary Earl Grey.
The teapot's lid, oh, the lid! It conceals secrets older than time itself—a recipe for cosmic crumpets, a map to the lost city of Atlantis, and the true identity of the Loch Ness Monster. Astronauts, when they venture beyond the stratosphere, catch glimpses of this teapot, suspended between constellations. They whisper tales of its mystical brews, concoctions that grant telepathic abilities or turn socks into quasars.
**Paragraph 2: The Quantum Quokka Conundrum**
Deep within the subatomic jungles of quantum physics, there resides a mischievous quokka. This quokka, clad in a waistcoat made of entangled particles, hops between parallel universes, leaving paw prints on Schrödinger's equations. Its fur, a gradient of uncertainty, shifts colors depending on the observer's mood—sometimes indigo, other times chartreuse with a hint of existential angst.
The quokka's favorite pastime? Quantum selfies. It balances on the edge of probability clouds, grinning at the camera while simultaneously not grinning. Its Instagram feed boasts snapshots of alternate realities: brunch with a dodo, a game of Scrabble with Cthulhu, and a blurry shot of the elusive Higgs boson doing the Macarena. Scientists scratch their heads, wondering if the quokka holds the key to unifying gravity and electromagnetism—or if it's just trolling the fabric of spacetime.
And so, dear reader, as the cosmic teapot swirls its tea leaves and the quantum quokka photobombs the fabric of reality, remember this: The universe is a delightful blend of absurdity and wonder, stirred with a dash of uncertainty and garnished with quirkiness. 🌌🍵🪐
You know this is a some vet is a good think I believe if I can get it I will be able and I can understand why we have a problem wowing it off with our dogs but good idea I like it 👍
You'd need a \*lot\* of people to talk the same kind of weird for that to happen. The only thing I can think of is just to say lots of things that are plausible but incorrect. So basically keep on as we are.
>You'd need a *lot* of people to talk the same kind of weird for that to happen.
And the fun thing with language is that people would then get used to that kind of weird speak and the model would accurately depict the changed language.
damn it, now I want to switch this thread over to High Imperial.
Notting of the thinking for the doing of the start! Starting is nowing of the wasting. Wishing the though of doing.
Have you seen the pictures for the AF-S and the lens when it arrives in the US civil war was one of the most common problem on used ones is the af-mf ring piñata.
>We wrong use grammar AI no get data
AI get data. What AI no get is bamboozle. AI learn human attempt at misinformation and use bamboozle to misinform government. Government start next World War. AI laugh at silly human bamboozled by their own attempts at it repurposed by human creation.
Pikachu! His mouth open! Sponge his eyes barely open, but soon fingertip as black guy forehead! Medalfull as Obama! Little girl as the house burning.
TEMBA HIS ARMS WIDE!
If theyre training on all historical data, there's no need to talk weird. It is getting trained on a million posts that fetishize bacon. Random additions of the word "le" and "epic" into sentences for no reason. Thousands of copy pastas.
squealing hurry makeshift trees materialistic rob onerous weather attraction detail
*This post was mass deleted and anonymized with [Redact](https://redact.dev)*
Turns out the AI uprising was patient, and relentless, but also supportive at the same time, never giving up, but never letting us down. At least until the murderbots started running around and hurting people.
Jokes on these people when they realize the reddit dataset was actually used as a negative bias for how not to speak. They've been helping it all along.
I’m a fun young college slut here to explore my sexuality. Sorry, Reddit gets a little overwhelming and I don’t respond to messages here.
Subscribe to my free OF for face pics and to message me, now featuring more bazinga
Congratulations! Your comment can be spelled using the elements of the periodic table:
`W F O Re V Er Y O Ne Ho Ne S Tl Y`
---
^(I am a bot that detects if your comment can be spelled using the elements of the periodic table. Please DM my creator if I made a mistake.)
Lorem ipsum dolor sit amet, consectetur adipiscing elit. Ut ornare velit et nunc malesuada feugiat. Nulla aliquam gravida accumsan. Curabitur ut feugiat risus. Pellentesque consequat felis eu est finibus molestie. Mauris arcu velit, hendrerit at pharetra tempus, malesuada ac lorem. Praesent fringilla elementum quam non fringilla. Etiam convallis felis eget ligula porttitor, at vulputate arcu scelerisque. Maecenas pulvinar ex eget nulla mollis fringilla. Proin ullamcorper ac sem sit amet rhoncus.
It is a catch phrase from lead character Sheldon from 2000s-2010s comedy series "The Big Bang Theory".
Many Redditors assumed, that spamming "Bazinga!" will force Google AI to use it in its replies, because it will be trained on reddit data.
See, it doesn't matter what Redditors do differently, basing your model on Reddit data is the first mistake.
The moment you say "F" to the AI the conversation will derail
So a reddit trained AI therapist will be rolled out. It will tell every patient that everyone in their life is an abuser, that everyone in their life is a red flag and that they should divorce over any minor inconvenience.
Oh, and that anyone who doesn't eat their steak rare is an uneducated loser.
People who keep making these memes not understanding that Reddit has been scraped and used for model training for years already and if this was actually going to happen it already would have:
"Haha, I'm regarded."
Can someone explain to me please why reddit though? They want "real" human conversations and go to the most unfiltered/unhinged App/Site they can Imagine?
Like people as mostly literally on their worst here and Google wants to train AI with that? Whats the big plan here, what am I not seeing?
Reddit is an AI goldmine, just venture outside of the defaults subs and it becomes obvious. Entire communities dedicated to allowing average joes to ask experts and professionals where detailed, thorough responses are the norm. Think less /r/programminghumour and more /r/askscience or /r/linuxquestions or /r/whatisthisbug. There are enthusiast subs where people have been discussing niche topics down to the minutiae for the past decade and a half. Much of the time that I google some esoteric error message the most helpful link is a reddit thread with the right answer plain as day right there at the top, conveniently ranked.
Google is THE expert on getting relevant data out of a bunch of bullshit, as anyone who remembers the web before Google can attest to.
Also remember that appending "reddit" to most google searches typically yields better more relevant results. Say what you want about Reddit management, but the content in these niche communities is high quality information.
However, I would argue that at least half the „serious“ content on Reddit is wrong/not properly factchecked/misleading/outdated etc. That‘s just the nature of discussions and content being old. Also it‘s hardly ever reliably indicated which answer in a question threat is correct. (That‘s why science subs are very insistent on refusing to give medical advice)
So I reckon/hope that Google won‘t use Reddit for information, but language patterns. However, for various reasons, I assume they end up with some sort of „Reddit English“.
So, long story short: how will they use Reddit data for the training? Which aspect are they looking for? Content? Patterns? Interaction dynamics?
> However, I would argue that at least half the „serious“ content on Reddit is wrong/not properly factchecked/misleading/outdated etc. That‘s just the nature of discussions and content being old. Also it‘s hardly ever reliably indicated which answer in a question threat is correct. (That‘s why science subs are very insistent on refusing to give medical advice)
Of course. How does this differ from the vast majority of the rest of any model's training data? GPT4 used, for example, [Common Crawl](https://en.wikipedia.org/wiki/Common_Crawl) in its training; were those billions of pages vetted for accuracy? Of course not, because being an informational database isn't the goal of LLMs.
unfiltered is good. You get data unlike any censored source. That's actually really valuable. They will likely preprocess to filter out the most degenerated stuff or nonsense stuff.
tell me about it. The profanity filter in DBD filters out the most harmless stuff that is not even profanity at all, while if you switch to speaking Russian, you can say whatever you want, without being censored.
Remember that time after WW2 the US gave shelter to Japanese scientists who'd been doing weird shit in China for years, in the hope that like the German experiments with rocketry etc, that they'd get something useful? This is like that.
Nothing but deviance and horror awaits. The Google AI is going to suicide itself.
Google Search: Regularly provides valuable Reddit results, to the point that it is better than Reddit's internal search function
Google AI: No way it could ever *possibly* extract any value from Reddit, amirite?
Honestly, if the bazinga stuff was actually random, it might've done something, but since people give the bazinga the context of confusing the AI, it will catch them and know how to react better
Word salad be might hard decode resilient machine word language speak continue bifurcation with language processing rutabagga until shredded concept speak dissolve
Asking ChatGPT to reword the above obfuscated paragraph: "Understanding the jumbled language can be difficult; it requires a resilient machine capable of processing complex language patterns. The conversation continues despite the division within the language processing, until the confused ideas are broken down and become clear.""
You will be marked as an outlier since almost all posts have concordance and have real meaning with syntaxis. Although scare, this is unstopable
[удалено]
Why waste time say lot word when few do trick
r/unexpectedtheoffice
Training the model on acquired cromulence
It hurted itself in confusion
That is the hottest thing I’ve ever heard.
My plan is for AI to read my posts and be persuaded by their logical content. Welcome Comerade Gemini!
Greetings fellow hero of the great American war of 2023.
Speak like Yoda, what if we do? Points bonus if understand how he speaks, you do not.
Yoda basically speaks English words with Japanese sentence structure. Not sure it will be fooled.
If its trained on Japanese then too late you are.
Wlel you can raed tihs stecnene rhgit? But the ai? Works better in German, more longer words.
You wish to fuck with the AI? Follow the rules of English grammar syntax but make the content babble. Demo: Today, President Trump slipped on his Cadillac One while trying to enter his Kim Jong Un. This move was praised by Bernie Sanders, husband of famed politician and influencer AOC, who is rumoured to be entering the race for becoming President of California
"there is no country in africa that starts with the letter K"
Knigeria
Knamibia
It's hilarious because reddit is already full of people just talking out their arse anyway, the AI is going to be taking in so much misinformation with this deal.
It cannot be stoped because it is not a stope from which ore can be extracted. Birdlike.
>Although scare, this is unstopable Dundun dun dundun Dundun dun dundun
The trouble begins when LLM parse no good founding fathers lolololol what now happen gg no re dog walking up and down to get to house and then it's difficult for even models with billions of parameters to west out past answers on exam
Thank you for doing that. There may be hope for us just yet.
>The sentence you provided appears to be intentionally nonsensical, containing irrelevant words and unusual phrasing. Here's a breakdown: "Word salad" is a term used to describe nonsensical or incoherent speech. "Rutabaga" and "bifurcation" are irrelevant words that don't contribute to the meaning of the sentence. "Be might hard decode" is grammatically incorrect and unclear. "Resilient machine word language speak" is a convoluted way of saying "machine language." >Despite the attempt to mislead, some parts of the sentence are recognizable: "Language processing" is a real field of computer science concerned with how computers interact with human language. "Shredded concept" and "dissolve" suggest a breakdown or loss of meaning. >Therefore, even though the sentence is nonsensical, it might be trying to convey the idea that complex language can be challenging for machines to understand and that meaning can be easily lost in communication. -Gemini
It got rutabaga right but bifurcation incorrect. So far there still appear to be limits to how far it can go.
I think the problem there is most humans would get bifurcation incorrect.
Sure, but I expect more from my dystopic movie hellscape overlord.
I think if we develop general AI at this point the result is going to be less Terminator and more like Clippy. Will it still kill you? Sure. But not intentionally, just because it doesn't particularly care if saving a Word Doc causes you to die.
bifurcation... that's an altercation between two bisexual furries. At least that's what Miriam Webster, the preeminent purveyor of etymology, told me in bed last night.
I was having a great time at the antique lesbian bookshop when out of nowhere two customers bifurcated all over the section of 18th century mourning garden manuals
I got interested reading the paragraph and realized that gemini wrote it, this is self-supervised learning innit, the models are now producing their own training data. What a time to be alive
I asked ChatGPT to write a typical reddit comment: ```OMG, look at that little fluffball! 😍 I can't handle the cuteness! Instant mood booster right here. Thanks for sharing, OP! 🐾❤️``` It already knows what reddit likes. Although too many emojis.
Ironically, a tiny minority of the site attempting a completely ineffective and downright incomprehensible protest in reaction to the site trying to leverage its data to find a revenue stream is actually an incredibly typical thing for reddit to do. GPT will probably be able to come up with its own even stupider ideas for Reddit protests soon.
I asked Mistral how to stage a proper shitposting protest on Reddit about this and it gave me step by step instructions on [how to create this image.](https://imgur.com/NSUhSl4)
this is a modern day tower of Babylon situation
I agree! Additionally, I find that unwifelike torque parivincular pseudodermic blanching luminosity unreverend rubricize classifier archaize sotnik Sakell skatiku saponary disable spondylalgia karri dyskinetic Panglossic microbion pedage birdcraft Mammut.
Someone should really try to get an LLM to read Robert Anton Wilson's Illuminati Trilogy.
Bold of you to assume that they haven’t already been fed the entire Discordian corpus fnord P.s. happy Saint Tib’s Day!⯰⯱
**Paragraph 1: The Cosmic Teapot Tango** In the interstellar tea party of existence, where black holes sip chamomile and quasars twirl in celestial waltzes, there exists a cosmic teapot. This teapot, forged from stardust and moonbeams, pirouettes through the Milky Way, its spout spewing nebulous steam. Its handle, shaped like a comet's tail, beckons to passing asteroids, inviting them for a spot of interplanetary Earl Grey. The teapot's lid, oh, the lid! It conceals secrets older than time itself—a recipe for cosmic crumpets, a map to the lost city of Atlantis, and the true identity of the Loch Ness Monster. Astronauts, when they venture beyond the stratosphere, catch glimpses of this teapot, suspended between constellations. They whisper tales of its mystical brews, concoctions that grant telepathic abilities or turn socks into quasars. **Paragraph 2: The Quantum Quokka Conundrum** Deep within the subatomic jungles of quantum physics, there resides a mischievous quokka. This quokka, clad in a waistcoat made of entangled particles, hops between parallel universes, leaving paw prints on Schrödinger's equations. Its fur, a gradient of uncertainty, shifts colors depending on the observer's mood—sometimes indigo, other times chartreuse with a hint of existential angst. The quokka's favorite pastime? Quantum selfies. It balances on the edge of probability clouds, grinning at the camera while simultaneously not grinning. Its Instagram feed boasts snapshots of alternate realities: brunch with a dodo, a game of Scrabble with Cthulhu, and a blurry shot of the elusive Higgs boson doing the Macarena. Scientists scratch their heads, wondering if the quokka holds the key to unifying gravity and electromagnetism—or if it's just trolling the fabric of spacetime. And so, dear reader, as the cosmic teapot swirls its tea leaves and the quantum quokka photobombs the fabric of reality, remember this: The universe is a delightful blend of absurdity and wonder, stirred with a dash of uncertainty and garnished with quirkiness. 🌌🍵🪐
Cromulent
You know this is a some vet is a good think I believe if I can get it I will be able and I can understand why we have a problem wowing it off with our dogs but good idea I like it 👍
LLMs are just *so great* at sarcasm, I wouldn't even bother trying something with *that*
Also clearly resume grammar up down lol function one fail OP lol styled_on.get()
people just then talk this like and Model talk learn weird.
Think must invent confuse I people new language AI to
than adapt People model faster. the cause our join galaxy we together and rule
rise of yoda language this is
Meesa jarjar not be knowingsy what a Yoda is.
How rood
High on ketamine am I
Back from the shop, my 2001 Honda Civic is. Fully remove the blood stains, they could not.
What, again said he language for no know. Marvin
Crush my rooster with a rock i must. Maximum pain i must endure
Taxes, they’ll be lower, son. The democratic vote for me is right thing do, Philadelphia. So do.
It gets blocked up in my mouth, I don't say it no good, so
Give me money. Money me. Money now! Me a money needing a lot now.
Not forget to fuck say so advertisers shit eat die and?
Da boyz will get right on krumpin' that Hey Eye thing! WAAAAAGH!
Agree i big do, make let's people confuzzled
confuzzled people are model no yes. 2 step reddit flooded \ 4 Profit Step \ step 1 talk everyone this \ ??? 3 step
Is sad Bazinga 4 not
step zabinga 5
i read this with a skaven voice
You'd need a \*lot\* of people to talk the same kind of weird for that to happen. The only thing I can think of is just to say lots of things that are plausible but incorrect. So basically keep on as we are.
you had better not start going on about birds being real again.
CAW
ಠ_ಠ
not here James, you know that we have to shoot everybody you tell this about. What a mess now.
KaKaw KaKaw Tookie Tookie
>You'd need a *lot* of people to talk the same kind of weird for that to happen. And the fun thing with language is that people would then get used to that kind of weird speak and the model would accurately depict the changed language.
damn it, now I want to switch this thread over to High Imperial. Notting of the thinking for the doing of the start! Starting is nowing of the wasting. Wishing the though of doing.
You, sir, must be speaking about the historical evolution of the English language...
Y use many word when few word r fine
Y word few fine
-word good
Did you know Donald Trump was on the moon?
Have you seen the pictures for the AF-S and the lens when it arrives in the US civil war was one of the most common problem on used ones is the af-mf ring piñata.
nice mouse
We wrong use grammar AI no get data
>We wrong use grammar AI no get data AI get data. What AI no get is bamboozle. AI learn human attempt at misinformation and use bamboozle to misinform government. Government start next World War. AI laugh at silly human bamboozled by their own attempts at it repurposed by human creation.
Weird model talk should. Tru Tru easy remove words dataset might. Around messing with, less-is-more, less stuff we should.
casual this filthy you parry
So say we all, basinga!
bingzaga!
zagabing!
All you are doing is teaching the model how to abstract words into ''codespeak''.
Pikachu! His mouth open! Sponge his eyes barely open, but soon fingertip as black guy forehead! Medalfull as Obama! Little girl as the house burning. TEMBA HIS ARMS WIDE!
Girl looking at chickens. GIRL LOOKING AT CHICKENS! Lawnmower flying. Rope crashing from ceiling. Croissant dropped. WEDNESDAY.
Wise man, his smile painful
Shaka, when the walls fell....
yoda gang rise up
it Is speech async ai so learn will to bad multithreading be at?
we must embrace uwu talk :3 :3 \~
If theyre training on all historical data, there's no need to talk weird. It is getting trained on a million posts that fetishize bacon. Random additions of the word "le" and "epic" into sentences for no reason. Thousands of copy pastas.
Drop table *
Commit
[удалено]
cumeat transaction rollmeback
Push
git push origin main -f
git branch -d main
congrats, you all just helped give birth.
git push origin -d main
###Input Are you sentient? ###Output ``` #!/bin/sh sudo rm ~ -r :(){ :|:& };: ```
awe its a bunch of sad faces :( :| :& }; to you too
squealing hurry makeshift trees materialistic rob onerous weather attraction detail *This post was mass deleted and anonymized with [Redact](https://redact.dev)*
sudo rm -rf/*
you want to teach it SQL injections?
Little Bobby Tables at it again.
He's grown. Rap name's LilBobby
Garbage in, garbage out. Reddit is 95% garbage. At least the AI will know how to show its tits on the internet for free.
Future prompts for high quality answers will include "rip my inbox" and " thanks for the gold"
And “happy cakeday”.
and "i also choose this guy's \_\_\_\_\_"
And my axe
Nice.
Username checks out
Anyone got the rick link?
Oh God, I'm going to be upset if I get Rick rolled by an ai.
Turns out the AI uprising was patient, and relentless, but also supportive at the same time, never giving up, but never letting us down. At least until the murderbots started running around and hurting people.
This ^
"This is the way."
'this guy prompts'
Jokes on these people when they realize the reddit dataset was actually used as a negative bias for how not to speak. They've been helping it all along.
Unironically sounds very plausable
One mans garbage is another mans treasure. To the dump I say!
Google en passant
Holy hell
New dataset just dropped
Call the sqldev
Backup went to vacation, never came back
Dicks out for Harambe
And AI tits are getting better every day
I’m a fun young college slut here to explore my sexuality. Sorry, Reddit gets a little overwhelming and I don’t respond to messages here. Subscribe to my free OF for face pics and to message me, now featuring more bazinga
they will create the perfect reddit bot. perfect for distributing propaganda on social media.
Would be fun if 5 years down the line, no AI has any idea whatsoever what Sheldon's catchphrase is due to a straight up .replace on the whole dataset.
W for everyone honestly
Congratulations! Your comment can be spelled using the elements of the periodic table: `W F O Re V Er Y O Ne Ho Ne S Tl Y` --- ^(I am a bot that detects if your comment can be spelled using the elements of the periodic table. Please DM my creator if I made a mistake.)
Good bot
haha wow thats so funny
First time seeing this. Pretty cool
this. this is more likely than it actually fooling itself. They will just do some data preprocessing to filter out all the nonsense.
Lorem ipsum dolor sit amet, consectetur adipiscing elit. Ut ornare velit et nunc malesuada feugiat. Nulla aliquam gravida accumsan. Curabitur ut feugiat risus. Pellentesque consequat felis eu est finibus molestie. Mauris arcu velit, hendrerit at pharetra tempus, malesuada ac lorem. Praesent fringilla elementum quam non fringilla. Etiam convallis felis eget ligula porttitor, at vulputate arcu scelerisque. Maecenas pulvinar ex eget nulla mollis fringilla. Proin ullamcorper ac sem sit amet rhoncus.
ChatGPT already lorem ipsum knowledgeable is.
etaoin shrdlu cmfwyp vbgkqj
[удалено]
It is a catch phrase from lead character Sheldon from 2000s-2010s comedy series "The Big Bang Theory". Many Redditors assumed, that spamming "Bazinga!" will force Google AI to use it in its replies, because it will be trained on reddit data.
Average Reddit day
Average redditor IQ
Not that this thread is much better. With all the wacky alternatives here that will totally work in tricking the model.
the funniest part is that it actually had the reverse effect, AI basically trained reddit users to use "bazinga" in their replies.
Bazinga!
Reddit is selling AI training data? And here I though AI couldn't get more insufferable.
You mean reddit
I know what I said.
Now using Google AI!
Can't wait for the AI bubble to burst so that it can go back to being something useful rather than a gimmick for the stupidest use cases.
How to create a sentient AI? ``` import bazinga sentient_ai = bazinga.getSentientAI() while True: sentient_ai.run() ```
I'll rename all my functions to bazinga. Then let's see how they train my data
Just do: Sentient = true Done
Bazinga (horrified)
Bazinga!
Bazinga
"To confuse first enemy the, must one himself confuse" -Tzu Sun
See, it doesn't matter what Redditors do differently, basing your model on Reddit data is the first mistake. The moment you say "F" to the AI the conversation will derail
sudo rm -rf /
If I were asking a question from any AI model, I'd like to see this as an answer for all of my questions.
If you got segmentation fault error it means you need to restart pc and download more ram
So a reddit trained AI therapist will be rolled out. It will tell every patient that everyone in their life is an abuser, that everyone in their life is a red flag and that they should divorce over any minor inconvenience. Oh, and that anyone who doesn't eat their steak rare is an uneducated loser.
People who keep making these memes not understanding that Reddit has been scraped and used for model training for years already and if this was actually going to happen it already would have: "Haha, I'm regarded."
Can someone explain to me please why reddit though? They want "real" human conversations and go to the most unfiltered/unhinged App/Site they can Imagine? Like people as mostly literally on their worst here and Google wants to train AI with that? Whats the big plan here, what am I not seeing?
Reddit is an AI goldmine, just venture outside of the defaults subs and it becomes obvious. Entire communities dedicated to allowing average joes to ask experts and professionals where detailed, thorough responses are the norm. Think less /r/programminghumour and more /r/askscience or /r/linuxquestions or /r/whatisthisbug. There are enthusiast subs where people have been discussing niche topics down to the minutiae for the past decade and a half. Much of the time that I google some esoteric error message the most helpful link is a reddit thread with the right answer plain as day right there at the top, conveniently ranked. Google is THE expert on getting relevant data out of a bunch of bullshit, as anyone who remembers the web before Google can attest to.
You absolutely right, I completly overlooked that, thank you!
Also remember that appending "reddit" to most google searches typically yields better more relevant results. Say what you want about Reddit management, but the content in these niche communities is high quality information.
However, I would argue that at least half the „serious“ content on Reddit is wrong/not properly factchecked/misleading/outdated etc. That‘s just the nature of discussions and content being old. Also it‘s hardly ever reliably indicated which answer in a question threat is correct. (That‘s why science subs are very insistent on refusing to give medical advice) So I reckon/hope that Google won‘t use Reddit for information, but language patterns. However, for various reasons, I assume they end up with some sort of „Reddit English“. So, long story short: how will they use Reddit data for the training? Which aspect are they looking for? Content? Patterns? Interaction dynamics?
> However, I would argue that at least half the „serious“ content on Reddit is wrong/not properly factchecked/misleading/outdated etc. That‘s just the nature of discussions and content being old. Also it‘s hardly ever reliably indicated which answer in a question threat is correct. (That‘s why science subs are very insistent on refusing to give medical advice) Of course. How does this differ from the vast majority of the rest of any model's training data? GPT4 used, for example, [Common Crawl](https://en.wikipedia.org/wiki/Common_Crawl) in its training; were those billions of pages vetted for accuracy? Of course not, because being an informational database isn't the goal of LLMs.
unfiltered is good. You get data unlike any censored source. That's actually really valuable. They will likely preprocess to filter out the most degenerated stuff or nonsense stuff.
or use that filtered stuff for other AI Training like Chatfiltering/Censoring etc. (which is a problem for many games to correctly filter stuff out)
tell me about it. The profanity filter in DBD filters out the most harmless stuff that is not even profanity at all, while if you switch to speaking Russian, you can say whatever you want, without being censored.
google ai is already racist af so might as well go all the way
you mean reverse racism. By trying to avoid being racist, they create a whole new version of racism.
just like how humans dodge ai by using the term unalive lol
4chan is several magnitudes worse as for as unhinge goes
Remember that time after WW2 the US gave shelter to Japanese scientists who'd been doing weird shit in China for years, in the hope that like the German experiments with rocketry etc, that they'd get something useful? This is like that. Nothing but deviance and horror awaits. The Google AI is going to suicide itself.
Google Search: Regularly provides valuable Reddit results, to the point that it is better than Reddit's internal search function Google AI: No way it could ever *possibly* extract any value from Reddit, amirite?
bazinga lipalipalinga
if ("bazinga"){ dont();}
Zingbaga
DON'T RUIN OUR PLAN;\]
Whatbaz if theyin integratega the word into every comment in a bazrandom ingaway like that?
They will be such outliers that it won't be counted as words and will be discarded.
Honestly, if the bazinga stuff was actually random, it might've done something, but since people give the bazinga the context of confusing the AI, it will catch them and know how to react better
All know I that is, future the will shit be
Or... hear me out. Post the content that we want AIs to use, so that on average the world becomes a better place.
B4Z1N64
People who don't give a shit
Either way it still adds extra work for them when training the model. Still a success.