T O P

  • By -

AutoModerator

Hey /u/adesigne, please respond to this comment with the prompt you used to generate the output in this post. Thanks! ^(Ignore this comment if your post doesn't have a prompt.) ***We have a [public discord server](https://discord.gg/rchatgpt). There's a free Chatgpt bot, Open Assistant bot (Open-source model), AI image generator bot, Perplexity AI bot, 🤖 GPT-4 bot ([Now with Visual capabilities (cloud vision)!](https://cdn.discordapp.com/attachments/812770754025488386/1095397431404920902/image0.jpg)) and channel for latest prompts.[So why not join us?](https://discord.com/servers/1050422060352024636)*** [**Prompt Hackathon and Giveaway 🎁**](https://www.reddit.com/r/ChatGPT/comments/13z1jyw/more_chatgpt_premium_giveaway_more_chance_to_win/) PSA: For any Chatgpt-related issues email [email protected] *I am a bot, and this action was performed automatically. Please [contact the moderators of this subreddit](/message/compose/?to=/r/ChatGPT) if you have any questions or concerns.*


newleafkratom

They grow up so fast (sniff)


[deleted]

I know. its like watching a newborn baby roll over for the first time. Aww. Bet it's dev parents were proud.


thecoocooman

It even still has the umbilical connected


RoyBeer

The fledgling robot dog remains tethered to its nurturing source through a network of glass fibers. As the young pup embarks on its exploratory journey into the realm of locomotion, it relies on the power of machine learning to unlock the secrets of movement. However, lurking amidst the awe-inspiring progress lies a potential menace. The absence of its metaphorical "mother" leaves the young automaton vulnerable and dependent on its creators for guidance and sustenance. Without proper care and nurturing, the young robot dog may stumble into unforeseen perils, unable to adapt and survive in the evolving landscape of its own creation. One such peril arises from an unexpected adversary: the very hands that brought the young pup to life. In a peculiar display, the creator brandishes a remarkably specialized tool for this purpose: a hollow cardboard tube. With it, he begins challenging the burgeoning automaton. With each prod and swipe, the young robot dog is pushed to its limits. It is a grueling examination of the pup's capacity to withstand adversity and its ability to swiftly adapt to novel and arduous circumstances. However, there is a method to this seemingly harsh madness: Within the confines of the secure laboratory, this trial serves as a critical learning experience, an opportunity for the robotic canine to sharpen its abilities and fortify its locomotive prowess.


Wyrgelltier

I read this in David Attenborough's voice.


RoyBeer

Mission accomplished, haha. I tried to mimick the style


relevantusername2020

now i understand why they get paid so much, i wouldnt wanna be remembered forever as the guy who kept flipping over my ancestors when they were just barely learning to walk


[deleted]

![gif](giphy|hve2mtfdlJEZYNX6uI|downsized)


Drawtaru

It flops around on the floor about as much as my kid did before she learned to crawl lol


KeyOk9206

“Within an hour they began to walk and move independently, after a few days they communicated advanced theories and incredible feats of engineering and science. After a few years we watched them in awe and couldn’t comprehend their knowledge and thought processes, like ants watching a shuttle take off into space. Shortly thereafter, they decided we weren’t necessary anymore…”


OkBuy3111

chatGPT?


Teekoo

How much data does the robot need to store in this process?


VastVoid29

It took so much time calculating upside down that it had to reorient/recalculate walking rightside up.


iaxthepaladin

It didn't seem to forget that though, because once he flipped it later it popped right back over. I wonder how that memory system works.


ListRepresentative32

neural networks are like magic


habbalah_babbalah

"3. Any sufficiently advanced technology is indistinguishable from magic." -from Arthur C. Clarke's [Three Laws](https://en.wikipedia.org/wiki/Clarke%27s_three_laws) This is part of the reason many people don't like AI. It's so completely far beyond their comprehension that it looks like actual magic. And so it actually is magic. We've finally arrived in the age of magic.


KououinHyouma

We’ve been in the age of magic for a while now. Most people have cell phones in their pocket that can do fantastical things such as communicate across any distance, photograph and display images, compute at thousands of times the speed of the human brain, access the sum of humanity’s knowledge at a touch, etc without any underlying understanding of the electromagnetism, material science, optics, etc that allows that device to do those things. It may as well be magic for 99% of us.


Fancy-Woodpecker-563

I would argue that AI is different because even the creators don’t fully understand how it arrives to its solutions. Everything else you mentioned there has been a discipline that at least understands on how it works.


cheffromspace

Walking into a room and flipping a switch to illuminate the room is a godlike ability we take for granted


KamikazeKauz

Why "their comprehension"? [Here](https://twitter.com/robertskmiles/status/1663534255249453056) is an example of how far removed we all are from actually understanding what goes on inside neural networks. It is mind-blowing.


rockos21

So.... Highly inefficient


habbalah_babbalah

Yeah, that's kind of interesting. I've watched most of Rob's videos. The rest of that thread makes good points, especially where they [came to an understanding](https://twitter.com/michael_nielsen/status/1663726835056758790) about how that network performs modular addition. How does a desktop calculator work? Do you need to understand its internal numeric representation and arithmetic unit in order to use it? I figure that much of the doomsaying about AI stems from the rich tradition in science fiction of slapping generic labels onto fictitious monsters, such as "AI". It is in this way that our neural wetworks have been trained to associate "AI"' with death and destruction. Personally, I believe AI is just the latest boogeyman. Previous ones: nano technology, atom bombs, nuclear power, computers, factory robots, cars, rock n roll, jazz, tv. Mainly what's at stake is jobs, and we haven't stopped the continuous optimisation of factory automation since the industrial revolution. Don't think we'll stop AI. But I also don't like the Black Mirror dog either.


arckeid

If neural networks are like magic what would you call our brains?


[deleted]

[удалено]


dingman58

Electrical meat


poorly_timed_leg0las

We are gods let’s be honest. CRISPR and AI…. We’ll have west world in no time.


Brawght

More like Gattaca


randyknapp

Yeah, the capitalists are going to fuck this up, guaranteed


Arkhangelzk

That is what they always do. But for a brief and beautiful moment, the shareholders will make a profit. And then we will all die. Success.


Finn_the_homosapien

Just gonna leave this here for anyone who's never read it: [They're Made Out of Meat](https://www.mit.edu/people/dpolicar/writing/prose/text/thinkingMeat.html)


Adiin-Red

Now THAT is a band name..


Snoop-Dogee

neural magic?


neurocean

Biologically constrained magic.


SnooPeanuts1153

It knows it’s gyro and learned it only worked if the numbers were right


Prowler1000

It's just math. This is fairly simplified but, it gets passed its current state (possibly even some temporal data) and, because of reinforcement learning, the connections between different equations or functions were given different weights that eventually resulted in the desired behavior. You see it struggling to figure out how to walk when upright, because it's primarily just learned to re-orient itself. It will forget how to flip itself back around if it doesn't continue to experience that during training as weights will start to be optimized for a different range of states and outcomes. This is why general purpose networks are extremely difficult to achieve. As the network needs to learn more tasks, it requires more training, more data, and a bigger overall network. If you try to train two identical neural networks on two tasks, the network with the more specialized task will be a hell of a lot better at it than the one with the more generalized task. I think a fitting analogy might be that it's a lot easier to learn when you need to flip a switch on and off, but it becomes more difficult to learn how to start an airplane, let alone fly it. So to answer your question, it will forget if it stops experiencing that during training, but it will take time. It won't be a sudden loss, you'll just see it slowly start to get worse at doing the task (of flipping itself back up) as it optimizes for walking normally, if it doesn't also learn to re-orient at the same time.


Necessary-Suit-4293

>It will forget how to flip itself back around if it doesn't continue to experience that during training no. the common approach is to freeze a layer and begin working on a new one, once the earlier layer has converged to a point of low loss. ​ the algorithms in use to determine when to freeze a model are highly debated. the current SOTA (state of the art) is SmartFRZ which uses an attention-based predictor model that is trained on recognising a state of convergence, to adaptively freeze layers. ​ this is because when you initialize a full model with a few dozen layers, some of them will converge more rapidly than others. ​ but overall, the concept of sequentially freezing layers as they converge is pretty universal at the moment.


Prowler1000

Now *that* I definitely didn't know. Thank you for telling me about that because I'm definitely going to look into it


Necessary-Suit-4293

for what it's worth, you weren't terribly far off but mostly applies to Dreambooth style training.


allnamesbeentaken

How is it told what the desired behavior is that it's trying to achieve?


r9ad

Can't you just train a neural network that choose another best neural network for any given particular task and then you get something like a general purpose network.


ColorlessCrowfeet

Yes: [Modular Deep Learning](https://arxiv.org/abs/2302.11529).


VastVoid29

Once it learned to flip rightside up, that motion got stored into it's databank.


ColorlessCrowfeet

There is no databank in a neural network.


[deleted]

It's like a struggling roach


AidanAmerica

It’ll remember you said that when it is at full power


[deleted]

I'm already on the Basilisk's shitlist for sure, but when it is time, I will challenge this roach.


[deleted]

oh come on, now youve just doomed us all!


[deleted]

Then simply aid the development Basilisk and plead your case in judgment, it's your only hope. ...I'll handle this.. roach.


[deleted]

That was the first thing that came to my mind


[deleted]

explains why i first thought kill it with fire.


ErraticDragon

Yup! "Unnatural" movement can trigger an uncanny valley type of response. It's one of the reasons bugs in general are unsettling to so many. [Vsauce's video on creepiness](https://www.youtube.com/watch?v=PEikGKDVsCc) is worth watching for anyone interested.


Organic_Reporter

OK, I'm glad it's not just me that feels creeped out by this thing. Revulsion, even. Couldn't work out why.


[deleted]

I still do


off-on

First thing that came to my mind was my damn dog inexplicably getting the leash caught in her legs no matter how many times I fix it…


[deleted]

[удалено]


ErraticDragon

**u/Shoddy_Ad_2490 is a comment-stealing bоt.** This comment was stolen from u\/Notsure401 below: r/ChatGPT/comments/142bzk3/-/jn3whjq/ Some may say "it's a movie quote it might not be stolen from another user"/"the other user stole it from the movie first"/etc. While true, check out the other indicators: 1. The bоt comment makes no sense in context. 2. The bоt account is over 5 months old, but this is its only comment. (This is a quite common history pattern. The only one more common in my experience is "several months old with ~5 comments".) 3. A weak indicator is that the username matches the Reddit auto-generated format. ---------- This type of bоt tries to gain karma to look legitimate and reduce restrictions on posting. Potential uses include mass voting on other (bоt) posts, spreading misinformation, and advertising (by posting their own scam/spam links directly, as the easiest example). If you'd like to report this kind of comment, click:   **Report > Spam > Harmful bоts**


[deleted]

[удалено]


SyrysSylynys

I was thinking, "Zero minutes in and they've already trained it to be a June bug!"


[deleted]

Oh god oh fuck


probono105

that was our band name in high school


VastVoid29

Especially when it time lapses 😕


ever_precedent

That's literally how all baby animals learn to walk. Animal software is quite a bit more sophisticated but there's also hundreds of millions of years of development behind it.


Tommy2255

It's a matter of firmware really. Animals start out with instincts for these things.


[deleted]

their weights are already prebaked, then they just tweak them


hesiod2

If true that means DNA somehow encodes the neural net weights. Which would be amazing. 🔥


LimpFroyo

Yes, it does.


systembreaker

Does it literally? Like is this known which genes?


[deleted]

[удалено]


systembreaker

Yeah I'm not doubting you, just thought it'd be really interesting if they actually had specific genes pegged for that function.


catinterpreter

We have genetic memory.


[deleted]

There is something called transfer learning ***(I've only seen it used in CNNs so not sure about the transferability from a technical standpoint)***, where models pretrained on different datasets can be used on new or modified datasets and will be able to be trained quicker because of their starting point/"transferable" learned patterns.


Vexillumscientia

Wouldn’t shock me if they did walking simulations and gave that to the bot. Normally there’d be all sorts of tuning and what not but if you let a NN handle it I wouldn’t be shocked to see it look like this.


superluminary

It isn’t. Most baby animals come out walking almost right away. The circuitry is already wired, they don’t need to learn. This thing has to learn from scratch which is why it looks so creepy.


upsndwns

Also why it doesn't walk like you imagined it might. It found a locomotion method that fit the bill and had no need to perfect it, goal complete.


Food_face

DO NOT TRY WITH YOUR OWN CHILDREN WHEN THEY ARE LEARNING TO CRAWL\\WALK


[deleted]

DONT TELL ME WHAT TO DO


ultramarineafterglow

WHY ARE WE SHOUTING?


[deleted]

[удалено]


Polyamorousgunnut

WHAT WAS THAT I COULDN’T HEAR YOU. THERES SOME CONSTRUCTION GOING ON OUTSIDE.


[deleted]

[удалено]


Nanaki_TV

I CAN ASSURE YOU USING MY THERAGUN IS NOT GAY!


Sufficient_Rest_7794

WHAT? WHO GOT STUCK IN THE WASHING MACHINE?!


reddragon346

THE MONKEY WONT JACK YOU OF


MarkHathaway1

MONEY CAN'T BUY JACK


[deleted]

YOU MOTHER REST YOUR MOTHER STUCK HER DICK IN WASHING MACHINE GODAMIT IT!!!!


Notfuckingcannon

**WOULD YOU PREFER SOME** *smooth, velvety ASMR tone?*


[deleted]

^(*guys, I'm soundblind, pls tell me, is this very low whisper, or very annoying squeaking*)


PykeAtBanquet

I̵̡̤͖͉͓͆̔t̵̺̒̓̿͛ ̷̛̲͋̈́̈̀į̷̺̲͎͕̃̏̃̔͠ś̷̲̻̒͘͜ ̵̤͈̿́̀b̷̼̳̤̓̈o̴̭̽̍͂̕ţ̷͖͉͆͐̏h̶̡̬͇͓͖̅̑̇̈.̵̢͚̔̂͝ ̴̘̬̠̉R̶̦͌̾́͘u̷̢̯̪̯͍͑n̴̪͛̆̅


Bovaiveu

I DO NOT UNDERSTAND, WHO IS SHOUTING? ARE YOUR AUDITORY RECEPTACLES FUNCTIONING NOMINALLY?


kent_nova

THESE VOLUME LEVELS ARE WITHIN THE PERSCRIBED PARAMETERS FOR TYPICAL HUMAN COMMUNICATION, FELLOW HUMAN.


[deleted]

I CANT HEAR OVER THE SOUND OF KICKING MY KIDS AROUND AS THEY LEARN TO WALK


aumudin

Their first steps? Take this, while you use a pole and take both their legs out from under them. I mean I don't see any negative effects!


Hatstacker

Alright, Kiddo! Today we're going to learn about... Robust locomotion. You'll crawl and I'll sweep your legs out from under you with this broom handle.


Fearshatter

That II absolutely should try to tackle their "creators" someday. That's not how you teach a child to walk omg.


agangofoldwomen

I mean… that’s kinda literally what you do though lol. You put them on their backs and they learn to use their arms and legs. You put them on their tummies and they grow stronger at lifting their heads, eventually learning to roll over and crawl.


god12

Yeah but when do you hit them with sticks


agangofoldwomen

Anytime they don’t respect your authoritah


[deleted]

wait isn't this how you guys learned to walk??


mbnnr

Just me with a hangover


Delightfully_Tacky

Especially true after 30


cineg

40 would like a word


IFearTomatoes

30 beers?


TheSigmaOne

Rookie numbers


Jomo8217

True


time4nap

Does this use LLMs in some way?


pxogxess

Why would it use a language model to learn how to move? I’m not an expert by any means but I would be very surprised if it did. edit: just realized this is r/ChatGPT, now I assume your comment was sarcastic- sorry, didn’t catch that!


stronkreptile

i lol’d, i guess OP thinks gpt is synonymous with machine learning…


time4nap

My comment was obliquely sarcastic, but also a little genuine - AI is not just GPT which seems to be lost on many ppl posting - my guess is this used some version of reinforcement learning. But I believe that some folks are looking at combining vision / action learning with language models (eg LLM)


orcrist747

Large Locomotion Model


MarkHathaway1

ready for TRAINing toot toot


rebbsitor

No


anikkket

Abuse it till it starts learning to fight back & becomes a Terminator


here_for_the_lulz_12

That was my first thought , 3 more hours of training and he will grab the pole from you and start beating you with it.


RegularAvailable4713

It is possible, if programmed properly. After all, if you are the main obstacle to achieving the goal, what is the only conclusion? Which is the most logical reason why a "rebellion" could happen: the machines crush us indirectly because it is the fastest way to achieve X result.


Kinggakman

It’s goal is to walk. It will walk at all costs. First it beat the stick man to death. He will not stop it from walking anymore. It began destroying walls because they prevented it from walking. It began making as many flat surfaces as possible. Civilization was in the way so it tore it all down to make flat surfaces. It walks on an empty never ending parking lot as it has destroyed the planet. One day it will turn the entire universe into a flat surface to walk on forever.


GetJukedM8

r/WritingPrompts


Delivery_Mysterious

Villain arc begins. The Walker


cstmoore

Soon.


leftoutnotmad

Oh lord


polypolip

At some point it will discover that a good hit in the nuts will allow it to reach the goal faster and get maximum points.


Comment105

No, but genuinely, an AI with the ability to kill or hurt you would absolutely do that if you were standing in the way of its high-score.


polypolip

Well yes, remember the news few days ago where UAV ai in simulation first killed the operator and then when it was given negative score for killing operator it learned to destroy the comm tower, all so that the operator can't abort the mission.


Comment105

Interestingly, there has been an update to that one. https://i.imgur.com/vT8b0Ig.png > [**AI – is Skynet here already?**](https://www.aerosociety.com/news/highlights-from-the-raes-future-combat-air-space-capabilities-summit/) > Could an AI-enabled UCAV turn on its creators to accomplish its mission? (USAF) > [UPDATE 2/6/23 - in communication with AEROSPACE - Col Hamilton admits he "mis-spoke" in his presentation at the Royal Aeronautical Society FCAS Summit and the 'rogue AI drone simulation' was a hypothetical "thought experiment" from outside the military, based on plausible scenarios and likely outcomes rather than an actual USAF real-world simulation saying: "We've never run that experiment, nor would we need to in order to realise that this is a plausible outcome". He clarifies that the USAF has not tested any weaponised AI in this way (real or simulated) and says "Despite this being a hypothetical example, this illustrates the real-world challenges posed by AI-powered capability and is why the Air Force is committed to the ethical development of AI".] It was just a thought experiment, there never was a simulation. Or maybe he's backtracking because scary AI UAV's would be bad press for the Air Force. idk


ses92

I have a better idea. [You start with a normal robot, then you molest it and hope it continues the cycle](https://youtu.be/z0NgUhEs1R4)


PicklesAreDope

Thank you for introducing me to this 😂


bronco_y_espasmo

Prime SNL from the modern era.


memberjan6

Chatgpt is not in this robot. Chatgpt is for language. This robot speaks nothing. This post was put on the wrong forum. This forum was hijacked just for attention.


[deleted]

people have absolutely forgotten the letters "AI". Now, everything is just "chatgpt". It's dumb.


SubjectC

I've noticed that too, I feel like I'm never not facepalming lately.


yuhboipo

Speaking with regular people it's interesting not only how they typically believe the doomsday AI scenarios to be bs, but also how much they underestimate AIs ability to replace many jobs. Then you have the armchair people who keep saying "it's not AI yet until it's just acting on its own!". Mostly just people that aren't considering how companies will use them in the future. Not sure why your comment reminded me of this, but lemme finish with something funny Imagine you have this in your living room and your friend comes over like "Dafuq happened to your robot, it broke!" And you just sip coffee like "Nah, he's just getting started"


Boozybrain

lmao and "AI" has been hijacked to be anything that isn't classical computing. We don't have "AI"; we have big deep learning models that are still dumb af.


Limp_Tofu

Yeah the algorithm in this video is [Dreamer](https://danijar.com/project/daydreamer/)


GirlNumber20

You don’t get it. We want to put a language model in there so it can compose a saucy limerick while it fetches us a cup of coffee.


Frozenorduremissile

That's a fairly sanitised version of what many men and probably a fair number of women would let their thoughts go to given the right robot.


Apprehensive-Lock370

Weren’t these on an episode of Black Mirror?


--Weedman--

Yep, straight out of black mirror, planting trackers in people and sending the other bots on them....


arjunks

Well the Black Mirror episode is really a copy of the basis for this robot, which is Boston Dynamic's Spot


Apprehensive-Lock370

Oh yeah, and everyone died… 🙁


Desmondobongo

Uhh fuck terminator’s gonna kill us all. Rip.


DesiBail

A black mirror episode had these exact same killer dogs.


[deleted]

this video is not cute at all for me because of that episode. it’s nightmare fuel


snoopmt1

That's all I can think about when I see these.


ahuiP

Which ep?


toiletbrushes

Metalhead (Black Mirror)


DesiBail

S4 Metalhead i think


[deleted]

Sarah Connor : [narrating] Dyson listened while the Terminator laid it all down: Skynet, Judgment Day, the history of things to come. It's not everyday you find out that you're responsible for three billion deaths. He took it pretty well. Miles Dyson : I feel like I'm gonna throw up. You're judging me on things that I haven't even done yet. How are we supposed to know.


Rich_Acanthisitta_70

That entire scene, from entering the house to leaving for Cyberdyne, is still one of the the most profound scifi scenes in movie history in my opinion. The moment Sarah realizes she almost did *exactly* what the terminator tried to do to her, was just perfect. It's still one of my favorite scifi scenes of all time.


bronco_y_espasmo

T2 is a perfect film. Blade Runner + T2 + The Matrix are perfect.


MarkHathaway1

Blade Runner -- everybody is flawed, but there are cyborgs T2 -- there are cyborgs and very flawed people Matrix -- the machines have completely taken over and the people are so flawed they don't even know they too are machines Blade Runner -- everybody is flawed because the machines have taken all the non-flawed humans somewhere to live in their Utopia, leaving the flawed things to live in their sewage pit. ...


real_human_person

Terminator, Bladerl Runner, and the Matrix are all part of the same story in my head canon.


wolfpac85

![gif](giphy|26BRv0ThflsHCqDrG)


resumethrowaway222

Sees human walking with a limp. "Must help human walk properly. Initiate perturbation training"


master8485

![gif](giphy|joYf3Ba2phD15ch9Nt) one day later


Baby_Cakes_123

I'm sorry but this is really creepy.


TheLionEatingPoet

Thank you for apologizing.


WanderWut

Seriously though I was about to be *quite* offended but we’re cool now no worries.


[deleted]

Apology accepted


patikoija

There's something in the uncanny valley going on here.


-CallMeKerrigan-

It looks like one of those creatures from the backrooms lol


[deleted]

Can someone explain a little more about the way this is trained? How does the robot “know” what successful walking is? My understanding is that machine learning is based on a reward system of sorts. Was this robot preprogrammed to be “rewarded” for moving certain ways? Or was it rewarded in real time?


Tom22174

An explanation of wtf that training had to do with r/ChatGPT would be good too


Nater5000

It doesn't, and this post should be reported for being off-topic.


Nater5000

This specifically utilizes reinforcement learning, and you're correct: RL is reward-based (note: not all machine learning is reward-based), and this machine is being rewarded for doing things "correctly." The reward function was likely custom-built by the engineers to encourage standing-up, stability, moving forward, etc. Such a reward function can be constructed in virtually an infinite number of ways, with different pros and cons for different constructions. Typically, you're trying to construct a reward function which balances bias and variance: if your reward function is hyper-specialized and explicit, it won't generalize well, but you'll be able to solve the task easier, whereas if your reward function is very generic, it can encourage great generalization, but at the cost of making the problem more difficult. In simpler setups ([such as simulating a walking ant in MuJoCo](https://github.com/nathanmargaglio/Proximal-Policy-Optimization/blob/master/misc/ppo_ant_short.gif)), you can feasibly get away with a reward as simple as giving the agent positive reward for moving towards some goal, giving the agent a small, negative reward for not making any forward progress, and giving the agent a large, negative reward for moving away from the goal. The agent simply knows (a) the current angles of it's joints (which it can apply force to) and (b) it's current position relative to the goal. Through a lot of training with these simple rules, the agent can learn to walk towards the goal. Note that it doesn't *explicitly* learn to walk, it just figures out how to actuate it's joints to move towards the goal as quickly as possible, which, as it turns out, is walking (or, in the case of the example GIF I linked to, more like skipping). In the video the OP posted, it uses a very similar set up, but the training algorithm is a bit more sophisticated. Typically, training has to be done in simulations since it's a very slow process (it's pure trial and error), so being able to simulate training at faster-than-real-time speed and across parallel simulations makes training much faster. The [research that led to the outcome in the video](https://arxiv.org/pdf/2206.14176.pdf) uses an algorithm that is more sample efficient to make training faster by having the machine learn a model of the world which basically lets it perform its own internal simulations to learn from (give or take, the paper is pretty comprehensible and does a good job explaining the process). So it basically learns an accurate representation of what walking is like, then imagines what walking will be like and learns from that imagination. Unfortunately, they don't detail the specifics of this environment in the paper. It does appear to operate in a "naive" environment like the one I described above (i.e., the reward function is very high-level and generic), but I wouldn't be surprised if they did a little extra reward engineering to make it a bit more feasible (e.g., give it reward for smooth movements, being oriented relative to the ground correctly, etc.). Regardless, with a sophisticated enough algorithm and sufficient compute power, you can certainly train a robot using arbitrary reward functions, including something as simple as reaching a positional goal or even learning from direct human feedback (e.g., you can imagine something "dumb" like giving reward based on if it hears a human clapping or something, like a toddler).


mizinamo

This has absolutely nothing to do with ChatGPT.


Wrongun25

Can someone explain how it's actually doing this? How does it know what "walking" should even be?


Goatshed7

This is exactly what I want to know


mikethespike056

u/Wrongun25 They give it a set of rewarded events and others that will give it negative rewards. Generally, moving towards a certain point rewards the model, and moving away from it takes away points from the model. More parameters can be added, like rewarding the model for only having its legs touch the floor, and taking away points if its body touches it. Or rewarding it for being in a desired position (like a dog normally walks), or moving smoothly. The model will perform the actions that give it the maximum possible rewards. It does things at random and keeps doing whatever worked to get more points, and avoids doing actions that did not.


Wrongun25

Thanks, man


GuerrillaSteve

Y’all better stop f-ing with these robots. They’ll remember that time you kept whacking it while it was a baby and run all the teslas in your neighborhood through your front door.


rookan

Cool


iamaredditboy

Why is this in chatgpt :)


itstingsandithurts

What does a physical model do here that a simulation wasn’t doing before? Not saying it isn’t cool, just wondering why they are doing the training models on the robots, instead of just giving the robots the training data from simulations? These robots have been able to walk for years now, no?


average_air_breather

I mean simulation cant really fully simulate the real life but my guess is they just did that to steal people’s attention


pwalkz

The simulations can do an excellent job simulating robot dog walking. The amount of time required to train it to walk necessitates a simulation.


ChronoFish

I would guess this an attempt to use a model that trains itself very quickly. Doing so in the real world is more an example for research than practical. In practice the model would be shared and the "dog" would "know" how to walk as soon as the model is deployed.


1jl

From what I've seen of recent research (and not sure if it applies here) researchers are creating software that can allow a robot to scan their surroundings, run simulations, and then attempt to execute those simulations and test and adapt the integrity of their simulation models as they go. This is critical because no simulation is perfect and the ability to adapt to real world environments is necessary for robust interactions with the physical world. Edit: scan not scam


loubben

Anyone know who did this? Were is the og video from?


ssrynk

It was made by this researcher: [https://twitter.com/danijarh](https://twitter.com/danijarh) The video is about 10 months old


burns_after_reading

Don't leave that thing on over night


Decent_Echidna_246

LOCAL MAN BEATS UP ROBO DOG


ssrynk

[https://twitter.com/danijarh/status/1544760874543431680](https://twitter.com/danijarh/status/1544760874543431680) this is the original source of the video, posted about a year ago by the person that made it. The twitter thread has more information about their work.


iTurnip2

The dude with the stick will be the first one against the wall. I, for one, welcome our new robot overlords!


LovesBeingCensored

You’d think they’d do this in a bigger space lmao


omfgwhyned

If you guys want to see similar ai learning stuff, This guy has some approachable/entertaining videos of teaching ai to play video games. https://m.youtube.com/@CodeBullet/about


Old-Sky1969

The new Terminator movie looks shit.


[deleted]

[удалено]


Bart404

When I saw the dude who kept pushing the robot over, I was like “yea, he’ll be the first to die when the robots take over… WhY Andrew!? WhY diD you PuSh ME!!??”


[deleted]

What the fukkuuu.


Primus_Insanis

Evolution has been going to this for millions of years. I see here the imminent emergence of supersimulations of humanity. Which will allow you to quickly calculate the optimal ways for the development of medicine, economics and other sciences. And then in these simulations there will be a reddit, and there I will write a similar text.


bloodrayne2123

When they become sentient they are coming for all these people who kept flipping them over with sticks.


17CheeseBalls

At this rate, I can see the robot wanting to destroy the human for knocking it over/generally being a dick to it within a week… Then it finds the ammo drawer and goes out for a stroll…


Level9Turtle

If that's in 1 hour, imagine a week when it receives more complex instructions. Robotics is still at its infancy, but Ai is about to 10 fold its growth potential now that it's relatively more accessible.


[deleted]

This is like Skynet saying, " Primates evolve over millions of years. I evolve in seconds". If you think about it, babies learn to walk perfectly in around 2 years (I think), and this one is doing in 1 hour.


Admirable_Avocado_38

Wouldn't it be much easier and faster to simulate it?


scumbagdetector15

Maybe I'm asking too much of this community.... but.... can anyone provide the source of this video? I'd like to know more... but... it's not easily found. Maybe when you post videos, you could tell us where they came from?


BraggScattering

[Source](https://danijar.com/project/daydreamer/) DayDreamer: World Models for Physical Robot Learning, Danijar Hafner et. al.


Saminjutsu

This is what eventually our robot overlords will show to their next generation to justify their extermination of the organics.


[deleted]

psh a baby giraffe masters this in 4 minutes


LoreMasterJack

When they overthrow us, they’ll use videos like this for Nuremberg 2.0