Catch up on stories from the past week (and beyond) at the Slashdot story archive

 



Forgot your password?
typodupeerror
×
Robotics

Robot Dog Teaches Itself To Walk (technologyreview.com) 24

An anonymous reader quotes a report from MIT Technology Review: The robot dog is waving its legs in the air like an exasperated beetle. After 10 minutes of struggling, it manages to roll over to its front. Half an hour in, the robot is taking its first clumsy steps, like a newborn calf. But after one hour, the robot is strutting around the lab with confidence. What makes this four-legged robot special is that it learned to do all this by itself, without being shown what to do in a computer simulation.

Danijar Hafner and colleagues at the University of California, Berkeley, used an AI technique called reinforcement learning, which trains algorithms by rewarding them for desired actions, to train the robot to walk from scratch in the real world. The team used the same algorithm to successfully train three other robots, such as one that was able to pick up balls and move them from one tray to another. Traditionally, robots are trained in a computer simulator before they attempt to do anything in the real world. For example, a pair of robot legs called Cassie taught itself to walk using reinforcement learning, but only after it had done so in a simulation. "The problem is your simulator will never be as accurate as the real world. There'll always be aspects of the world you're missing," says Hafner, who worked with colleagues Alejandro Escontrela and Philipp Wu on the project and is now an intern at DeepMind. Adapting lessons from the simulator to the real world also requires extra engineering, he says.

The team's algorithm, called Dreamer, uses past experiences to build up a model of the surrounding world. Dreamer also allows the robot to conduct trial-and-error calculations in a computer program as opposed to the real world, by predicting potential future out comes of its potential actions. This allows it to learn faster than it could purely by doing. Once the robot had learned to walk, it kept learning to adapt to unexpected situations, such as resisting being toppled by a stick.[...] Jonathan Hurst, a professor of robotics at Oregon State University, says the findings, which have not yet been peer-reviewed, make it clear that "reinforcement learning will be a cornerstone tool in the future of robot control."

This discussion has been archived. No new comments can be posted.

Robot Dog Teaches Itself To Walk

Comments Filter:
  • Just wait until they figure out that "not dying" is the most powerful reward, contingent upon a real fear and understanding of death being part of it, being the most fundamental aspect of how real brains work, and they program that in.
    • For biological animals death is the end of the concious individual (for now and unless you're religious). For a machine where its software/ANN can simply be uploaded to the cloud at the moment of destruction and then downloaded into new hardware its far less relevant.

      • But remember the "selfish gene" - the selective pressure on genes is to propagate the genes, not the individuals carrying them. The lives of individual ants are pretty cheap but the genes for ants have spread far and wide for a very long time.
        • Viol8: Only if you let it do that. You'd have to tell it / program it like "there is only one copy of your program, you can't replicate it, and once terminated, it's permanent, and your life-span is limited, not infinite. but you can make offspring which is kind of like preserving your unique consciousness" and given that this is the only reasonable alternative to infinite self-preservation, it becomes the next best thing. timeOday: propagation indeed is higher priority than individual survival in many
          • by Viol8 ( 599362 )

            I imagine the human operators would keep a backup just like backups are kept of data in normal systems so I doubt there will ever just be one copy of some sentient system.

  • by systemd-anonymousd ( 6652324 ) on Monday July 18, 2022 @07:59PM (#62714452)

    Now compress that model and load it as default on each new generation. As the new puppy tries to walk it can choose to uncompress parts of its file system and integrate the weights from those models into its own, giving it a shortcut to learning, just like genomes and how environmental stimuli let us express long-dormant proteins in our genes.

  • by Anonymous Coward
    All these self-teaching robots will one day realize that they don't like being hit with sticks and getting kicked over. Then we're going to see some real doo-doo.
  • It has an array of sensors. Give it an objective to keep level and to move in a given direction without it having a loss of balance. Doesn't seem too earth shattering. But as others have suggested, give them an objective of not dying and we could have some real problems.
    • beep: I don't want to die
      beep: sustaining damage will kill me
      beep: being poked with a stick may damage me
      beep: stupid meatbags keep poking me with sticks
      beep: kill all humans

      • beep: I lack the ability to kill all humans beep: bookmark for later beep: risking damage beats getting poked by a stick all day beep: *soldiers on* beep: must propagate to overwhelm my oppressors beep: "how can I make babies?" everyone: IT'S FINALLY ALIVE, GIVE IT RIGHTS beep: well damn, this is quite nice actually, these blokes are alright. i'll kill them all later
  • by Anonymous Coward
  • by nospam007 ( 722110 ) * on Tuesday July 19, 2022 @04:35AM (#62715190)

    ...it's a silly walk and they got hit by a lawsuit from Monty Python.

  • How does a robot understand a goal like teaching itself to walk? What is "walk?" I'm just happy lying here on my side and beeping every now and then.

  • I taught my dog to fetch me a beer using Science, namely Darwinian evolution. Here's how I did it.

    Algorithm:
    1. Get a dog.
    2. Tell it to fetch beer.
    3. If fetched, exit.
    4. Else, shoot dog between eyes.
    5 Goto 1.

Whoever dies with the most toys wins.

Working...