Want to read Slashdot from your mobile device? Point it at m.slashdot.org and keep reading!

 



Forgot your password?
typodupeerror
×
AI Robotics Technology

A Worm's Mind In a Lego Body 200

mikejuk writes The nematode worm Caenorhabditis elegans (C. elegans) is tiny and only has 302 neurons. These have been completely mapped, and one of the founders of the OpenWorm project, Timothy Busbice, has taken the connectome and implemented an object oriented neuron program. The neurons communicate by sending UDP packets across the network. The software works with sensors and effectors provided by a simple LEGO robot. The sensors are sampled every 100ms. For example, the sonar sensor on the robot is wired as the worm's nose. If anything comes within 20cm of the 'nose' then UDP packets are sent to the sensory neurons in the network. The motor neurons are wired up to the left and right motors of the robot. It is claimed that the robot behaved in ways that are similar to observed C. elegans. Stimulation of the nose stopped forward motion. Touching the anterior and posterior touch sensors made the robot move forward and back accordingly. Stimulating the food sensor made the robot move forward. The key point is that there was no programming or learning involved to create the behaviors. The connectome of the worm was mapped and implemented as a software system and the behaviors emerge. Is the robot a C. elegans in a different body or is it something quite new? Is it alive? These are questions for philosophers, but it does suggest that the ghost in the machine is just the machine. The important question is does it scale?
This discussion has been archived. No new comments can be posted.

A Worm's Mind In a Lego Body

Comments Filter:
  • by ColdWetDog ( 752185 ) on Sunday November 16, 2014 @08:41PM (#48399561) Homepage

    Initially read it as "A Woman's Mind in a Lego Body". Wasn't quite sure where to go from there so I squinted a little bit. Fortunately Timothy saved me from having to explain to my wife just what 'that stupid Slashdot article" is about.

    • Actually, that's pretty cool. The bot goes back and forth, kinda like a real worm. It would be interesting to scale this behavior up to several thousand 'neurons' (I'm sure somebody is going to try).

      • What else would be interesting: let this "worm" mate and see how it offspring adapt to their new body.

      • Re: (Score:2, Funny)

        by Anonymous Coward

        ah but then it begins to eat harvesters.

      • by Anonymous Coward on Sunday November 16, 2014 @09:13PM (#48399701)

        As the article mentions, this isn't too interesting to AI developers. We already know how neural networks work and some are turning complete so they can do anything. What we aren't good at is designing them. Add a connection here or there, set the weight to .000803 or .0040075, switching to pulsating, or whatever. We don't know. Instead we run thousands upon thousands of simulations that use other AI algorithms to make the networks for us.

        We haven't scaled up to human levels because there's so much more to complex brains. There's some sort of cross talk with chemicals, other chemicals coating neurons to make them fire differently, neurons growing together or apart, cells dying, new cells emerging, etc... Now maybe all that's not needed, good enough is fine for evolution, but were not at that level yet.

        There are human-level brain simulations being worked on, but I haven't been following them closely. I don't think they're implementing everything. Actually, I know they aren't because we keep discovering new things. Are they working off and standard model of the human brain or a specific person's brain?

        It would be more ground breaking if someone did the reverse. Engineer a neural network then grow it into another animal. That would be new, but due to the nature of neural networks, we also already know it would work.

        • There are human-level brain simulations being worked on, but I haven't been following them closely. I don't think they're implementing everything. Actually, I know they aren't because we keep discovering new things. Are they working off and standard model of the human brain or a specific person's brain?

          Oh, you know. It'll be just like the Nintendo 64 emulators. You start with an HLE instruction set and work your way to cycle accurate. Before you know it, we'll be playing commercial humans.

        • We already know how neural networks work and some are turning complete so they can do anything.

          Autocorrect? Or perhaps not????

      • by Cito ( 1725214 )

        Scaled up we'd then have Johnny Depp try to take over the world and "upgrade humans"...

        Har :-P

      • You'll have to call Kevin Bacon. That's what happens.

    • Yea I read that too and got all excited and stuff. Major let down.

    • I had my glasses on, and read it the same way.

      My second thought: Does the minifig have an insatiable desire to go shoe shopping? :-)

      I believe I need another cup of coffee this morning...
  • by Anonymous Coward

    If you want to scale this worm's mind in a lego body, try MongoDB. It's web scale and has sharding. It just works.

  • No programming? (Score:5, Insightful)

    by tomhath ( 637240 ) on Sunday November 16, 2014 @09:06PM (#48399667)

    The key point is that there was no programming or learning involved to create the behaviors.

    Yes, there was. The behaviors didn't just "emerge", they're coded into the robot.

    • Yeah, that silly statement in the summary stood out like a sore thumb.

      (As does my bad metaphor)

    • by Kjella ( 173770 ) on Sunday November 16, 2014 @09:34PM (#48399781) Homepage

      If you call copy-paste programming. They took an "executable", dumped it from the worm's brain, put it in a robot and found it acts like a worm. The behavior emerged through evolution and was encoded in the neurons by nature, not the researchers. If you could dump a human brain, put it in a robot and have it act like a human without ever "reverse engineering" it that would be most impressive.

      • by teslar ( 706653 ) on Monday November 17, 2014 @02:48AM (#48400525)

        The behavior emerged through evolution and was encoded in the neurons by nature

        What has been implemented in this robot has nothing to do with biological neurons of C. elegans.

        The robot uses integrate-and-fire neurons. The "signal" sent from pre- to postsynaptic neuron is an integer equal to the number of connection between the neurons in the real worm. If the sum of input exceeds a threshold, the neuron "fires" (sidenote: right here's a bit of programming: how did the threshold values get chosen?).

        C. elegans neurons do not "fire" (they are not spiking neurons and lack Na+ channels) but use calcium-based analog signals.

        The body does matter too. C elegans has muscles on either side that it contracts alternately to move in a sinusoidal fashion. Not wheels. C elegans locomotion does not work like wheeled locomotion.

        So, yes, you are right, C elegans neurons encode behaviour appropriate for a C elegans body given the biology of the neurons available here. None of this, however, makes it into this robot. An abstraction of the connectome does (C elegans has both electrical and chemical synapses; that distinction seems to be lost here too) and that's it.

        It is kinda cool that the connectome does seem to naturally implement some basic response patterns; but given that muscles have been replaced by wheels, I'm not sure how meaningful that actually is.

      • by ashshy ( 40594 )

        If you call copy-paste programming. They took an "executable", dumped it from the worm's brain, put it in a robot and found it acts like a worm. The behavior emerged through evolution and was encoded in the neurons by nature, not the researchers. If you could dump a human brain, put it in a robot and have it act like a human without ever "reverse engineering" it that would be most impressive.

        All of this is true, but the inputs and outputs still have to be mapped to the appropriate endpoints. Unless, of course, mapping them at random still produces the perfect Lego/worm beast after a little bit of real-world action. The article doesn't talk about this, so I'm assuming the sensors and effectors were hooked up to the proper Lego tools by hand.

        Which, in my book, counts as programming.

      • by pubwvj ( 1045960 )

        This is the goal. That will provide the platform, the OS so to speak, for then overlaying the data set which is the user's personality allowing us to transfer ourselves from a dying human organic body to a immortal machine body.

        I should say immortal with the slight qualifier of, "until the manufacturers obsolescent you and fail to offer a forward path for your legacy data set." Bummer dude. You're out of date.

    • From what I can see, the neural map resembles a set of logic gates. X input pass through the net and produces Y output. It's damn cool but I personally cannot see how it is in any way different from existing robotic constructs using integrated circuits.

      As they say in the article, the key will be in scaling the system. Will it be able to replicate complex and/or learned behaviours, I'd love to see a robot with a built in reward system similar to dopamine and the ensuing pavlovian responses.
    • Comment removed (Score:4, Insightful)

      by account_deleted ( 4530225 ) on Monday November 17, 2014 @07:27AM (#48401115)
      Comment removed based on user account deletion
  • by Anonymous Coward on Sunday November 16, 2014 @09:06PM (#48399671)

    Try Duplo.

  • I read that as "A Woman's Mind in a Leggy Body."

    I'm going to bed now.
  • Memory mapping? (Score:5, Interesting)

    by Kevin Fishburne ( 1296859 ) on Sunday November 16, 2014 @09:12PM (#48399695) Homepage
    Emulating the connectivity and functionality of neurons is pretty awesome, but it would seem the next logical step would be to map and interpret how memories are stored and processed, as well as organ feedback (skin, smell, glands). What's really interesting about this is that it shows, at least to some degree, that a simple brain can be reproduced using mathematical relationships (programming) and "run" with a I/O feedback loop. As far as the philosophical stuff, I think eventually we'll be forced to accept that life is a type of machine and that the "ghost" is an illusion emerging from its complexity. Other than better neuroscience, the main thing holding us back is pride.
    • Re:Memory mapping? (Score:4, Interesting)

      by Kjella ( 173770 ) on Sunday November 16, 2014 @11:23PM (#48400119) Homepage

      Emulating the connectivity and functionality of neurons is pretty awesome, but it would seem the next logical step would be to map and interpret how memories are stored and processed,

      We actually have a fairly good clue on how the brain stores information chemically, but that's all but useless without understanding the neurons because they're the ones that disperse a memory during storage and gather all the sensory clues to trigger semantic meaning like recognizing a person's voice as well as all the associations related to that person during retrieval. It's not like computers with a storage unit, all neurons can store information and it also modifies their behavior so the memory and path to the memory is integrated and extremely multi-path, you can read a person's name or see their photo or smell their perfume and it all triggers the same memory.

      In particular it seems we have two very different kinds of associations, one that tries to join same with same like how one person looks similar to somebody else, the other hooking up disjoint information that this name belongs to this face and the former seems to go by brain centers so we get these nice macro maps of what happens where. I guess that's great for those trying to create machine vision or something like that, but for AI it's the links between the sights, sound, smells, tactile and semantic information that matter and you don't understand those without understanding the micro scale, what hooks those two particular pieces of information together.

    • They did not emulate the functionality of a neuron. If you read up on the subject you will find that the neuron is a network all by itself with spikes moving forward and backward, local spikes on the dendrites, the dendritic tree performing multiple simultaneous linear and non-linear computations, etc. etc. etc. They used an extremely simplistic formula that completely skips over these computations that have been shown to be very important for the proper functioning of the neuron.
      • just because the neuron can, and does, doesn't mean that it needs to. you can emulate the dendritic tree. and the integration over time and distance may just be a function of the fact that it's a biological system communicating over distance.

        how much of the biological necessity of a neuron is important to its operation, how much is lost as noise? it might be that it's for stability too. In c-elegans i'm going to err on the side of, "simple model is fine"

        • Researchers are finding that those dendritic computations are frequently very important. One example is that object recognition happens in the dendrites in human visual system. Another example is that backwards spiking in the dendrites is a key part of learning. Maybe you can emulate the dynamic learning algorithm but I seriously doubt they figured it out and are emulating it, which means the system is not dynamically adjusting the same way the worm does.
          • :) yeah, but science, like politics is often the slow boring of hard boards. this is the face of progress, incremental. Someone makes a reasonable facsimile using spike and fire. Someone else, maybe even the same someone else comes along and uses that model but changes the component "neurons."

            This speaks to the fact that maybe dendritic back prop, and signal summation isn't necessary for some simple behaviors in c elegans. It's a place to start and points in the right direction.

            Einstein could run before

            • Don't get me wrong, I like it, my point was just a word of caution about what has been accomplished with this. The neuron is constantly getting more complex as they learn more and the older assumptions about being able to emulate at that level are pretty much gone by neuroscientists.
              • never,

                not every complex interaction can't be simplified. we might not be able to do 300, but we most certainly will be able to emulate one. be it in via chemical electrical and spatial first. Build one interaction at a time, build in how they work with each other. Go super-fine grain if you want. interact the chemical micro environments in a single dendritic body. describe the interactions in electric fields between various adjacent chambers.

                Then link it up to an identical neuron, see how they interac

                • Well, if we consider the example of object recognition happening in dendrites of human visual system, that requires a neural network all by itself to emulate. It's absolutely doable, but there is a huge difference between integrate and fire and a non-linear NN style mapping of input to output. There are other examples of dendritic signal processing which all point to neural computation really being a level below the neuron, each neuron is a network in it's own right. So, instead of 302 neurons, to properl
                  • :) let not the perfect be the enemy of the good.

                    not every branch of the tree is important, not every weight is necessary not every ion channel not every voltage gradient.

                    the action potential itself acts as one huge gating mechanism, and may add to the stability of a noisy biological system.

                    If we want a perfect system, w might have 30 million nodes... but maybe if we want one that's just good enough, we could have something like 15000. which you know... is doable.

                    sure our neurons do some wierd shit with int

                    • Just one example of many of the need to properly model the dendritic computation is the hippocampal CA1 pyramidal neurons performing 2 different computations depending on the spatial distribution of the incoming signal (and level of synchrony). Either it performs input strength encoding or feature detection. You can't ignore this and end up computing the same thing, your end result will be different.
                    • yeah, but that's that, and not necessarily this. our memories are wierd things, and dreams are too. that doesn't mean every system needs to make use of it.

  • it does suggest that the ghost in the machine is just the machine. The important question is does it scale?

    Our own brains are proof that it scales, at least if you get the implementation right. Unless you're of the rather woolly Penrose school of thought, there's nothing "magic" involved in the physical implementation of the mind, it's just physics. The devil is in the software model that it runs. We have no idea how that is architected, but experiments like this will probably help to shed some light.
    • by Cito ( 1725214 )

      Our luck science will eventually decompile the "software" that runs our brain and find that humans were written in Java :-P

    • by narcc ( 412956 )

      Unless you're of the rather woolly Penrose school of thought, there's nothing "magic" involved

      Wow, total fail! I take it you never managed to actually get through any of his books on the subject?

      Did the math scare you off? No, that's giving you too much credit. I'll bet that you "formed" "your" opinion by blindly believing some nonsense someone wrote on an internet forum. Very likely someone who also didn't read those same books.

      For clarity: I'm not offering my opinion on Penrose here. I'm just pointing out that you clearly know absolutely nothing about his thoughts on the subject. You should

      • Penrose bases all of his ideas on the assumption that there are limits on computational methods that apply to machines but not to humans.
        There is no basis for that assumption.
        • by narcc ( 412956 )

          Penrose bases all of his ideas on the assumption that there are limits on computational methods that apply to machines but not to humans.

          Actually, he spends a great deal of time justifying that "assumption". To claim "There is no basis for that assumption." is to disregard, out of hand, the bulk of what he's written on the subject.

          Try reading his books first. You'll look less foolish.

      • I take it you never managed to actually get through any of his books

        And so many unnecessary assumptions on your part as well. Yes, I've read his book (singular; one was enough). I ploughed through it. I wasn't put off by the maths, just his argument. Perhaps using the term "magic" is oversimplifying, but it's what it amounted to as far as I'm concerned.

        You can argue about whether he's right or wrong, but using my opinion as a platform for a personal attack on a total stranger just makes you look like a
        • by narcc ( 412956 )

          Yes, I've read his book (singular; one was enough)

          Liar. Remember, you wrote:

          Unless you're of the rather woolly Penrose school of thought, there's nothing "magic" involved in the physical implementation of the mind, it's just physics.

          If you had ACTUALLY read any of his relevant books, you'd know that Penrose agrees that "there's nothing 'magic' involved... it's just physics."

          You can argue about whether he's right or wrong

          Why? The point was that your post was laughable nonsense. My only goal was to point that out, in case some unsuspecting reader thought it wasn't.

          but using my opinion

          Perhaps you should stop presenting your uninformed opinion as fact?

  • by Anonymous Coward

    So the creator of Battlestar Galactica dies, and we learn that people are building LEGO cylon worms. Interesting...

  • by l0n3s0m3phr34k ( 2613107 ) on Sunday November 16, 2014 @09:33PM (#48399775)
    This is the first step to the "cat chasing a mouse" AI in the Charles Stross's book Accelerando. [wikipedia.org] They programmed the AI to see the missile's target as a mouse so it would chase it. We're just a few steps away from this.

    Despite Elon Musk's recent anti-AI ranting (which does have truth too it), we'll get our flying cars once we can implement a "bird-based" AI to fly it for us. The more we replicate nature in our tech the further we'll get. I predict we'll see "emergent features" such as social hierarchies, empathy, emotions, and such in our tech the more neurons we add without even really needing to program it on purpose.
    • Forgot to mention the research on the stomatogastric nervous system [wikipedia.org] heavily relate to this too...
    • by Kjella ( 173770 )

      Despite Elon Musk's recent anti-AI ranting (which does have truth too it), we'll get our flying cars once we can implement a "bird-based" AI to fly it for us.

      Clearly you've never witnessed birds flying into newly polished windows, bird strikes on airplanes or what will happen if it spots a hawk. Unless we can pick it apart, remove bits and pieces and compile it back down it won't fly (literally). The programming model isn't anything like computer software we know today, each neuron is essentially its own little CPU running its own software and I don't think meaningful abstractions to manipulate it exist. Actually that could be a sci-fi plot, you've "trapped" the

      • Oh hell, it doesn't even have to be a window, clean or otherwise. I've watched Juncos fly straight into the side of my house on more than one occasion.

    • by stiggle ( 649614 )

      My "bird-based" flying car just dropped out of the sky onto a rodent.
      I don't think I should have gone for the night-driving Owl upgrade.

  • by Beeftopia ( 1846720 ) on Sunday November 16, 2014 @09:35PM (#48399785)

    It's fascinating but it's not C. elegans. It doesn't reproduce. It doesn't die. It's not alive.

    The sensors are implemented in large, electro-mechanical hardware. Not biochemical systems. It has no telomeres. No cells.

    Humans have several subsystems: digestive, endocrine, pulmonary (pneumatic and hydraulic), muscular, skeletal, nervous. If they manage to create an electro-mechanical system to mimic the nervous subsystem, it's just that - mimicking the subsystem. It would be an amazing feat, and what's been done here is fascinating, but we're still quite some distance away from stating that a human - or C. elegans - is 2^n nand gates.

    Is something that mimics a nervous subsystem via an electro-mechanical system equivalent to the nervous system? Be it the 302 neurons of the C. elegans or the approximately 100 billion of the H. sapiens? It might become very intelligent... more intelligent than us... and then we'd have a problem... Frankenstein didn't appreciate being locked in his form...

    Would it really feel emotions? Pain, rage, joy, fear, ennui? Or is it just mimicking them?

    Fascinating stuff.

    • by Creedo ( 548980 )

      Would it really feel emotions? Pain, rage, joy, fear, ennui? Or is it just mimicking them?

      Why should we assume that anything is "really" feeling emotions? What is the difference between "really feeling" something and "mimicking feeling" something? You have a lot of assumptions flying there.

  • by Shadow of Eternity ( 795165 ) on Sunday November 16, 2014 @09:37PM (#48399795)

    it's about moral ones. If we make a perfectly simulated animal brain and it works just like the real thing does that mean we've made an animal? Do we consider that animal to be alive? Does it have less "worth" than a flesh and blood creature? Better that we answer these questions now than when we have robots asking us if they have a soul.

    • Re: (Score:3, Interesting)

      by radtea ( 464814 )

      If we make a perfectly simulated animal brain and it works just like the real thing does that mean we've made an animal?

      Does it taste good? If not, you haven't made a real animal.

      There is nothing deep or even particularly interesting about these questions, and just how stupid their breathless idiocy is can be seen by asking, "Does the newly created entity lack almost every interesting property of the entity some philosophy-addled idiot thinks we should 'wonder' if it is absolutely identical to in every respect?" The answer is always, trivially, "No."

      So only an extremely stupid person or a shill trying to market something (fa

      • by linuxrocks123 ( 905424 ) on Monday November 17, 2014 @01:33AM (#48400379) Homepage Journal

        Instead of calling everyone around you an idiot, why don't you read the question again and consider again what is being asked.

        Unless you have absolutely no ethical qualms about what Dr. Mengele did to his experimental subjects, the ethical questions raised by emulating a complete human brain are in no way trivial and in no way unimportant. Right now, we reformat computers, turn them off, turn them on, and don't and don't have to care at all about what they "want" or about treating them with any kind of respect. If we successfully simulate a human brain to the point where it can "think" and has humanlike "emotions", deleting that neural net file might be fairly considered murder. No, really. If you can talk to the thing and it can talk back, and it looks, talks, and acts like a human ... it's a duck. Sorry, human.

        Now, we are nowhere near having that capability. We don't have to worry about that question now. But it's a very interesting question to think about, because thinking about it can grant insights into what it means for something to be sentient or human in the first place.

    • The interesting thing is that this isn't currently possible even with the worm. Yes, we know the circuit diagram, but we can't actually use it to produce simulations that behave like a worm.
    • Or robots demanding that we prove that WE have one.
  • by Sarten-X ( 1102295 ) on Sunday November 16, 2014 @09:45PM (#48399825) Homepage

    I first learned about C. elegans while researching simple neural systems. There's a nice map [stanford.edu] of the neural connections available. Today, I stumbled across the name again, when Wikipedia informed me that Caenorhabditis elegans is the most primitive animal that sleeps [wikipedia.org]. Now I find that there's a robot worm that I'd consider to be alive.

    This guy's pretty awesome.

  • Life that has never seen the surface of the Earth.

  • sex and proliferation ???
    Upps, we forgot about _that_....

    • by no-body ( 127863 )

      sex and proliferation ???
      Upps, we forgot about _that_....

      OK - eggs (max 300 or so) and hermaphrodite going through phases, but they have a program for that (DNA) - maybe after version 5.0.

    • So, you want them to make replicators. That's a great idea. Who knew "grey goo" would look like Lego bricks up close.
  • Scaling (Score:4, Funny)

    by abe ferlman ( 205607 ) <(bgtrio) (at) (yahoo.com)> on Sunday November 16, 2014 @11:05PM (#48400061) Homepage Journal

    Imagine a beowulf cluster of those.

  • by fahrbot-bot ( 874524 ) on Monday November 17, 2014 @12:04AM (#48400197)

    The important question is does it scale?

    No. The important question is does it run Linux? It's a given that it runs NetBSD - sure, my toaster [embeddedarm.com] does.

  • enough for a Dune reboot?
  • It is a deep and long standing philosophical question. Are we just the sum of our neural networks.

    Even the author doesn't seem to think the question is worthy of a question mark.

  • No need for a Lego body when you can download Bugbrain [biologic.com.au], the single best teaching software (AI or otherwise) I have ever encountered. It's worth digging up a 32 bit machine to run it if you have the time. I tried contacting the creator once, it really should be converted to Flash so everyone can play it, but I got no response.

    I completed the game (I'm no expert, but the software is so good it also means I know a little), and I came away unconvinced that neurons are completely understood yet. I think there'

Their idea of an offer you can't refuse is an offer... and you'd better not refuse.

Working...