Inceptionism: Mind Blown By What Neural Nets Think They See

Dr. Robert Hecht-Nielsen, inventor of one of the first neurocomputers, defines a neural network as:

“…a computing system made up of a number of simple, highly interconnected processing elements, which process information by their dynamic state response to external inputs.”

These ‘processing elements’ are generally arranged in layers – where you have an input layer, an output layer and a bunch of layers in between. Google has been doing a lot of research with neural networks for image processing. They start with a network 10 to 30 layers thick. One at a time, millions of training images are fed into the network. After a little tweaking, the output layer spits out what they want – an identification of what’s in a picture.

The layers have a hierarchical structure. The input layer will recognize simple line segments. The next layer might recognize basic shapes. The one after that might recognize simple objects, such as a wheel. The final layer will recognize whole structures, like a car for instance. As you climb the hierarchy, you transition from fast changing low level patterns to slow changing high level patterns. If this sounds familiar, we’ve talked out about it before.

Now, none of this is new and exciting. We all know what neural networks are and do. What is going to blow your knightmind, however, is a simple question Google asked, and the resulting answer. To better understand the process, they wanted to know what was going on in the inner layers. They feed the network a picture of a truck, and out comes the word “truck”. But they didn’t know exactly how the network came to its conclusion. To answer this question, they showed the network an image, and then extracted what the network was seeing at different layers in the hierarchy. Sort of like putting a serial.print in your code to see what it’s doing.

They then took the results and had the network enhance what it thought it detected. Lower levels would enhance low level features, such as lines and basic shapes. The higher levels would enhance actual structures, such as faces and trees. ibisThis technique gives them the level of abstraction for different layers in the hierarchy and reveals its primitive understanding of the image. They call this process inceptionism.


Be sure to check out the gallery of images produced by the process. Some have called the images dream like, hallucinogenic and even disturbing. Does this process reveal the inner workings of our mind? After all, our brains are indeed neural networks. Has Google unlocked the mind’s creative process?  Or is this just a neat way to make computer generated abstract art.

So here comes the big question: Is it the computer chosing these end-product photos or a google engineer pawing through thousands (or orders of magnitude more) to find the ones we will all drool over?

72 thoughts on “Inceptionism: Mind Blown By What Neural Nets Think They See

    1. Cybernetic psychedelia. Those pics aren’t a million miles away from some of the visuals you get from LSD.

      Taking drugs is interesting too, it’s like studying stroke patients to find out how the missing part of their brain affects them. Except less drastic. You can catch the bits of your brains, or the process of a thought growing and blossoming. Sometimes, at least. It’s something you can become aware of, and then notice in everyday life.

    1. For what? For the pictures the guy chose to show? Not for the process of image recognition itself tho I’d suppose. And that’s the machine’s day job, so people aren’t interfering with that.

  1. The interesting parts are where they gave it an image of just noise and ran the network backwards, asking it what it would be seeing if they said it was an image of a banana, etc. Then the higher level layers produced those images out of seemingly nothing. That is where the training set becomes more visible in the results, but more interesting is the fact they were able to run a neural network backwards when it’s classically described as a one-way algorithm from input to output. I studied that stuff for a semester and have been reading more about it for years and I still can’t work out how they did that one part. Absolutely astounding!

    1. Definitely an ANN-barely-understander, myself, but I wonder if there’s some mixing of two separate concepts, here… both of which I (barely understandingly) consider astounding
      A) backwardness: Is it that they’re actually feeding data backwards through the previous layers, or that they’re feeding the data forwards through a branch (inception?) layer which, essentially, is a reverse of the first layer…? e.g. the number of inputs and outputs on the first layer correspond to the number of outputs and inputs, respectively, on the “branch-layer”, with what I assume would be identical (or maybe inverse?) weights…?
      B) in which case, feeding a “noise” image through a normal one-way system, WITH the “banana” training-input set… and “incepting” that normal one-way system’s output, might make sense in the “classical one-way” sense…
      OTOH, if such a “incepting”/branch-layer could be created as in (A), then it might make sense that those layers, tied-together, could essentially be the “reverse” of the normal-one-way system…
      This is purely-hypothesizing by someone who hasn’t even tried a one-neuron ANN, and has never gotten through an entire article/document on “how they work”/”how they’re implemented” despite trying so many times (maybe that’s how they’re supposed to be learned…? “This is an ANN” training-sets in the multitudes wherein no single “this is an ANN” example is understood…? “This is a banana” “This is a banana” … “This is noise a banana, now show me the banana”) ;)
      I can’t *quite* see how it’d be possible to create an inverse/branch/inception-*layer* if such thing actually exists (rather than reverse-feed-through as they seem to describe it)

  2. This is truly fascinating! Interestingly the net tries to apply faces to anything it can which is EXACTLY what the human mind does. Faces in the clouds, tress, things that go bump in the night. Also how all the lake picture has huts and people as if imagining the potential of the space.

    The Japanese towers on a snowy mountain and blue/green twisty pagodas are worthy of paintings btw.

    1. We have survival imperatives to quickly recognize danger, to seek other humans, and to find food and shelter. These are the things that our meat processors are trained to identify. These neural networks don’t factor their own survival into their training. However they were mostly trained on pictures of things that we tend to photograph or paint: landscapes, buildings, people, and animals. The researchers could have used geometric shapes or objects we find less interesting instead.

    2. Now the NN needs what real brains have, bullshit rejectors, most of the data we take in is discarded, these NNs need to do the same so they won’t keep thinking they see faces in the clouds. But I do agree with you, some of these images are pretty artistic, so when NNs start making art who gets the royalties?

  3. My favorite bit is where we put computers in charge of drones and give them guns and missiles.

    Man to Computer: Dude, you blew up a kid competing in a steeplechase!
    Computer: I was aiming at the tiny terrorist dog he was sitting on. The kid and horse were just collateral damage.

  4. Wow, I’ve seen this story 5 times now, and hack a day has been the only source to actually tell my how/why it’s an interesting thing. the other stories were just a quick write up about “wow look at the crazy shit computers made” hackaday was the first to tell me the images were a product of a neural network.

      1. yeah they post the source, but that doesn’t give them any points, I wouldn’t read the source article of the news post makes it sound completely boring and pointless.

        1. Haha, I actually do the opposite – I don’t read the article but go directly to the source. In many cases article’s “interpretation” is flat out wrong. If subject interests me I rather read the original info. So pretty much I use these websites as aggregators of links on a certain area of interest.

      1. It’s not a new thing. We usually credit people for the tips they’ve sent into the tip line.

        In this instance, Michele sent in the tip on June 23, 2015 at 6:51 am (pacific).Will actually claimed this topic in the writer’s backend a little before this, and even though this post was finished on June 23, 2015 at 8:20 am (pacific), he started writing it a bit before hand.

        In fact, yesterday, I got hit with the same complaint., even though the post was written before a tip was sent in.

        If this is a new thing and people want credit for tipping us off, even though the tip had no influence over the post, then that’s a conversation we need to have. The editors and writers don’t have the tools to organize this, so expect some tips to slip through the cracks.

  5. Does it answer the question: Why do we have to sleep? Why do we dream?
    It’s known that a good NN has to be initialized with a random values for good result, is it possible that parts of the brain have to be reseted (while sleeping) with random noise (creating dreams) to get effective again when we wake up?

    Are there some Brain Hackers around here?

      1. “Are there some Brain Hackers around here?”

        Not saying there is or isn’t, but really;
        a) Install a firewall already.
        b) Thats some disturbing stuff you got in there!

    1. No. Sleep is fundamental to allow our minds to do its garbage collecting. Basically, it is during a night of sleep that your brain will weaken the neural connections corresponding to information that you did not use recently. It will also make stronger the neural connections corresponding to information that you recently learned (or revisited). This is why that you tend to memorize things more easily when you sleep after it.
      Without sleeping, our minds wouldn’t be able to “forget”. All the neural pathways corresponding to different pieces of information would just end up being all garbled and you would have serious trouble to remember or learn things. Ever wondered why people go insane with weeks of sleep deprivation? Now you know why.

      1. As mentioned bellow, the 60% shriveling of brain cell, is like a necessary chemical washing/cleaning.( Somewhere I can see this mechanism also as a possibility for the reinforcing of the neuronal connections.)
        But such cleaned brain cells, would have “noise” values (due to the re-expansion of cells), which would triggers dreams in the wake up phase of sleep. My bad to ask about sleep, while I meant the reason of dreaming.

    2. Regarding why we dream… My theory is that when we go to sleep, we are shutting down our senses and using the full capacity of our brains to sort out short term and long term memories. We process EVERYTHING we get throughout the day. This is an enormous amount of data. The reason that we dream is that this data is being passed along parts of the brain we use to hear, to see, etc.. and our brain tries to interpret it as we pass into a higher consciousness cycle. This essentially becomes short term memory which is why it is usually difficult to remember a dream after a few minutes. Though sometimes things get shoved into long term memory and we sometimes get a dream confused with a real experience.

      Apparently Dejavu is a hiccup in the process of filing long term and short term memories. It also tends to happen most often when we are sleep deprived.

      I am a bit of a brain hacker, but haven’t seriously messed around with it in about 12 years.

      1. You’re right with the consolidation of memory, but it doesn’t make sense to reshuffle ( or reprocess) everything back through the sensory parts of the brain, and I never experienced anything like that in dreams.
        I don’t say we aren’t reprocessing all the data, but I doubt that dreams are crucial for this.

        As more we sleep (in one day) the more we dream which is a bit contradictory to the reprocessing idea, and dreaming is the last thing we do while we sleep.
        But if the deleting of our short memory is done with random noise (the init state of any NN), and when this noise get processed, it could explain why we get dreams (which can get amazingly crazy when I can get to control them).

        Brain hacker: I find it interesting to listen to white/pink noise and hearing voices and even music inside these noises or using binaural soundtracks to get in a “trance” like state or watch the darkness/optical noise (were I can see patterns in the neuronal tissues covering/in front of the light receptors inside the eyes)

        1. Exactly, I’m more inclined to believe that dreams are just a consequence of our mind having reduced input stimuli (eyes shut, no noise, low light, constant temperature …). Lower power on stimulus equates to our mind picking up more background noise… With such low SNR, the brain would just pick up noise which is turned into dreams, akin to what the Google guys attempted in one of their examples.
          Pretty cool stuff.

    3. It’s obvious we sleep so that our brains and bodies have downtime for things like long term memory reinforcements and muscle repair etc. Most stuff about us you can just ask “why would evolution make this happen” and get the answer.

      1. In a predatory world, sleep doesn’t make much evolutionary sense. Why would we close your eyes, lie down and don’t move for 1/3 of a day, with the danger to get eaten?
        Last studies do point out that the brain want’s to get rid of waste, every brain cell shrivels up to 60% while we are asleep. Kind of washing the braincells from the accumulated neurotransmitter fluids.
        But you’re right about then memory consolidation and cell repair.

    4. My understanding was always that dreams is sort of “lazy loading”. There is just too much info to process in real life. So brain buffers that data and then, during sleep, it starts feeding that data to the network. As associations form, they produce visions (dreams). And since it’s not a simple act of remembering, but rather re-arranging the network on multiple layers to include new patterns, those visions can be quite weird.
      So I’d like to see this project modified a little bit to visualize what the network goes through when it is learning. That would be the exact analogy of a dream. What we’re seeing now should probably be called “imagination”

        1. It doesn’t really. Lucid dreaming simply means that something went wrong and your brain decided to start this process while you were awake. It’s an exception which further supports the rule.
          “no relation to real life” – it’s your sane interpretation. On some level of the neural network a house you went by or someone’s eyes could be imprinted as a rectangle or triangle. Dreams always have connection to real life.. either recent experiences or older ones (because recent events may cause previously established associations to update).
          But, of course, full disclosure – all this are just speculations at this point.

  6. I am far from an expert on this subject, but doesn’t a neural network rely on feedback to evaluate success? If so, then all the network is really doing is trying to please us. The more we tell it we like a certain thing, the more it will produce that certain thing.

    Let’s say we create a fairly simple network that inputs small writings and we want to evaluate whether the network finds them funny or not. If the network reads a short story about a someone grieving over the loss of a loved one and then it classify that as funny, do WE determine that it was successful or not?

    It’s the classical limitation in modern efforts in AI, IMO. The intelligence is limited by the creator. It can’t become smarter than us. We could make it smarter than any one person by using feedback from many people. But how then does it ever become smarter than a human?

    Then again, our own consciousness is pretty much built the same way, though much more difficult to pin-point. What we enjoy or hate depends heavily on our experiences and social interactions. And what we like or hate changes with time based on those experiences and social interactions.

    1. In your example about a neural network (NN) that evaluates texts as being funny or not, the network would first need to be trained. This means that you would feed the NN with a different text, one a time and for each you would “tell” the NN if the text is funny or not. After a certain large number of training iterations, your NN would be able to evaluate if any text (any text is just a random variation of a text inside the training set) is funny or not.
      The are problems with NNs, like “overtraining”. For example, if you only use classic-literature texts to train your NN and then ask your NN to evaluate if a SMS text is funny or not, the NN would much probably fail at task.
      Another problem is overfitting. For example, if you train your NN using texts in English, French, Chinese and German, the NN would not be able to learn much from the training set because your texts are too different. Using the wording in this article, it would mean that the network wouldn’t be able to store much invariant data from your training set.

      Nobody is talking about consciousness here, that’s the holy grail of AI. Currently neural networks are just an easy way to do pattern matching in very very complex signals.

    2. “But how then does it ever become smarter than a human?”

      The positive feedback doesn’t have to be supplied by a human directly.
      You could put your neural network in a “virtual environment”, for example, with how well it does in that environment being the factor that strengthens and weekends connections. Humans might be supplying the initial conditions, but theres no reason to say that limits the AI to the same intelligence as humans as a cap.

      Intelligence in this case is just “fitness for the task” – and while that fitness can be directly determain by human-dictated truths (“this book is funny”), it can also come from anything humans can measure even if humans cant predict or control it themselves (say, the weather). In this way its determined, but not limited by the creator.

  7. The fact that they fed its own processed images back through itself shows that we have a lot of ANN research left to do. If we arrange an ANN as a torus-shaped ring of layers with spokes coming off to accept input or process output then we might have a lot of interesting feedback loops and a more realistic simulation of how gray matter really works.

  8. The retina has similar layers of processing, finding lines shapes etc.
    LSD creates feedback and cross coupling in the brain, synesthesia.
    This explains the visuals oft reported.

  9. Awesome article that Google wrote about this. Funny to see it here after reading it a week ago on google news. Never seen an article in the news before it was featured on HAD before. LOL @ people quarreling over who ‘tipped’ it to HaD first. Irony.

  10. “Has Google unlocked the mind’s creative process?”

    No, this is a common technique researchers use to study neural networks. It provides insight into how to optimise an ANN.

  11. It also has a Rorschach-test aspect when you look at them, perhaps they can run these things through human brains and average the reaction and see what comes out when it’s reversed like that through a ‘real’ neural network.

    Incidentally, based on this and similar social media discussions I now reached a point where I feel justified in estimating the amount of americans that use LSD to be about 20% minimum.. And yet it’s never discussed in anti-drugs discussions or used as a thing in cop shows.

  12. They should show these images and other half processed ones to people with brain damage, autism and try to understand what they see. I thing they are on to something here.

Leave a Reply

Please be kind and respectful to help make the comments section excellent. (Comment Policy)

This site uses Akismet to reduce spam. Learn how your comment data is processed.