crdrost 5 years ago

A very similar effect is harnessed in current machine learning models and is called "dropout". The basic idea in context is:

- you take your reference data and reserve a random chunk of it to figure out how good you're doing, this data is only used to grade performance -- the rest is for training.

- you notice that as you train on the training data, you start improving your performance on the evaluation data, but only up to a point: then as you train more, the neural net starts getting worse on the evaluation data.

- this problem is called "overfitting", your neural net is now trying so hard to get the "details" right that it is losing accuracy on the "big picture".

- so we create a circumstance where the neural net cannot properly do overfitting because those "details" cannot really be resolved, by just dropping nodes at random out of our neural network as we're training it. It will have to pick up a level of redundancy in the nodes and the associated noise of losing nodes should stop the later stages from really having a "focused view" of the training data, such that it can overfit that data.

If you're interested see e.g. Adrian's Medium post about it, here: https://medium.com/@bingobee01/a-review-of-dropout-as-applie...

  • cr4zy 5 years ago

    Dropout is traditionally only applied at training time: c.f. "At test time the weights are multiplied by their probability of their associated units’ dropout." from the medium post you linked. I _suspect_ that leaving it on during test has been accidentally tried a bunch of times though :) - so probably is not effective, although it could be interesting to experiment more intentionally with test-time noise. Our brains do not separate so cleanly between train and test, and so it seems more natural for the noise to be always on (if that's even the case as it seems hard to know what's noise when you haven't completely decoded the brain's activity). Decoding the brain seems to be super promising though and I hope we get good enough so as to upload our minds someday. Here's an interesting related article I read recently towards that end: http://www.foldl.me/uploads/papers/ccn2018.pdf

  • zozbot123 5 years ago

    Dropout is computationally expensive though. It does address overfitting (by doing something like training an ensemble of weaker models, none of which will individually overfit, but all of which capture different "views" of the dataset), but it's not without its drawbacks, and people seem to be using it less these days.

    • sdenton4 5 years ago

      Qua? I was under the impression that it was just going into architecture with less remark - ubiquitous, rather than dying. It's a great step towards building sparse models, as well, which is important for client side deployments.

  • hyperpallium 5 years ago

    Overfitting occurs when details are modeled that aren't justified by the data. i.e. not statistically significant.

    Dropping nodes seems a roundabout way of addressing this, and you'd get a similar effect by training less.

cabaalis 5 years ago

> When a new image is sufficiently different from the set of training images, deep learning visual recognition stumbles, even if the difference comes down to a simple rotation or obstruction.

I'm about as far from an ai expert as you can get.

When I see and recognize a school bus, it seems that object remains a school bus to me until there is very significant evidence otherwise, whether it is ahead, beside, tipped over, or behind as referenced in the example.

It would seem ai on a single image is problematic, and needs classification over time to gain "confidence" instead of a single attribution.

Edit/additional thought: It also seems to me that I know and accept that it's a "bus" before I know it's a "school bus" while another person might immediately recognize a "school bus" and then think "that's a type of bus." How wonderful to think of how those arrangements of hierarchies leads to differing opinions and creative abilities in humans.

  • phkahler 5 years ago

    Humans are never really shown still images. We are trained on real-world "video input" as we move around or manipulate objects. There is a sense that an object in your hand is the same object even if we rotate it, so "sameness" from different perspectives is learned even without knowing what it IS. Different people have different levels of ability to imagine an object in a different orientation, and I suspect this is related to our ability to identify objects in other situations. Also, if you've never seen the underside of a schoolbus I don't know why you'd be able to identify one from a bottom-only view. Large wheeled vehicle? Yeah but you'd probably have to think about aspect ratio and position of the 4 wheels and such. I'm thinking a more conscious effort and thinking through might be needed to identify it correctly rather than relying on the magic of lower level visual system.

  • kurthr 5 years ago

    If you want human like recognition (eg rotation and orientation invarience) of objects it seems like you would want to at least train them on image sequence data with multiple views of the same object (like video).

    It's not like humans learn image matching based on a sequence of disjointed 2D images. We train on binocular moving images of changing distance and orientation.

    Maybe the training sets are not well chosen for this sort if issue. Certainly expanding the set with rotations translations and scalings isn't difficult, but different orientation views would require a bit more effort.

  • jerf 5 years ago

    "It would seem ai on a single image is problematic,"

    It means that whatever it is that these latest models are doing, it still isn't what we are doing as humans.

    What exactly that difference is... well, if you could confidently and even more importantly, correctly tell me, in a way so detailed and correct it was implementable, you'd be able to become very rich.

    • taneq 5 years ago

      Well for starters, humans work on continuous video streams rather than still images, so there's a ton more information there. Even when we're identifying a still image, we're looking at a video stream of an object showing a still image (which is why a photo can look "exactly like the real thing" but we're never in any doubt that it's a photo and not the real thing.)

    • bena 5 years ago

      Yeah, I've talked with some people who believe we're close to general artificial intelligence and that we're fairly confident that we know what intelligence means but I'm not so sure we understand it.

      When we finally understand how we think, then we'll be able to re-implement it in software. But I don't think we're anywhere near understanding how we think.

  • vokep 5 years ago

    note that a human doesn't recognize a bus after seeing one once. A human takes thousands, maybe millions, of examples of various things before it learns to look at a new thing and recognize it afterwards

    Actually, not sure if humans are even really good at recognizing a never-before-seen object...hmm

    • whatshisface 5 years ago

      >note that a human doesn't recognize a bus after seeing one once.

      I disagree. If I showed you a single picture of a distinctive aircraft, you could recognize it on a runway. Likewise, birdwatcher's books rarely have more than a few pictures of each bird (nowhere near thousands), and birdwatchers seem to be able to identify the birds they are shown there.

      • ksdale 5 years ago

        We’ve seen lots of aircraft and birds though, even if we haven’t spent much time actively thinking about them.

        Even by the time we’re young children, we’ve been exposed continuously to something like tens or hundreds of terabytes worth of visual and aural information that informs our ability to recognize things. I think it’s very rare that people see something that they have no framework for recognizing.

        I know personally that I could identify distinctive aircraft from a single viewing because I’ve paid attention to a lot of aircraft, but I struggle with bird identification because I haven’t ever spent much time looking at birds. Even given a picture of a bird, I’m not that confident because I don’t know what characteristics could be common to other similar birds and what are distinctive.

        Birdwatchers are able to easily identify birds based on a couple pictures because they have seen thousands and thousands of birds.

        This feels sort of related to the study that showed that chess grandmasters had much better than average memories for the positions of pieces, but if the positions were random and not from an actual game, their memories were no better than amateurs. We rely heavily on things we “know” even when that knowledge isn’t exactly conscious.

        • TeMPOraL 5 years ago

          The aircraft/bird example is good. I too, can recognize aircraft easily, because I've seen a lot of them (both real and in photos/videos/drawings/3d models), and the distinctions mattered to me. Show me a bird photo, and a while later, give me a book containing this very photo + 50 others, and I most likely won't be able to find the one I've seen. Definitely not by any clues on the bird itself. It seems to me that you need some commonality with a whole category of objects before you even start paying attention to details of individual objects.

rdlecler1 5 years ago

I once built an evolutionary algorithm to evolve the neural network of a virtual robot. When I saved to disk and tried to rerun it produced a different result. Turns out the performance of the network was sensitive to floating point precision errors. I added some slight noise to the input values which made it much more robust.

perseusprime11 5 years ago

Brain is lot more complex. Trying to equate Brain to machine learning or neural network is not correct. The key ingredient of the brain is the reptile brain which is based on fight or flight and highly optimized for survival. Everything uses this part. In order to build similar, we have to build a neural network based on this survival instinct.

ggm 5 years ago

The most important sentence is the last sentence:

There’s still a considerable gap between real intelligence and so-called artificial intelligence

My Tl;Dr on the article is that it's a lot of maybe, we don't know, we don't know why, we are not sure, maybe.

But I certainly agree with the final sentence.