• Research reflects how AI sees through th

    From ScienceDaily@1337:3/111 to All on Thu Jul 2 21:35:30 2020
    Research reflects how AI sees through the looking glass

    Date:
    July 2, 2020
    Source:
    Cornell University
    Summary:
    Intrigued by how reflection changes images in subtle and
    not-so-subtle ways, a team of researchers used artificial
    intelligence to investigate what sets originals apart from their
    reflections. Their algorithms learned to pick up on unexpected
    clues such as hair parts, gaze direction and, surprisingly, beards
    - findings with implications for training machine learning models
    and detecting faked images.



    FULL STORY ========================================================================== Things are different on the other side of the mirror.


    ==========================================================================
    Text is backward. Clocks run counterclockwise. Cars drive on the wrong
    side of the road. Right hands become left hands.

    Intrigued by how reflection changes images in subtle and not-so-subtle
    ways, a team of Cornell University researchers used artificial
    intelligence to investigate what sets originals apart from their
    reflections. Their algorithms learned to pick up on unexpected clues
    such as hair parts, gaze direction and, surprisingly, beards -- findings
    with implications for training machine learning models and detecting
    faked images.

    "The universe is not symmetrical. If you flip an image, there are
    differences," said Noah Snavely, associate professor of computer science
    at Cornell Tech and senior author of the study, "Visual Chirality,"
    presented at the 2020 Conference on Computer Vision and Pattern
    Recognition, held virtually June 14- 19. "I'm intrigued by the discoveries
    you can make with new ways of gleaning information." Zhiqui Lin is the
    paper's first author; co-authors are Abe Davis, assistant professor of
    computer science, and Cornell Tech postdoctoral researcher Jin Sun.

    Differentiating between original images and reflections is a surprisingly
    easy task for AI, Snavely said -- a basic deep learning algorithm can
    quickly learn how to classify if an image has been flipped with 60%
    to 90% accuracy, depending on the kinds of images used to train the
    algorithm. Many of the clues it picks up on are difficult for humans
    to notice.



    ==========================================================================
    For this study, the team developed technology to create a heat map that indicates the parts of the image that are of interest to the algorithm,
    to gain insight into how it makes these decisions.

    They discovered, not surprisingly, that the most commonly used clue was
    text, which looks different backward in every written language. To learn
    more, they removed images with text from their data set, and found that
    the next set of characteristics the model focused on included wrist
    watches, shirt collars (buttons tend to be on the left side), faces
    and phones -- which most people tend to carry in their right hands --
    as well as other factors revealing right- handedness.

    The researchers were intrigued by the algorithm's tendency to focus on
    faces, which don't seem obviously asymmetrical. "In some ways, it left
    more questions than answers," Snavely said.

    They then conducted another study focusing on faces and found that the
    heat map lit up on areas including hair part, eye gaze -- most people,
    for reasons the researchers don't know, gaze to the left in portrait
    photos -- and beards.

    Snavely said he and his team members have no idea what information the algorithm is finding in beards, but they hypothesized that the way people
    comb or shave their faces could reveal handedness.



    ========================================================================== "It's a form of visual discovery," Snavely said. "If you can run machine learning at scale on millions and millions of images, maybe you can start
    to discover new facts about the world." Each of these clues individually
    may be unreliable, but the algorithm can build greater confidence by
    combining multiple clues, the findings showed. The researchers also
    found that the algorithm uses low-level signals, stemming from the way
    cameras process images, to make its decisions.

    Though more study is needed, the findings could impact the way machine
    learning models are trained. These models need vast numbers of images
    in order to learn how to classify and identify pictures, so computer
    scientists often use reflections of existing images to effectively double
    their datasets.

    Examining how these reflected images differ from the originals could
    reveal information about possible biases in machine learning that might
    lead to inaccurate results, Snavely said.

    "This leads to an open question for the computer vision community,
    which is, when is it OK to do this flipping to augment your dataset, and
    when is it not OK?" he said. "I'm hoping this will get people to think
    more about these questions and start to develop tools to understand how
    it's biasing the algorithm." Understanding how reflection changes an
    image could also help use AI to identify images that have been faked or doctored -- an issue of growing concern on the internet.

    "This is perhaps a new tool or insight that can be used in the universe
    of image forensics, if you want to tell if something is real or not,"
    Snavely said.

    The research was supported in part by philanthropists Eric Schmidt,
    former CEO of Google, and Wendy Schmidt.


    ========================================================================== Story Source: Materials provided by Cornell_University. Original written
    by Melanie Lefkowitz. Note: Content may be edited for style and length.


    ==========================================================================


    Link to news story: https://www.sciencedaily.com/releases/2020/07/200702152445.htm

    --- up 23 weeks, 2 days, 2 hours, 39 minutes
    * Origin: -=> Castle Rock BBS <=- Now Husky HPT Powered! (1337:3/111)