Visit

By Melanie Lefkowitz

Things are different on the other side of the mirror.

Text is backward. Clocks run counterclockwise. Cars drive on the wrong side of the road. Right hands become left hands.

Intrigued by how reflection changes images in subtle and not-so-subtle ways, a team of Cornell researchers used artificial intelligence to investigate what sets originals apart from their reflections. Their algorithms learned to pick up on unexpected clues such as hair parts, gaze direction and, surprisingly, beards – findings with implications for training machine learning models and detecting faked images.

“The universe is not symmetrical. If you flip an image, there are differences,” said Noah Snavely, associate professor of computer science at Cornell Tech and senior author of the study, “Visual Chirality,” presented at the 2020 Conference on Computer Vision and Pattern Recognition, held virtually June 14-19. “I’m intrigued by the discoveries you can make with new ways of gleaning information.”

Zhiqui Lin ’20 is the paper’s first author; co-authors are Abe Davis, assistant professor of computer science, and Cornell Tech postdoctoral researcher Jin Sun.

Differentiating between original images and reflections is a surprisingly easy task for AI, Snavely said – a basic deep learning algorithm can quickly learn how to classify if an image has been flipped with 60% to 90% accuracy, depending on the kinds of images used to train the algorithm. Many of the clues it picks up on are difficult for humans to notice.

For this study, the team developed technology to create a heat map that indicates the parts of the image that are of interest to the algorithm, to gain insight into how it makes these decisions.

They discovered, not surprisingly, that the most commonly used clue was text, which looks different backward in every written language. To learn more, they removed images with text from their data set, and found that the next set of characteristics the model focused on included wrist watches, shirt collars (buttons tend to be on the left side), faces and phones – which most people tend to carry in their right hands – as well as other factors revealing right-handedness.

The researchers were intrigued by the algorithm’s tendency to focus on faces, which don’t seem obviously asymmetrical. “In some ways, it left more questions than answers,” Snavely said.

They then conducted another study focusing on faces and found that the heat map lit up on areas including hair part, eye gaze – most people, for reasons the researchers don’t know, gaze to the left in portrait photos – and beards.

Snavely said he and his team members have no idea what information the algorithm is finding in beards, but they hypothesized that the way people comb or shave their faces could reveal handedness.

“It’s a form of visual discovery,” Snavely said. “If you can run machine learning at scale on millions and millions of images, maybe you can start to discover new facts about the world.”

Each of these clues individually may be unreliable, but the algorithm can build greater confidence by combining multiple clues, the findings showed. The researchers also found that the algorithm uses low-level signals, stemming from the way cameras process images, to make its decisions.

Though more study is needed, the findings could impact the way machine learning models are trained. These models need vast numbers of images in order to learn how to classify and identify pictures, so computer scientists often use reflections of existing images to effectively double their datasets.

Examining how these reflected images differ from the originals could reveal information about possible biases in machine learning that might lead to inaccurate results, Snavely said.

“This leads to an open question for the computer vision community, which is, when is it OK to do this flipping to augment your dataset, and when is it not OK?” he said. “I’m hoping this will get people to think more about these questions and start to develop tools to understand how it’s biasing the algorithm.”

Understanding how reflection changes an image could also help use AI to identify images that have been faked or doctored – an issue of growing concern on the internet.

“This is perhaps a new tool or insight that can be used in the universe of image forensics, if you want to tell if something is real or not,” Snavely said.

The research was supported in part by philanthropists Eric Schmidt, former CEO of Google, and Wendy Schmidt.

This story originally appeared in the Cornell Chronicle.







As society works through the widespread impact of COVID-19, Runway Startup Postdocs at the Jacobs Technion-Cornell Institute have been working on the front lines of the pandemic. At a recent Cornell Tech Community Conversation organized for the Roosevelt Island community and moderated by Director of Runway and Spinouts Fernando Gómez-Baquero, these extraordinary entrepreneurs described their work — ranging from a community-focused mental health platform to the development of a COVID-19 immunity monitoring test.

Learn more about how four of these extraordinary Runway Startup Postdoc Fellows have pivoted to help conquer the COVID-19 pandemic.

Mental Wellbeing

According to Runway Fellow and AwareHealth founder Prathamesh Kulkarni, lawyers are 3.6 times more likely to suffer from depression than those among the average population. Since the outbreak of COVID-19, the team at AwareHealth has reached out to law professionals, conducting research on what kind of support they feel they need in the wake of social distancing and self-isolation — two factors that have led to the acceleration of mental health awareness in the field. The response from users was clear: lawyers want help when it comes to fighting stigma and reaching out.

AwareHealth is a fully anonymous mental wellbeing platform for those in the legal profession, where byte-sized exercises, support from peers, and community-matched expert coaching are used to provide interactive therapy tools and support to its users. During COVID-19, the focus has been on two key aspects of this: community and proactivity. Through the in-app support groups and chat features, AwareHealth provides a therapeutic sense of community to those utilizing the platform, and by finding people who are going through the same issues and talking to mental health professionals, users can find ways to be more aware and productive toward bettering their mental health.

Immunity Testing

Runway Fellow Rebecca Brachman quickly recognized that there have been three common unmet needs when it comes to current testing for COVID-19. First, there have not been enough tests to go around. Second, tests can give false positives for the virus, since viral RNA can remain residually for several weeks after the infection is cleared. Third, the presence of antibodies does not guarantee immunity, as immunity depends on the antibody’s type, target, and quantity and current tests are not specific enough to be certain.

Around mid-March, when news of COVID-19 began to spread, Brachman contacted fellow Runway Fellow Server Ertem to discuss ways they could help to solve some of these issues with testing. Ertem’s biotech startup, Katena, was originally focused primarily on oncology supplies. However, they eventually realized that one of the current cancer detection tools under development at Katena could be modified to be an inexpensive and accurate test for coronavirus immunity — one that could address the unmet needs of the current tests. By being easily printable, showing results in 15 minutes, and targeting specifically COVID-19 neutralizing antibodies, Katena’s solution can bring higher quality point-of-care tests to the public in larger quantities. This will not only help those who are sick but lead to an increase in data for more accurate contact tracing.

Workplace Operations

One common problem that business owners are currently facing is the uncertainty of when they will be able to fully reopen — and how to do so safely. Will floor plans need to change? Worker hours? It is hard to predict how things will unfold. With these uncertainties looming overhead, Runway Fellow Davide Schaumann’s company, Spacemate, is hoping to find the most effective solutions for each unique situation.

Spacemate uses AI-powered automation and human behavior analytics to optimize workplace design and operations strategies. In the time of COVID-19, Schaumann became more concrete in the company’s objectives — now intent on discovering how to reconfigure the workspace, protect the workforce, and revive the operations for its users in a health-conscious manner. The AI does this by running multiple “what if” simulations of daily worker movements and congestion points against a Social Distancing Index (SDI). These results can help business owners set occupancy limits, desk assignments, work hours, and more — and the AI can be used in a range of businesses, like offices, distribution centers, hospitals, universities, transportation hubs, and food markets.


Powerful new technologies emerge when human experts and Artificial Intelligence (AI) collaborate. Cornell Tech Associate Dean Serge Belongie is a pioneer in this approach to innovation. Belongie helped establish the tech campus in 2013 and has been a faculty member ever since. In April, he was appointed the inaugural Andrew H. and Ann R. Tisch Professor in the Department of Computer Science

Belongie’s research in computer vision, machine learning, and augmented reality is motivated by his desire to build technology with a human dimension that serves people in beneficial ways. “That was part of the vision when the campus was created. I’ve clicked with it and I feel very at home,” he said. 

When computer vision meets bird watching

Belongie completed his Ph.D. in Electrical Engineering & Computer Sciences at UC Berkeley in 2000. He spent twelve years as faculty at UC San Diego before embarking on a collaboration with the Cornell Lab of Ornithology. The partnership grew out of Belongie’s interest in fine-grained visual categorization, a technique that combines computing and expert human knowledge to identify objects within subcategories– such as bird species– rather than at generic levels.

“To go down that path, you need to have powerful AI and really smart human experts and so a bunch of different people pointed me toward birding as the place to find that,” said Belongie.

The project, called Merlin, was a collaboration between Cornell Lab of Ornithology, UC San Diego, Northeastern University, Caltech, and UC Berkeley.  The partners had already produced an app that classified birds using a system of field guide questions, such as breast color and bill shape. Belongie’s students embedded with the team to develop a classification method based on uploaded images. “We borrowed their infrastructure of asking those questions but we provided the image and the computer vision system tried to help make it faster,” he said.

Two years into Belongie’s work on Merlin, deep learning exploded onto the scene and revolutionized the identification process. “It turned out you didn’t need to label anything other than the whole image,” he said. “We just threw it all away. No more questions, no more field guide stuff. Just put in the image, pop out the answer.”

This allowed Belongie and his team to move their focus from recognition algorithms onto methods for capturing and sharing visual expertise. The outcome was Visipedia.

Visipedia uses machine learning to harness crowdsourced expertise for the classification of visual data, such as images of flora and fauna. Communities of experts– from keen hobbyists to academics– gather and annotate data while machine learning trains and evaluates the system.

Visipedia’s approach is to go beyond basics and focus on more complex levels of categorization that are of interest to specialist communities. When expert crowds are challenged they become mobilized to contribute. “They don’t want to waste their time on tons of American robins. Their attitude might be more like, ‘Save it. My time is valuable. When you have a really tricky case, bring it to my attention,'” said Belongie.  

The system has been popularly deployed in Merlin’s Bird Photo ID app and iNaturalist which allow users to identify 1000s of species of birds, insects, and animals using smartphones. 

Future vision and entrepreneurship 

Google supported Visipedia for six years and Belongie remains a member of the Visiting Faculty Researcher program at Google Research. Belongie and the team behind Visipedia are now looking at how they can support collaborations between expert communities and big tech companies, allowing the approach to be scaled up on a level that would not be possible within merely an academic lab.

On campus, Belongie is encouraging his students to look to Augmented Reality and Virtual Reality (AR/VR). For years, computer vision has been based on data sets that are captured with cell phones or pulled off the internet, but that is about to change, he said. “That’s how things have been done for a long time but when you’re talking about AR/VR, it’s always-on wearable cameras constantly moving around the world.”

To support future research into AR/VR, computer vision, and human-computer interaction, Belongie has been involved in the creation of a new cross-campus initiative called the Mixed Reality Collaboratory.

Cornell Tech is the ideal environment for this interdisciplinary approach to human-focused technology and entrepreneurship, he said. “At any given time, I’m advising lots of little start-ups that pop out of Cornell Tech, and the Google interaction has been a fantastic way to amplify what Visipedia was doing at a small scale.”


Alumni startup Otari, an interactive workout mat company, was founded in Startup Studio in spring 2019 by Chris Kruger and Skyler Erickson, Masters of Engineering in Computer Science ‘19. After winning one of four Cornell Tech Startup Awards, the team has spent the last year refining their smart fitness mat and preparing it for the market. 

Kruger and Skyler recently launched an Indiegogo campaign in order to ship their first run of mats, which was fully funded within 30 minutes. Otari intends to start shipping units to its backers in October.

Learn more about Otari and how it has evolved in its first year from CEO and Co-founder Chris Kruger.

What is Otari?

The Otari Studio is a recently-launched smart fitness mat that streams unlimited strength, cardio, and yoga classes with personalized real-time feedback and automatic rep counting. The built-in camera scans the user’s form on-device to provide AI-driven recommendations, pose modifications, and comprehensive progress analytics. Our class platform bridges the attention of a personal training session with the community of an in-person class. 

How has Otari evolved since you won a Startup Award last year?

Otari has grown so much since the Startup Awards! At the time we had a (mostly) functioning prototype that weighed almost 30 pounds because it was made up of whatever we could find at Home Depot and carry back on the Tram. Since then, we’ve had some professional product design and manufacturing support to get the Otari Studio to a more user-friendly (only 12 pounds now!) and scalable (established manufacturing partners) place. 

Our team has also grown since winning the Startup Award, with our first full-time hire at the beginning of the year and an all-star cast of 4 yoga and strength instructors. Most importantly, we’ve grown our community of early adopters who have been excited to share their stories and why the Otari Studio is the right product for them. 

How do you feel your Cornell Tech experience prepared you to build Otari?

The Studio experience at Cornell Tech was fundamental to getting Otari started. Our team quickly learned how to identify and focus on a meaningful problem, strengthened our technical skills to execute on our solution, and developed the startup acumen needed to begin pitching and raising money for our early-stage company. We also benefited immensely from resources like the MakerLAB in our earliest days and are sad to no longer have access to all those 3D printers! 

Startup Studio taught our team that value isn’t rooted in the idea or concept behind a startup, but rather the ability of the team behind that idea to execute and bring it to life. 

Can you share one or two learning experiences you’ve had in the first year of Otari?

Bringing on outside help, whether through independent contractors or entire agencies, should always be a closely managed relationship. At the end of the day, only you will always have your company’s best interests in mind. It’s easy to rely too much on external stakeholders that weren’t there struggling with you in the earliest days of your company, but no one knows or cares about what is best for your brand than the internal team who is living and breathing the brand. Finding the right balance between trusting external parties, while making sure the best care is being taken of your brand, is an ongoing pursuit that we are still learning to navigate.

Nurturing a community and moving fast are essential for an early-stage consumer startup. With everything that goes into building a startup, it can be easy to forget that every action you take should be heading in the direction of building products that improve people’s lives. Never lose sight of who matters most, listen to them, and work quickly to meet their wants and needs. 

How does Otari establish community and company culture this early on?

By reaching out to people directly! In the beginning, we were nervous to solicit feedback directly from our early audience as a brand that is still in its early phases. It is nerve-wracking to ask for feedback on something that you know is still incomplete. Once we overcame that concern, things really started to come together and we were surprised that 9 times out of 10, people are happy to share their opinion. The feedback we received was both helpful in shaping our growth, and deeply affirming in what we are doing. Individual reach-outs eventually grew into a private Otari Facebook group, which has become a valuable sounding board between us and our early adopters!

After winning the Startup Award we were so focused on developing the product that we quickly realized everything was being done in an ad hoc manner. Lacking any set rules or boundaries is definitely a really exciting part of startup life and to be expected early on, but unfortunately, that doesn’t really scale. To guide the madness, we’ve developed a set of internal Otari principles that guide our decisions: foster inclusivity, celebrate individuality, drive personal growth, and most importantly, have fun! We have been able to retain the freedom and excitement of early startup life, but having values to guide the process has helped product, business, and community decisions immensely. 

Any final thoughts?

Always be open to new perspectives, but also don’t miss your opportunity for the sake of collecting more data. Trust your gut — if you have a question ask it — but at a certain point, when you are continuously asking questions that you know the answer to, it’s time to get moving on that idea.


For people with limited or low vision, walks to the grocery store, post office or a friend’s house can be an extraordinary challenge. The vast amount of visual cues that saturate our physical environment and assist with orientation and wayfinding can be of little help to people who experience limited sight clarity, low peripheral vision, or challenges in perceiving depth.

Emerging technology in vision enhancement poses promising avenues for navigational assistance for people with low vision, but research and development in this area is lacking.

Shiri Azenkot intends to change this, drawing on her expertise in human computer interaction, accessibility, and computer vision to design a head-mounted display (HMD) system to help people with low vision get around outdoors independently and safely.

An assistant professor at the Jacobs Technion-Cornell Institute at Cornell Tech and in Information Science, Azenkot is the recent recipient of the National Science Foundation’s CAREER Award for her proposal, “Designing Head-Mounted Display Systems to Support People with Low Vision in Outdoor Navigation.” Azenkot’s CAREER Award brings Info Sci’s tally this year to two; colleague Malte Jung also received one for his research into human-robot collaboration and teamwork dynamics.

Head-mounted displays, or HMDs, are headsets worn over the eyes that enhance elements of the physical world. They offer promising advances in tech-assisted navigation for those with low vision, even as the technology itself is fairly nascent and limited in terms of weight, speed, and resolution. Thus far, though, HMD research has been limited. As Azenkot notes, people with low vision are rarely considered in computing research, a striking omission since the majority of people with visual impairments have some degree of usable vision. Secondly, HMD technology often enhances and augments a user’s entire field of view, which can prove disorienting and overwhelming for a user who’s attempting to focus on a single object in the environment, like a street sign, for instance.

Azenkot’s NSF-funded research will guide HMD technology into the area of outdoor navigation with the design of a platform with both visual and audio cues for people with low vision.

Azenkot’s NSF CAREER project builds off her previous research with CueSee, an augmented-reality, HMD system she designed to help low-vision users find specific products in a supermarket. When standing in front of a shelving unit brimming with various products, the user can tell CueSee what product to search for. Guided by computer vision algorithms, CueSee scans the shelves, locates the product and then enlarges the product image in the user’s field of view. Azenkot intends to further this work and redirect it toward navigational assistance, namely by enhancing cues to avoid obstacles, navigate elevation changes, read signs, and follow routing guidance.

This story originally appeared on the Cornell CIS website.