Home Community Creating Multi-View Optical Illusions with Machine Learning: Exploring Zero-Shot Methods for Dynamic Image Transformation

Creating Multi-View Optical Illusions with Machine Learning: Exploring Zero-Shot Methods for Dynamic Image Transformation

0
Creating Multi-View Optical Illusions with Machine Learning: Exploring Zero-Shot Methods for Dynamic Image Transformation

Anagrams are images that change their appearance once you take a look at them from different angles or flip them around.  Creating such illusions normally involves understanding after which tricking our visual perception. Nonetheless, a brand new approach has emerged, offering a straightforward and effective strategy to generate these charming multi-view optical illusions.

Many approaches exist for creating optical illusions, but most depend on specific assumptions about how humans perceive images. These assumptions often result in complex models which will only sometimes capture the essence of our visual experience. Researchers from the University of Michigan have proposed a brand new solution. As an alternative of constructing a model based on how humans see things, it uses a text-to-image diffusion model. This model doesn’t assume anything about human perception; it learns from data alone.

The strategy introduces a novel strategy to generate classic illusions, resembling images that transform when flipped or rotated. Moreover, it ventures right into a recent territory of illusions termed “visual anagrams,” where images change appearance once you rearrange their pixels. This encompasses flips, rotations, and more intricate permutations, like creating jigsaw puzzles with multiple solutions, referred to as “polymorphic jigsaws.” The strategy even extends to a few and 4 views, broadening the scope of those intriguing visual transformations.

The important thing to creating this method work is rigorously choosing views. The transformations applied to the photographs must preserve the statistical properties of the noise. It’s because the model is trained under the idea of random, independent, and identically distributed Gaussian noise. 

The strategy utilizes a diffusion model to denoise a picture from various views, creating multiple noise estimates. These estimates are then combined to form a single noise estimate, facilitating a step within the reverse diffusion process. The paper presents empirical evidence supporting the effectiveness of those views, showcasing each the standard and suppleness of the generated illusions.

In conclusion, this straightforward yet powerful method opens up recent possibilities for creating charming multi-view optical illusions. By sidestepping assumptions about human perception and leveraging the capabilities of diffusion models, it provides a fresh and accessible approach to the fascinating world of visual transformations. Whether flips, rotations, or polymorphic jigsaws, this method offers a flexible tool for crafting illusions that captivate and challenge our visual understanding.


Take a look at the Paper and Project. All credit for this research goes to the researchers of this project. Also, don’t forget to affix our 33k+ ML SubReddit, 41k+ Facebook Community, Discord Channel, and Email Newsletter, where we share the most recent AI research news, cool AI projects, and more.

When you like our work, you’ll love our newsletter..


Niharika

” data-medium-file=”https://www.marktechpost.com/wp-content/uploads/2023/01/1674480782181-Niharika-Singh-264×300.jpg” data-large-file=”https://www.marktechpost.com/wp-content/uploads/2023/01/1674480782181-Niharika-Singh-902×1024.jpg”>

Niharika is a Technical consulting intern at Marktechpost. She is a 3rd 12 months undergraduate, currently pursuing her B.Tech from Indian Institute of Technology(IIT), Kharagpur. She is a highly enthusiastic individual with a keen interest in Machine learning, Data science and AI and an avid reader of the most recent developments in these fields.


🐝 [Free Webinar] LLMs in Banking: Constructing Predictive Analytics for Loan Approvals (Dec 13 2023)

LEAVE A REPLY

Please enter your comment!
Please enter your name here