• Researchers shrink camera to the size of

    From ScienceDaily@1:317/3 to All on Mon Nov 29 21:30:32 2021
    Researchers shrink camera to the size of a salt grain

    Date:
    November 29, 2021
    Source:
    Princeton University, Engineering School
    Summary:
    Researchers have developed an ultracompact camera the size of a
    coarse grain of salt. The new system can produce crisp, full-color
    images on par with a conventional compound camera lens 500,000
    times larger in volume.



    FULL STORY ========================================================================== Micro-sized cameras have great potential to spot problems in the human
    body and enable sensing for super-small robots, but past approaches
    captured fuzzy, distorted images with limited fields of view.


    ==========================================================================
    Now, researchers at Princeton University and the University of Washington
    have overcome these obstacles with an ultracompact camera the size of
    a coarse grain of salt. The new system can produce crisp, full-color
    images on par with a conventional compound camera lens 500,000 times
    larger in volume, the researchers reported in a paper published Nov. 29 inNature Communications.

    Enabled by a joint design of the camera's hardware and computational processing, the system could enable minimally invasive endoscopy with
    medical robots to diagnose and treat diseases, and improve imaging for
    other robots with size and weight constraints. Arrays of thousands of
    such cameras could be used for full-scene sensing, turning surfaces
    into cameras.

    While a traditional camera uses a series of curved glass or plastic
    lenses to bend light rays into focus, the new optical system relies on
    a technology called a metasurface, which can be produced much like a
    computer chip. Just half a millimeter wide, the metasurface is studded
    with 1.6 million cylindrical posts, each roughly the size of the human immunodeficiency virus (HIV).

    Each post has a unique geometry, and functions like an optical
    antenna. Varying the design of each post is necessary to correctly shape
    the entire optical wavefront. With the help of machine learning-based algorithms, the posts' interactions with light combine to produce
    the highest-quality images and widest field of view for a full-color metasurface camera developed to date.

    A key innovation in the camera's creation was the integrated design of
    the optical surface and the signal processing algorithms that produce
    the image.

    This boosted the camera's performance in natural light conditions, in
    contrast to previous metasurface cameras that required the pure laser
    light of a laboratory or other ideal conditions to produce high-quality
    images, said Felix Heide, the study's senior author and an assistant
    professor of computer science at Princeton.



    ==========================================================================
    The researchers compared images produced with their system to the
    results of previous metasurface cameras, as well as images captured
    by a conventional compound optic that uses a series of six refractive
    lenses. Aside from a bit of blurring at the edges of the frame, the
    nano-sized camera's images were comparable to those of the traditional
    lens setup, which is more than 500,000 times larger in volume.

    Other ultracompact metasurface lenses have suffered from major image distortions, small fields of view, and limited ability to capture the
    full spectrum of visible light -- referred to as RGB imaging because it combines red, green and blue to produce different hues.

    "It's been a challenge to design and configure these little
    microstructures to do what you want," said Ethan Tseng, a computer science Ph.D. student at Princeton who co-led the study. "For this specific task
    of capturing large field of view RGB images, it's challenging because
    there are millions of these little microstructures, and it's not clear
    how to design them in an optimal way." Co-lead author Shane Colburn
    tackled this challenge by creating a computational simulator to automate testing of different nano-antenna configurations. Because of the number
    of antennas and the complexity of their interactions with light, this
    type of simulation can use "massive amounts of memory and time," said
    Colburn. He developed a model to efficiently approximate the metasurfaces' image production capabilities with sufficient accuracy.

    Colburn, who conducted the work as a Ph.D. student at the University of Washington Department of Electrical & Computer Engineering (UW ECE), where
    he is now an affiliate assistant professor. He also directs system design
    at Tunoptix, a Seattle-based company that is commercializing metasurface imaging technologies. Tunoptix was cofounded by Colburn's graduate adviser
    Arka Majumdar, an associate professor at the University of Washington
    in the ECE and physics departments and a coauthor of the study.



    ========================================================================== Coauthor James Whitehead, a Ph.D. student at UW ECE, fabricated the metasurfaces, which are based on silicon nitride, a glass-like material
    that is compatible with standard semiconductor manufacturing methods
    used for computer chips -- meaning that a given metasurface design could
    be easily mass-produced at lower cost than the lenses in conventional
    cameras.

    "Although the approach to optical design is not new, this is the first
    system that uses a surface optical technology in the front end and
    neural-based processing in the back," said Joseph Mait, a consultant
    at Mait-Optik and a former senior researcher and chief scientist at the
    U.S. Army Research Laboratory.

    "The significance of the published work is completing the Herculean task
    to jointly design the size, shape and location of the metasurface's
    million features and the parameters of the post-detection processing
    to achieve the desired imaging performance," added Mait, who was not
    involved in the study.

    Heide and his colleagues are now working to add more computational
    abilities to the camera itself. Beyond optimizing image quality, they
    would like to add capabilities for object detection and other sensing modalities relevant for medicine and robotics.

    Heide also envisions using ultracompact imagers to create "surfaces
    as sensors." "We could turn individual surfaces into cameras that have ultra-high resolution, so you wouldn't need three cameras on the back
    of your phone anymore, but the whole back of your phone would become
    one giant camera. We can think of completely different ways to build
    devices in the future," he said.

    Besides Tseng, Colburn, Whitehead, Majumdar and Heide, the study's authors include Luocheng Huang, a Ph.D. student at the University of Washington;
    and Seung-Hwan Baek, a postdoctoral research associate at Princeton.

    The work was supported in part by the National Science Foundation,
    the U.S.

    Department of Defense, the UW Reality Lab, Facebook, Google, Futurewei Technologies, and Amazon.

    ========================================================================== Story Source: Materials provided by
    Princeton_University,_Engineering_School. Original written by Molly
    Sharlach. Note: Content may be edited for style and length.


    ========================================================================== Journal Reference:
    1. Ethan Tseng, Shane Colburn, James Whitehead, Luocheng Huang,
    Seung-Hwan
    Baek, Arka Majumdar, Felix Heide. Neural nano-optics for
    high-quality thin lens imaging. Nature Communications, 2021; 12
    (1) DOI: 10.1038/ s41467-021-26443-0 ==========================================================================

    Link to news story: https://www.sciencedaily.com/releases/2021/11/211129122716.htm

    --- up 2 weeks, 4 days, 2 hours, 54 minutes
    * Origin: -=> Castle Rock BBS <=- Now Husky HPT Powered! (1:317/3)