AI Researchers Present Neural Mixtures of Planar Experts (NeurMiPs): A Novel Planar-Based Scene Representation for Geometry and Appearance Modeling

This Article Is Based On The Research Paper  'NeurMiPs: Neural Mixture of Planar Experts for View Synthesis'. All Credit For This Research Goes To The Researchers Of This Paper 👏👏👏

Please Don't Forget To Join Our ML Subreddit

The technology is booming, so in the future there will be a scenario where people can explore the world by being in their bedroom itself. As people move forward, details appear, and as they move laterally, occluded regions reappear. Although this scenario seems tempting, its realization requires innovations in several areas. One such area is Novel View Synthesis (NVS), which is high quality, can work in real time and is also memory efficient. In the NVS system, the challenge is to render the scene from different viewpoints in a photorealistic way. In addition, the system must be light and fast to be used everywhere. Experts have proposed several methods by which the visual world can be reproduced to meet this challenge. One approach is to perform image-based rendering (IBR) by modeling the geometry and point clouds of the scene. This approach adapts the visual characteristics of other existing views and can render high quality images. However, this approach consumes a lot of memory and requires proxy geometry.

In contrast to this approach, neural radiation fields synthesize very realistic images and consume less memory. It can handle complex geometry and scene effects that are difficult for other traditional methods. However, it has a surface modeling challenge. If the surface modeling is not appropriate, the geometry of the scene cannot be accurately captured, resulting in the generation of artifacts.

This article aims to discover other expressive, efficient, compact and generalizable 3D scene representations. This work models the surface using plane geometry. It models real-world surfaces through piecewise local planar structures. Unlike multiplanar imaging, the approach allows each plane to have arbitrary direction, position, and size. This allows for fast rendering and the calculation in empty spaces is eliminated.

Free 2 Minute AI NewsletterJoin over 500,000 AI people


This work proposed a new neural representation called planar expert mixing. It also presents a design of a neural rendering method using NeurMiPs. In this work, the scene is initially represented as a mixture of local planar surfaces, an oriented 2D rectangle. A neural radiation field function is used for each plane to encode its appearance and transparency depending on the view. Additionally, the input images are used to learn geometry and luminance fields. A radius rectangle intersection is checked during render time. The coordinate of the point of intersection is used to evaluate color and transparency. Finally, the ray color is calculated by combining the colors of all intersecting subjects using alpha blending. The proposed 3D architecture and other neural rendering approaches are shown in Figure 1.


For each planar expert model, this work uses Multilayer Perceptron (MLP) with three fully connected hidden layers, ReLU activation for each hidden layer, and sigmoid activation for the final output. It uses a mixture of planar experts to adapt to the geometry of the surface. During the test, a pixel is rendered by shooting a ray from the eye, and the luminance is evaluated along the ray. An intersection is made between the given ray and a plane of infinite size to decide if a local plane intersects the ray. Only intersecting rectangles are reserved on which the color and transparency of the ray are evaluated. Alpha compositing is done to get the ultimate approximation of the ray color.

NeurMiPs require both the geometry of the plane and the radiation to be optimized for training. The geometric parameters of the plane are structurally projected from motion using the coarse 3D point cloud. Splinter and geometry are optimized together. A large capacity NeRF model is trained as a teacher model to distill knowledge. After overfitting to the teacher’s grating, the plan parameters are fixed and the students’ radiation field models are refined to improve rendering quality.


This work proposes to pre-render the alpha values ​​for each rectangular plane firing, the pre-rendered alpha values. Early ray termination is avoided to prevent later evaluation of the network. A custom CUDA kernel is designed to intersect ray plane, alpha compositing, and model inference.


The proposed approach is evaluated on two different datasets, Replica and Tanks & Temples. Moreover, the replica is a simulated dataset with various indoor scenarios. This search selects seven random scenes, and 50 training frames and 100 test frames are rendered for each scene. Each scene has high resolution geometry and photorealistic textures. Blender Proc is suitable as a physical renderer and the image resolution is 512*512 pixels. Tanks & Temples covers five high-resolution real-world scenarios taken from the surrounding 360oh see.


Work is quantitatively assessed using the Peak Signal-to-Noise Ratio (PSNR) approach, Perceptual Metrics (LPIPS), and Structural Similarity Index (SSIM). The algorithm is compared to the most promising techniques, neural radiation field (NeRF), MPI-based method (NeX) and real-time hybrid methods such as NVSF, KiloNeRF and PlenOctrees.


This article proposes NeurMiPs, a new 3D illustration for a new synthesis of views. Unlike neural surface rendering because it consumes less memory and time. Also, the approach is considerably more sample-efficient with better extrapolation than volume rendering. The proposed method effectively reflects the geometry of multi-plane images. Furthermore, this approach achieves superior performance compared to state-of-the-art techniques on a challenging new benchmark for view extrapolation.


James G. Williams