102 points | by jasondavies19 hours ago
Love this a whole heck of a lot more than NeRF, or any other "lol lets just throw a huge network at it" approach.
Well yes, but that's what gaussian splatting also was. The question is: are their claims to be so much better than gsplat accurate?
https://news.ycombinator.com/item?id=43120582
Like photogrammetry. But, handles a much wider range of materials.
I'm familiar with the premise of NeRF "grab a bunch of relatively low resolution images by walking in a circle around a subject/moving through a space", and then rendering novel view points,
but on the landing page here the videos are very impressive (though the volumetric fog in the classical building is entertaining as a corner case!),
but I have no idea what the input is.
I assume if you work in this domain it's understood,
"oh these are all standard comparitive output, source from <thing>, which if you must know are a series of N still images taken... " or "...excerpted image from consumer camera video while moving through the space" and N is understood to be 1, or more likely, 10, or 100...
...but what I want to know is,
are these video- or still-image input;
and how much/many?
Pretty sure the input is the same as for NeRFS, GS and photogrammetry: as many high rez photos from as many angles as you have the patience to collect.
I think the example scenes are from a common collection of photos that are being widely used as a common reference point.