Best Projector For Projection Mapping

best projector for projection mappingLooking for the Best Projector For Projection Mapping? You’ve come to the right place

We were talking on the podcast about how people at Christmastime or holidays in general they take these Lights lasers and they project them on the side of their house and it just looks kind of cheesy I mean it looks very warped and and you know not like projection mapping should Right where the light actually conforms itself to the geometry of the object Your project exactly exactly.

Best Projector For Projection Mapping

So I was thinking would it be great if like somebody made a tool that allowed to anybody to? projection map onto their house or in in their rooms or an object or whatever you want and Somebody has a new company called light form. Just stay beautiful Consumers and prosumers so we went to visit their HQ and chat with our CEO about their .

technology and check out their light form projection mapping device Brett thank you so much for having us here. Yeah, of course. I’m super stoked to see light form So you guys make this device and working on it for a couple years. It is a projection mapping device. Yes, exactly So it’s a hardware and a software solution purpose-built for projected .

AR or projection mapping So what the light form is is it’s a high resolution camera. So it’s a 4k color camera There’s a computer inside of this and you take this and you actually mount it onto Almost any projector and so now it’s attached and what we’ve done is we’ve turned this projector into a 3d scanner So we can take this we can point it at .

any object in the real world We get a 3d scan of it and then we can apply magical effects or we can ply information on to the real world using projection I’ve seen projection mapping in a variety of places and some amusement parks like Disneyland It’s in retail locations even like eSports events, but not really in the consumer space This device looks .

really small and compact. How does this compare to something that like a disney would use for a theme park? Yep Yeah, so we actually got into projection about 10 years ago while at Disney Imagineering So I was adjusting measuring fill was touring with Skrillex doing big projected visuals So light form is all about taking those big kind .

theme-park level Magical experiences and then boiling down into a solution that is accessible for any designer to create project today R so we’ll talk about kind of what that scanning process is so light form is designed to Scan static scenes and then within a couple of minutes create super compelling Magical effects. We have a couple minutes right .

now. Can you walk me through? Yes Devlin’s processes Alright, so we’ll set this one to the side and go to our demo in unit over here So, let me just grab a stormtrooper helmet here Alright, and then we actually 3d printed a tested logo or on the form one printer over there so I’m just gonna set these objects down and Then what we’re gonna do is we’re. 

gonna scan it So right now the camera on the laptop you can actually see a live feed of what the camera is seeing and then Phil is gonna trigger a Scan, so we’ll see what that looks like. So projection mapping really it’s kind of backwards you map first and then you project on the computer’s understanding of the shape and sizes of the world and that’s actually why we call this projected AR and not projection mapping because We think the mapping out of projection .

mapping So it’s more about the experience in the content that you can create and less about that painful process of mapping pixels amount objects in a lot of AR devices we’ve seen mapping the world Understanding the shapes and size of the world’s a huge part of it. How does this do it? Because here I see it’s it’s one lens Yep, right, you’re not using like what a Kinect would have in terms of like IR sensor as blasters or stereo Even how does this understand .

the shape of that? Yep, so it is stereo and it’s stereo because You think of stereo is like left eye right eye right and that’s how we see in 3d But the left eye is the camera of the light form and the right eye is the projector So they’re actually talking to each other through light So those patterns that we saw were black and white patterns those were zeros and ones so it was actually sending binary code to each other through light and at the end of the day what we .

end up getting is A projector resolution scan so you can get a 1080p scan of whatever you point at so much higher than a real-time depth sensor and we can also use a projector with different lenses so you can actually zoom in and out of The lens on that projector you could use different lensing Projectors and we are capable of scanning any scene from 3 feet to infinity so you can scan a coffee cup Stormtrooper mask or a building across the street. And so your .

projector agnostic. Yep, and projector agnostic So any normal throw projector is what the LF one supports? productors have a vive wide range of throws like a Small pico projector and the angles you but you could mount this on a small thing habit behind your television and and it would still Work fine. Yep. Yeah, so it’s actually designed to be mounted with adhesive So you literally just peel off a sticker plop it on to any projector You can put on the top or .

the bottom plug in HDMI and the power and then you’re good to go You turn your projector into a 3d scanner the calibration though. What about the calibration between the scanner do that? all automatically using computer vision So we’re all PhDs in computer vision and design and spent two years building a system Read have to worry about any of that. We just do it for you. All right Well, the scan seems to be done those get much faster than a couple minutes .

What does the computer know about this world now? So if we on fill screen you can actually see that we have the scan so we have the color information And then we also have the depth behind that Okay, so we have a projector resolution depth and color scan and we’re going to use that to then Outline an object in the scene and apply a .

magical effect to it. You can was saying projector resolution So the higher the resolution of a projector the better fidelity your scan is correct up to 1080p So the hardware is designed to be cheap. And so we support up to 1080p so there’s no understanding of what the scene is that this is a helmet or it’s a head or you know? It’s a box. You’re manually masking But what tools does the system in able to help you with that masking? So we have a bunch of easy. 

things that you expect in like Adobe suite So we have a pen tool that we spend a bunch of time on for quickly outlining objects You just have to kind of trace an object. We also have quick select tool so quick select a magic wand like shop has But instead of just having color we use depth as well So it’s not just looking for outlines or dark shapes so you could select a white coffee cup on a white table Okay, select the coffee cup using depth in center. Okay, so you want to they’re selecting the outline here right picking out those those shapes You can tweak I assume you know and refine your mask and then with that mask Put a photo or a video on this so you can do images or videos so y.

ou can use your existing assets But what we’re really excited about is these projected AR effects So these are intelligent effects that are actually using the color and the 3d scan of the scene So this is actually wiping in depth through the object. We call this one digital fade So there’s a library of these effects and these effects are actually real time So they’re running on the device on the computer. That is the the light form lf1 not compiling Export in this .

video. Yeah. These are shaders, right? So there are shaders that are running on the device in real time. And what we can actually do is we can control them So you can go over here of that and actually control different parameters of the effects and Then Phil will show you Mapping we have this tested logo here and Phil show the pen tool mapping the tested logo and applying another effect Very so we’re publishing that right now So it’s actually saving the project .

file transmitting it to the light form and then playing it back into the real world So that’s the workflow right? You’re working. You’re creating your mass you are adding an effect Your library effects may be adjusting some variables and they are stacking the effects as possible And then once you get published that all sends to the light form, you don’t need to have a connected computer, correct? So the coolest thing is Phil could actually close his laptop and walk .

away and we just made permanently deployed magic. Mmm It’s all runs locally. So but what’s inside here essentially is like a small computer. Yeah HDMI output It’s a quad core a nine processor. It’s designed to be permantly installed. Mmm So, you know if you’re an AV installer, this is actually a fanless design and there’s on-board storage So it’s not like an SD card that’s gonna fail over time Yeah, the idea is that it’s computer vision hardware that makes it really .

easy to create compelling scenes in seconds But then you can permanently deploy this if you’re an art gallery You know in your house or at a retail store now because you’re running real-time effects on here and even combining that potentially with static images videos What’s the load of this machine? Can you that at some point do you lose .

framerate? Is there a maximum? Complexity you have for this. Yeah It’s very easy for an artist to throw things at a computer to make it run really slow Because it’s basically a mobile phone processor And so what we do is we intelligently try and kind of optimize the scene. So we’ll take if you create a bunch of really Complex effects like a particle system will actually render that to a video file and send the video file to the device Okay, so then you can .

have a mix of passive and interactive content running on the device, but you don’t have to worry about You know kind of optimizing your content we do that for you that this is super cool I know this is something you guys bought and you have run this demo before but the whole Scan to finish was just a couple minutes. And this looks really neat. Yeah, so we actually Have you brought in an object? Yes, so now we’re gonna map an object that you chose not us .

And I want to get my hands on that software. Yes, definitely. All right. Check it out. Sounds good So we brought something from our office, of course the zorb is the f1 from the Fifth Element nice static prop pretty complex and Phil here you guys have scanned this Yeah, we have so we have our projector scan as our background here so again, it’s it’s really easy to actually go in and use this as a reference as we’re tracing the outlines We can do a quick .

selection. Just kind of select all the gray areas and then convert that to vector in this case we’ve actually gone and Created a creative mask ourselves Took only a few minutes and what we can do Is if we want to actually interactively edit this mask I can stream the preview to the device So I have a crosshair now that’s out in the real world and if I wanted to just tweak some of that barrel right there I could go and Get that really precisely aligned and so everything. 

that’s bright here is Within the mask, correct? Yes. So we’re projecting white to represent the the mask of the area that you’re gonna be mapping Okay, and then what’s really handy? Is this interactive crosshair? So as I’m kind of editing these points in the real world, I can see exactly where they lie so we bring this point in a little bit reading that point out a little bit and then we have a Kind of sub pixel accuracy here so we can scroll way in and we can actually .

get down to you know our Bezier handles and this is when you really want to get it nice and precise but for most applications you can just do kind of a rough selection and that’s gonna be portent because the angle this projector is gonna be Different and where the shadow is kind of overlap. Yeah, you want to fine-tune that if I shift this though, you’ll need the rescan Yeah, correct so by vectorizing the mask outline It’s very easy for us to just select that whole .

shape and then do a transformation on it and move it On our roadmap. We have we call Auto fine-tune which will account for small shifts in projector movement or object movement So typically when you have an installation going for a while You know things heat up cool down buildings move a little bit right there and you can notice that You know five pixel offset so with computer vision We’ll actually be able to help solve effort and that’s for kind of lateral .

movements or just fixed planes scale changes If I do a big rotation or even a slight rotation, you’ll need to tweak that mask manually Correct. You will you will want to go in and there’s significant changes to the scene Tweak your Bezier handles a little bit. But what’s nice about our instant effects. Is that the content underlying you don’t have to recreate that You’re actually able to Just apply that and with the new scan and the new underlying projector pixels .

that content just kind of works right away exactly How you would seen it before correct? Yeah, so This one. We actually you know have a some nice depth data there. That’s cool so what I’ll do is I’ll bring up our effects pane and we’ll go and Use a depth trace effect. So this is an effect. That’s actually using the underlying depth data. I’ll go ahead and insert that and Then we can actually publish that So you do see a preview of the effect graph is playing using that. 

same depth at the same visual data the edge data Correct, and we actually have in our software. We also have a preview tab, which will allow you to preview What you know kind of a more photo accurate representation of the scene would look like while you’re in the software Let’s uh, let’s also give our favorite digital fade a try and then we’ll .

switch back to our color there and Publish that and then what can you do after you’ve published the scene that kind of tweaks? Can you do to this animation? Yeah, so I actually have this MIDI controller hooked up. So we support OSC as you saw before MIDI and then other kind of input/output devices so I can actually interactively adjust the speed and color of this effect So I can change it to more like a yellowish or a reddish And then actually speed it up or .

slow it down a little bit So we could spend a fair amount of time going through and actually mapping the various components and portions of this device, but we end up with You know quite a few effects and we also have stock video integration So yeah, as I would say everything here so far has been rendered. Yeah, but let’s see what video looks like or yep So I’ll go ahead and I can import a label into our software Cool, and so now we’ve got kind of what .

you would expect for a traditional Mapping tool so I can actually go ahead and I can then Adjust our mask. I can go ahead and adjust the underlying shape or structure. And so if I wanted to actually Add a little call-out so say we’re doing like a museum exhibit. Mm-hmm for our for our collection Go ahead and map that Luckily, I have this nice grid on the pegboard there They can help me align it Otherwise mapping text is like a pretty difficult problem just to .

get you know your alignment with your scene properly but having that underlying scan there is really handy for this and the type of transforms you can do or beyond just the skew Correct. Yeah, so I could actually also scale the content within this. We also have like a mesh warping So say you had like a cylindrical or spherical object, you know, you could actually take your content and warp it around that very easily You know seeing text there occurs to me that. 

Dynamic content to be really interesting. Yes right now its effects which is rendering real-time and imported images but what type of dynamic content could potentially put on this so We’ll have a fair amount of Social media integrations so you could display your Twitter feed. Yeah and have you know, say you’re a business and you wanted to Display a hashtag you could actually have that kind of integrated and then that would be dynamically populated .