the story of Gunnar, David and Schorsch…

Kubentypen03This isn’t Gunnar but made by Gunnar. Gunnar is a friend who makes actionfigures and just made that little guy. He is made of wood and around 50mm high.

Kubentypen02This isn’t David but it’s scanned using the David-Laserscannersoftware with an Optoma PK301 and a Microsoft Lifecam-Studio. It took about 1.5hrs to get the hole figure but i think you can make it in 30min. Here he sits in Blender, From 750k faces (50MB out of David) to 10% (75k faces).

Kubentypen04This isn’t Schorsch but SchorschXY printed it.

kisslicer:

layers_0.2mm

speed_20mm/s

ABS

That little guy hasn’t got a name jet, i think. How about “Lampshade”:

Kubentypen05

the last picture shows the two Prints of Lampshade a little better than the earlier one:

Kubentypen01

The white spots are where the support stucked on the model, you can easy get rid of that using acetone and your finger… A few missing perimeters are caused by low-quality-ABS (waterbubbles) and single-walled-perimeter (left one) The right one is a reduced one with only ~7500 triangles.

MK7 sneaks in

schorschxy24

E3D v4

GRRF modular

MK7  (ON 20 VOLTS!) only got 190°C here… (data send by a friend)

description (older post)

each square = 10° x 10s

Starting @ 20°C

Stopped @ 240°C

E3D v4: 120sek

GRRF modular: 350sek

shure it doesn’t say anything about printing-quality, but when i look at that GRRF-curve i doubt that it makes 270°C in a non-ridiculous amount of time… and i think they might have made it for PLA-only or 24V.

For this test i just startet my RAMPS1.4 with Marlin rc3 and standard PID. I don’t know how to adjust my PID so you can see it on the Table near 240°C.

more kinect-scans…

In order to understand how to increase Kinect-Fusion-scans i tried several things today.

First i wanted to digitize my bike:

06

First thing to say here is: that’s not a bike! the wheels are half missing! That’s right, it’s missing cause i set the volume to scan a little smaller than my bike. You can make it bigger but you will loose resolution: with my computer it can handle 640 X 640 X 512 = 209715200 Voxels. Put that into a cube of 1 m³ and you end up @ >2mm/Voxel.

09

08

5

thats it for now, next time it’s schorschtime again…

…’bout the resolution, baby…

05

Due to structured light the hand is recognized better than the keyboard (wich looks inverted here).

Smoother surfaces are reproduced more accurate.

It’s hard to tell about the resolution as it seems quite nice at the hand, good at the (reflecting) table and ‘funny'(nerdy) at the Keyboard at the same time! I have to measure it but it’s obvious that edges won’t come up like planes, the lesson of structured light?

Scan took about 20-30sek resulting in a ~50mb stl. snanned by me, not my desktop or hand.

Finally: Kinect Fusion

 

 

 

kinectfusio01

You may have heard that ‘Kinect-Fusion’ as seen at siggraph 2011 is now available through Microsofts Kinect SDK. It’s now possible to scan larger objects with just a Kinect. I use a Kinect for Xbox 360 (with Poweradapter, for old Xbox360) on a dualcore AMD Athlon64 x2 5200+ @2.71GHz and a GTX480. With that setup it is not simple to scan a hole human cause if you try to move the kinect around the person it looses tracking an rebuilds whatever you had already scanned.  nils5

The original resolution of the scan is around 2-3mm as seen on the next Picture:

nils4After a bit of playing around with the Demo i guess you have to adjust Clipping and resolution as well as the ‘dynamic’ of the object you want to scan. If i set all faders to maximum i end up with quite a small scanvolume. I dont know if that is due to my 1.5GB-Nvidia-Card or the Athlon or because it’s an Xbox360-Kinect.

I didn’t find a method to scan smaller objects by rotating them in front of the camera jet but this one: set min and max clipping to ‘isolate’ your object, than the software can only use the object itself to reorientate each frame:

micha1This picture of me is taken by myself using a static Kinect and an office-chair.