[Brian Korsedal] and his company Arcology Now! have developed a great geodesic building system which makes architectural structures that aren’t just limited to domes. They 3D scan the terrain, generate plans, and make geodesic steel space frame structures which are easy to assemble and can be in any shape imaginable.
Their clever design software can create any shape and incorporate uneven terrains into the plans. The structures are really easy to construct with basic tools, and assembly is extremely straight forward because the pole labels are generated by the design software. Watch this construction time lapse video.
At the moment, ordering a structure fabricated by the company is your only option. But it shouldn’t be too hard to fabricate something similar if you have access to a hackerspace. It may even be worth getting in touch with Arcology now! as they do seem happy collaborating to make art like the Amyloid Project, and architectural structures for public spaces and festivals like Lucidity. Find out what they are up to on the Arcology Now! Facebook page.
Would this be perfect for what you’ve been thinking about building? Let us know what that ‘something’ is in the comments below. Continue reading “Geodesic Structures that aren’t just Domes”
Ever heard of DesignSpark? They are releasing a powerful CAD package on September 16th — for free!
The company is owned by RS Components, a distributor of electronics and maintenance products. They offer a large library of 3D models of parts that they sell, dubbed the ModelSource. So if you are wondering how they are giving out software for free, that’s how. They also have free PCB designing software, and something called DesignShare which hosts open-source project collaboration, sharing and discussions.
By the looks of the demo video, DesignSpark Mechanical is a well laid out CAD package that is rich in features. The software allows for the import and export of several file types, and it looks like ECAD, OBJ, Sketchup, STEP, DXF and STL are all there, as well as the native file types. While it looks like you can import any files, we are willing to bet adding ModelSource files are by far the easiest and most convenient because of the integrated ModelSource library. But we think that’s a small price to pay for an alternative to SketchUp. After all, the component models will be useful for assemblies, even if you don’t order through them. Oh, and it’s perfect for making free models for 3D printing as it includes the ability to export STL files.
Watch the software demo after the break.
Continue reading “DesignSpark Mechanical – The Gift of Invention”
As 3D printing continues to grow, people are developing more and more ways to get 3D models. From the hardware based scanners like the Microsoft Kinect to software based like 123D Catch there are a lot of ways to create a 3D model from a series of images. But what if you could make a 3D model out of a single image? Sound crazy? Maybe not. A team of researchers have created 3-Sweep, an interactive technique for turning objects in 2D images into 3D models that can be manipulated.
To be clear, the recognition of 3D components within a single image is a bit out of reach for computer algorithms alone. But by combining the cognitive abilities of a person with the computational accuracy of a computer they have been able to create a very simple tool for extracting 3D models. This is done by outlining the shape similar to how one might model in a CAD package — once the outline is complete, the algorithm takes over and creates a model.
The software was debuted at Siggraph Asia 2013 and has caused quite a stir on the internet. Watch the fascinating video that demonstrates the software process after the break!
Continue reading “3-Sweep: Turning 2D images into 3D models”
Anyone can grab a projector, plug it in, and fire a movie at the wall. If, however, you want to add some depth to your work–both metaphorical and physical–you’d better start projection mapping. Intricate surfaces like these slabs of styrofoam are excellent candidates for a stunning display, but not without introducing additional complexity to your setup. [Grady] hopes to alleviate some tedium with the TightLight (Warning: “music”).
The video shows the entire mapping process of which the Arduino plays a specific role toward the end. Before tackling any projector calibration, [Grady] needs an accurate 3D model of the projection surface, and boy does it look complicated. Good thing he has a NextEngine 3D laser scanner, which you’ll see lighting the surface red as it cruises along.
Enter the TightLight: essentially 20 CdS photocells hooked up to a Duemilanove, each of which is placed at a previously-marked point on the 3D surface. A quick calibration scan scrolls light from the projector across the X then Y axis, hitting each sensor to determine its exact position. [Grady] then merges the photocell location data with the earlier 3D model using the TouchDesigner platform, and bam: everything lines up and plays nice.
Robots can easily make their way across a factory floor; with painted lines on the floor, a factory makes for an ideal environment for a robot to navigate. A much more difficult test of computer vision lies in your living room. Finding a way around a coffee table and not knocking over a lamp present a huge challenge for any autonomous robot. Researchers at the Royal Institute of Technology in Sweden are working on this problem, but they need your help.
[Alper Aydemir], [Rasmus Göransson] and Prof. [Patric Jensfelt] at the Centre for Autonomous Systems in Stockholm created Kinect@Home. The idea is simple: by modeling hundreds of living rooms in 3D, the computer vision and robotics researchers will have a fantastic library to train their algorithms.
To help out the Kinect@Home team, all that is needed is a Kinect, just like the one lying disused in your cupboard. After signing up on the Kinect@Home site, you’re able to create a 3D model of your living room, den, or office right in your browser. This 3D model is then added to the Kinect@Home library for CV researchers around the world.
September is coming, and soon college freshmen the world over will be decorating their dorm room walls with Dark Side of the Moon posters and [M.C. Escher] prints. Anyone can go out and simply buy a prism, but what if you wanted a real-life version of objects and buildings from [Escher]’s universe? Professor [Gershon Elber] at the Technion at the Israel Institute of Technology decided to turn [Escher]’s prints into reality.
First beginning with simple shapes such as a Penrose Triangle and a Necker Cube, [Elber] decided to branch out into much more impossible shapes such as [Escher]’s Waterfall, Belvedere, and Relativity. These buildings are extremely hard to visualize in any traditional computer design program, so [Elber] wrote a plugin for his IRIT computer modeling program to design the buildings before committing them to a 3D printer.
In the video after the break, you can see a few rotating views of the resulting [Escher] buildings. Of course they only work from exactly one point of view – and even then, only with one eye closed – but it’s amazing to see these famous architectural studies brought into the real world.
Continue reading “Turning [M. C. Escher] prints into real objects”
Here’s an oldie but a goodie that passed us up the first time it went around the Internet. [Qi Pan], (former) PhD student at Cambridge, made a 3D modeling program using only a simple webcam. Not only does this make very fast work of building 3D models, the real texture is also rendered on the virtual object.
The project is called ProFORMA, and to get some idea of exactly how fast it is, the model of a church seen above was captured and rendered in a little over a minute. To get the incredible speed of ProFORMA, [Qi] had his webcam take a series of keyframes. When the model is rotated about 10°, another keyframe is taken and the corners are triangulated with some very fancy math.
Even though [Qi]’s project is from 2009, it seems like it would be better than the ReconstructMe, the Kinect-able 3D scanning we saw a while ago. There’s a great video of [Qi] modeling a papercraft church after the break, but check out the actual paper for a better idea of how ProFORMA works.
Continue reading “Getting a textured 3D scan from just a webcam”