algorithmic modeling for Rhino
Now using two webcams (the other is above, out of view) to get 3D info out of my gestures. Grasshopper detects which object in the Rhino scene I am closest to, and then by pausing on that object, it is automatically selected in Rhino itself (moving out of camera view deselects everything). The goal of this is to create gestural workability within the pre-existing Rhino commands.
Major credit goes to Andy Payne for his Firefly components, and to Andrew Heumann and Chris Tietjen for scripting help.
Tags:
Comment
Very cool. BTW, you're going to love the LEAP controller :)
Thanks - I've fooled around a bit with gestural stuff before: http://www.grasshopper3d.com/video/object-creation-line-and-extrusion
I might go back to it eventually, but at the moment I'm trying to avoid setting up a preprogrammed set of gestural commands. It would be feasible to do a simple set, but Rhino's complexity of commands is just too great to be completely gestural....
Great work!
How about some movement based gestures as well - such as tapping or circling ?
Very Impressive dude, minority report. Though i'm slightly distracted by that book. Its Huge!
Welcome to
Grasshopper
Added by Parametric House 0 Comments 0 Likes
Added by Parametric House 0 Comments 0 Likes
Added by Parametric House 0 Comments 0 Likes
Added by Parametric House 0 Comments 0 Likes
© 2024 Created by Scott Davidson. Powered by
You need to be a member of Grasshopper to add comments!