Skip to content

Latest commit

 

History

History
69 lines (39 loc) · 1.98 KB

README.md

File metadata and controls

69 lines (39 loc) · 1.98 KB

Handzplay

handzplay.mp4

Handzplay is a playground for hand gesture interactions. You can create gesture interactions by adding rules and interfaces.

Experiment by Seungmee Lee Contact via Twitter or Instagram for any questions.

How it works

Your hands in a camera feed are recognized and visualized by TensorFlow Handpose. You can try pre-built templates with a set of rules and interfaces, or you can create your own.

Currently, there're 3 ways of creating conditions.

Add a rule

create_rule.mp4

It's an 'if-then' structure rule that you can define a type of gesture to trigger the interaction, and a type of reaction to be triggered.
For trigger,

  • fingers : Choose two fingers and the distance between them. If they're within the range, the interaction is triggered.
  • pose : Choose which hand (left, right, or both) and the pose.
  • action : Choose which hand (left, right, or both) and the swiping direction in the air.

For reaction,

  • audio
  • draw
  • voice
  • shape
  • stamp

Add an interface

interface-trim.mp4

You can create a visual element that reacts to actions, like swiping, hover, and click, almost similar to how you'd interact with interfaces with a mouse.

Canvas mode (beta)

canvasmode.mp4

Canvas mode offers the easiest way to create an interface, but now it only supports hover interaction at this moment.



References

Tensorflow Handpose doc
Fingerpose doc