Difference between revisions of "FingerSpark"

From ESE205 Wiki
Jump to navigation Jump to search
Line 12: Line 12:
 
Our goal in creating FingerSpark is to work towards eliminating the barriers to perfectly natural user control of electronics.  We believe that our product will be an essential next step in developing three-dimensional operating systems, creating robots that can flawlessly mimic the fine motor skills of humans, and producing interactive augmented reality technologies.
 
Our goal in creating FingerSpark is to work towards eliminating the barriers to perfectly natural user control of electronics.  We believe that our product will be an essential next step in developing three-dimensional operating systems, creating robots that can flawlessly mimic the fine motor skills of humans, and producing interactive augmented reality technologies.
  
Our demonstration at the end of the semester will consist of us demonstrating the ability to move our hand in the glove with colored fingertips in front of the camera, while a monitor reflects the location of each fingertip in two-dimensional space with each fingertip represented as a point on a uniformly colored background updating in real timeFor our demo, we will develop a simple coloring app that will allow the user to “paint” different colors onscreen using our product.
+
Our demonstration at the end of the semester will consist of us demonstrating the ability to move our hand in the glove with colored fingertips in front of the camera, while our program processes the video input and identifies the gesture made by the userOnce the gesture is detected, our program will output what type of gesture the user made.
  
  

Revision as of 18:45, 7 April 2016

Overview

FingerSpark, our product, will be a program that will allow a camera to track a user’s hand and individual fingers. The user will wear a glove with different colored fingertips, making it easier for the camera to pinpoint the two-dimensional location of each finger. To use our product, the user will position his hand 2-3 feet in front of the camera, with a white backdrop behind his hand. The camera will also utilize a Raspberry Pi B+ CPU, which will run our program and gather the data from the camera, forming an integrated system that will be covered by a 3-D printed plastic case and mounted on a tripod for the user’s convenience. The Raspberry Pi Camera Module has modes that support 1080p video at 30 frames/second and 720p video at 60 frames/second, either of which would be more than sufficient to serve the purpose of our project (detecting the movement of brightly colored points at approximately 2-3 feet away from the camera).

To achieve the desired functionality, we plan to use an image processing program to process individual frames from the camera’s video feed which we will then analyze to determine the coordinates of each of the user’s fingers. If we are able to achieve this goal with a substantial amount of time before the deadline, we hope to begin work on a program to interpret human gestures using this data. Transforming the finger coordinates to gestures will require real-time extrapolation of changes in the x- and y-locations of each finger while allowing for slight inaccuracies in the movement of the user's hand.

Team Members

  • David Battel
  • Connor Goggins
  • Kjartan Brownell (TA)

Objective

Our goal in creating FingerSpark is to work towards eliminating the barriers to perfectly natural user control of electronics. We believe that our product will be an essential next step in developing three-dimensional operating systems, creating robots that can flawlessly mimic the fine motor skills of humans, and producing interactive augmented reality technologies.

Our demonstration at the end of the semester will consist of us demonstrating the ability to move our hand in the glove with colored fingertips in front of the camera, while our program processes the video input and identifies the gesture made by the user. Once the gesture is detected, our program will output what type of gesture the user made.


Budget

  • Tripod: $19.99 (Need to purchase)

TOTAL: $86.25


Gantt Chart

The Gant Chart for our project.