Authors of Project: Heeded
Lillian Gales, Evan Brecht-Curry, Eric Huang, Brad Elliot, Chance Harmon
-
Description: Using Facial Recognition, develop a formula to be able to report the level of engagement from a group to assist the user in assessing the success in there speech.
-
Domain: As a speaker or instructor, I would like to improve my skills in captivating and holding the attention of my audience. Current technology depends on indvidual feedback from the audience, after the speech is already delivered. I would like to be able to have realtime feedback to maximize my delivery.
-
Resultions: Using new hardware, and newly developed software, we were able to capture the audience engagement in realtime, and return feedback to the user for them to make adjustments as needed.
- Rasberry Pi
- Webcam
- Node.js
- AWS Cognito
- AWS S3 Bucket
- AWS Rekognition
Have our Pi, a webcam, an AWS account, a verifiable email, and fortitude.
By intializing localhost3000, you will be directed to AWS Cognito to sign in or sign up. From there it will take you the dynamic results page where will begin to see live feedback. The process is that once Oauth is completed, it triggers the Pi to start taking pictures, and begins sending those photos to the S3 bucket. As soon the S3 Bucket recieves a photo, our function retrieves the photo, passes it along to AWS Rekognize for facial analyisis. Rekognize sends us back a JSON object of the analyisis, we parse through the data, picking out what we have determined to be our magic formula. After we pass the data through the formula, the ouput value is plotted on a graph in realtime, using an aggregate of how many faces deteced versus how many are engaged.