I tried to add a shorter version of this comment to the announcement of the Pioneer award winners. But I kept getting a message about my post triggering the spam filter so I'm posting it here as a new message. It's quite a bit longer than the original because this is really bothering me.
I was fine with not winning the top tier prize awards. I recognize that my app was not "pretty" and I'm sure that was one of the minimum threshold requirements for the top tier prizes. However that changed with the announcement of the Pioneer awards, especially since the declared criteria for the the Pioneer award is as follows:
"The Checkpoint Award recognizes developer profiles which showed Clarity of Idea, an overall grasp of Perceptual Computing and a clear understanding of Intel's SDK."
My app plays Rock, Paper, Scissors with you while a real robotic head talks to you as you play. In addition there was a live chat mode where you could talk with the head while it responded to you with funny GLADoS quips in response to your voice queries. (GLADoS is the hostile AI presence in the game Portal). To accomplish this the following technologies were integrated and implemented in the app:
- Nuance Text To Speech that spoke to you while training you how to play the game
- The open source Festival Text To Speech engine that was used for the voice of GLADoS
- Nuance speech recognition to recognize your speech during live chat mode
- PercSDK machine vision to detect when you shook your head NO or nodded your head YES
- Gesture recognition to detect the Rock, Paper, Scissors, and Palm gestures. This included custom pattern recognition work because those gestures are not detected by the PercSDK.
- Chatbot technology used during live chat mode
In addition, the integration of the animatronic robot head was a unique and advanced technology implementation in itself. The app is capable of being "fed" any block of text, and the app chooses appropriate facial expressions, eye movements, and jaw animations automatically to animate the head while the Festival TTS engine waveform plays. The reason I used the Festival TTS engine for the robot head was the detailed timing files the Festival engine produces. These timing files contain the time offsets for each phoneme into the synthesized waveform, thus allowing my app to tightly animate the jaw and sync it to the waveform. This is why the Nuance TTS engine was relegated to the training mode since that engine does not produce those timing files, however it was indeed used in the app. It was a lucky coincidence that the Festival TTS also does a fairly good and funny "impression" of the GLADoS AI voice.
I feel insulted that my app doesn't meet the criteria indicated for the Pioneer award. How does an app that uses the PercSDK speech recognition, text to speech, machine vision, custom gesture recognition (depth camera usage and finger geometry usage), fail to show that I have clear understanding of the Intel PercSDK or that I don't understand Perceptual Computing? What PercSDK core tech did I leave out? And how does an app strictly focused on providing an advanced robotic head that talks to you while playing Rock, Paper, Scissors fail to be perceived as a clear idea? On top of all that is a custom robotic rig with 3 cooperating modules that independently control eye movement, face motors, and jaw movements to intelligently synthesize lip-sync'ed facial expressions to dynamically synthesized speech waveforms, with all servos independently coordinated by the app in real time. How does this not meet the criteria for a Pioneer? As the old joke says, I certainly feel like one now because they "get all the arrows in the back".
So my reward for months of effort on my admittedly "ugly" app is Intel telling me that I don't know their SDK and my app didn't show any innovation or clarity for the ideas it implemented. I guess it's time to download the Kinect SDK.