Microsoft Project Oxford and VOIS

Microsoft have brought out a new face and emotion recognition library called Project Oxford that performs most of the functions of the C# library I was developing for VOIS, therefore I’ll be re-engineering my demonstration to take account of this new library.

microsofts-project-oxford-can-now-detect-emotions-in-photos

At first glance though, the Microsoft library performs much better than my own, especially with poorer quality images. Having said that, I don’t know what the pricing structure for the library is.

If it is expensive, then I may revert back to my own library and eventually release it on SourceForge.

VOIS Development Part 3 – Procedure for recording emotions

It’s been a long time since I’ve previously posted about VOIS. I have a new job and this has taken up most of my time.

At the end of September I’m undertaking a Masters at Huddersfield University so I’m definitely hoping the technologies behind VOIS could be taken forward with the help of the University’s facilities.

So given everything above my work on VOIS has been limited, so I’ve been thinking more about when to use audio-based emotion recognition and when to use video-based emotion recognition. The diagram below is my current thoughts which I’ll be testing later on in the year when I have time.

Emotion process

VOIS Development Part 2 – Choosing an Audio-based Emotion Recogniser

Been pretty busy since my last post and have:

  • Improved the EigenFace based facial emotion recognition by continually streamlining the facial dataset I use. This has resulted in improved results, although in the future, I think I will still need to move to a FACS (Facial Action Coding System) variant.
  • Decided to shelve my plans for using Xamarin for now, due to the increased workload it puts on me, not to mention the fact I’ll need to upgrade my old macbook which now is too old for the latest Xcode build.
  • Created a windows phone variant of the original tablet app I created, running on my new Nokia Lumia 630.

I’m now moving onto coming up with a plan for how I will integrate audio-based emotion recognition into VOIS. Having reviewed a number of interesting papers I’m going to implement the approach outlined in the paper “Perceptual cues in nonverbal vocal expressions of emotion”, UK published in 2010.

audio1

The paper outlines measurements in acoustic cues to define emotion.

I’ll provide another update on how successful I’ve been implementing the above approach and whether I can marry up the visual emotion feed I currently have, with a new audio-based one. At the moment I can’t see anyone that has done this successfully so it will certainly be a real triumph if it’s a success.

Bye for now..

Microsoft Build 2015 – Hololens for Education

It’s great that Microsoft are pushing education as a major stream for Hololens, I would though have probably hoped for a slightly better Build 2015 demo as what was demonstrated, something with location context sensitivity.

But still, good times ahead and I’m looking forward to the day I can try out VOIS with Hololens!

VOIS Development Part I – Facial and Emotion Recognition

Initially my VOIS development involves creating a series of prototypes that demonstrate each part of the complex functionality required, such as: facial recognition, video/audio emotion recognition, speech recognition etc.

First to tackle has been developing a mobile app for facial and emotion recognition.

VOIS1

Detecting Andy and his facial expressions

The app (shown in the screenshot above) was developed in C# and uses the concept of EigenFaces to initially detect someone the user knows (from a library they must accrue on their mobile device) then as the person talks the app will detect the emotion in their facial expressions.

Currently the app runs as a windows tablet app with my next task being to re-develop it in xamarin, so that it can be published to Android or iOS devices.

Bye for now.

An introduction to VOIS (The Visual Ontological Imitation System)

Maybe there are people who just don’t get where you’re coming from. This is an everyday reality for people with autism, but it is not just disabled people who need help with understanding emotion in others.

VOIS is an innovative design for an application (the brainchild of a friend of mine – Jonathan Bishop), which will assist autistic people in recognising the facial moods of people they are talking to and suggest appropriate responses.

Given that VOIS will work irrespective of what language is being spoken, there are obviously cross-over opportunities to use it in areas such as:

  • Defence
    Soldiers who have regular contact with, say, a tribal elder, could use it to see whether the elder is being evasive as well as how well his mood changes over time.
  • Security
    During interrogation of suspected terrorists, along with standard questioning, VOIS could pick-up evasiveness and suggest more questions in certain areas.
  • Immigration
    Again VOIS could help in questioning asylum seekers here too.

Future versions of VOIS could be used via a head camera or fixed camera for surveillance roles. Jonathan is himself autistic.

I’ve agreed to help Jonathan with the development of a prototype version that will run on a range of mobile devices and intend on charting my progress via my blog.

VOIS screen

 

openEar – Open Source Emotion Recognition Toolkit

I’ve been looking at using the openEar open source emotion recognition toolkit within a Windows 8 development, in the first case, to create a simplistic emotion recognition interface, that could be used within a military simulation or even a smartphone game.

emotional-smartphone-2

At the moment I’m getting my head around it, from within Visual Studio, but once I’ve saved some pennies I will try and integrate it into Xamarin to try and create a cross platform smartphone library.

Expressing yourself in a virtual world

Earlier in the year I attended a small C2ISTAR workshop at Qinetiq where I discussed the usefulness of collaborative virtual worlds such as VastPark, within military training. I was asked at the time if there was anyway of improving facial responses in avatars along with improved gesturing.

So I was very interested when I came across a tool from Mixamo called Face Plus, which can capture a users expressions using their webcam and use that within a simulation in realtime.

http://www.youtube.com/watch?v=yNaympgBVpQ

I think if this type of system could be incorporated into a collaborative virtual world to help improve key areas of training such as cultural awareness and the issues around the contemporary operating environment. Sure network traffic (sending/receiving facial gestures) issues would need to be overcome, but still, I see real potential for training applications.

Chalk and Talk in a Virtual World

Over the past few years VBS2 has swiftly become pretty much the standard 3D simulation tool in use within most NATO countries, whether that be for training or analytical purposes.

It’s upcoming new release will see the ability to use simulation as a presentation background for instructors in classrooms via its new chalkboard feature. Information and associated military symbology can be drawn on top of any 3D terrain as well as on top of previously played scenarios.

chalk1

Giving instructors the ability to create visually engaging lessons that overwise might have been delivered with old fashioned powerpoint can really make for compelling viewing. Especially if an instructor could write on top of previous scenarios done by groups of students, highlighting correct and incorrect practices.

chalk2

I personally feel that this concept could be extended to produce high quality Computer Aided Instruction (CAI) quickly on a large scale.

chalk3

GameTweet

I was watching a geek Tv programme last night where a team developed something they called TwitterCat for a woman who wanted to keep tabs on her Tabby! They installed sensors (eg in the cat flap) that sent twitter feeds to Tabby’s twitter page.

It got me thinking whether any games publishers or virtual world platforms have done the same with their titles.

Say you’ve attained a certain accreditation within the game, it could tweet it to your twitter page or send an accreditation post to your facebook/google+ account.

The same process could also be used as a within business systems employing Gamification/Captology priniciples.

If this is already been done, I’d love to hear from anyone who can give me more information.