We’ve written a lot about artificial intelligence (AI) here at Nanalyze, and just when we feel like there’s not much more we can add to the topic, we find loads more interesting companies to write about. There has been a lot of talk lately about how machines just won’t be able to capture that “human element” of emotions or “emotional intelligence” as it is often called. The act of building an emotional quotient or EQ as a layer on top of AI is being referred to as
affective computing, a topic we covered before. The first step towards AI being able to demonstrate emotional intelligence, is that it needs to see emotions in our behaviour, hear our voices, and feel our anxieties. To do this, AI must be able to extract emotional cues or data from us through conventional means like eye tracking, galvanic skin response, voice and
written word analysis,
brain activity via EEG, facial mapping, and even gait analysis.
Here are 7 startups attempting to seamlessly integrate AI, machine learning, and emotional intelligence.
Founded in 2012, Israeli startup Beyond Verbal has taken in $10.1 million in 4 rounds of funding to develop a technology that “analyzes emotions from vocal intonations”. The Company uses a simple microphone and a set of questions to elicit 11 emotions that describe a person’s character traits. This form of emotional analytics “listens” to vocal intonations as they are spoken. The technology is intended for marketing professionals to run customer service experience and communication campaigns while measuring things like:
In addition to marketing, there are also possible healthcare applications for this technology. The Mayo Clinic is using Beyond Verbal to identify vocal biomarkers that can help identify Coronary Artery Disease (CAD), Autism Spectrum Disorder (ASD), and Parkinson’s Disease (PD) with the belief that this technology can help provide an early diagnosis. With 2.3 million recorded voices across 170 countries and 21 years of research, they might just be in a position to make the FDA pay attention.
Want to see a demo? Here’s the Beyond Verbal technology analyzing Donald Trump and Megyn Kelly go at it:
Don’t read too much into our choice of a demo video here. We simply took the one demo off of the Company website that we thought you’d be most likely to click. If you want to try the technology on yourself, they have a free app called Moodies that you can speak into and it will tell you what emotions you’re exhibiting. Any newly minted MBAs reading this should use that technology to build a breathalyzer that you can speak into on your smartphone, then send it to us so we can use it on our writers when they come in to work in the morning.
Founded in 2005, nViso is a privately-held Swiss startup with a $750K grant which was originally meant to develop an automatic prediction process for accurately categorizing patients needing tracheal intubation (plastic tube down the throat) for surgeries involving general anesthesia, a very costly procedure. Like CrowdEmotion, nViso’s technology tracks the movement of 43 facial muscles using a simple webcam and then uses AI to interpret your emotions.
The Bank of New Zealand tried nViso’s technology on about 200,000 people (about 6% of all adults in New Zealand) and found that they rose from a #4 position to the #1 most preferred bank consumers would move to. BNZ also moved from #5 to #2 national rankings in terms of being a good place to manage clients money when it deployed nViso’s
EmotionAdvisor. It’s kind of hard to see how that would even be possible but here’s
a case study with all the details.
While the tone of your voice may disclose how you’re feeling, did you know that what you write can as well? Founded in 2014, Toronto-based startup, Receptiviti, has taken in $690K in total funding to develop a proprietary technology they called Linguistic Inquiry and Word Count or LIWC2015 which looks at what you write and then uses it to gather insights about your character, emotions, level of deception, or even how you make decisions. The Company uses a branch of artificial intelligence called Natural Language Processing (NLP) techniques to capture people’s emotions, social concerns, thinking styles, psychology, and even their use of parts of speech. Here’s an analysis they did on some of the literary genius produced by Taylor Swift on Twitter:
Receptiviti analysis of Taylor Swift’s Tweets from Jen Golbeck as it appears in Psychology Today.
While one obvious use might be to
make chatbots actually tolerable, the technology can be used as a foundation for doing some serious armchair psychology on customers, employees, job applicants or even the president. That’s right, as it turns out
President Trump’s tweets are likely coming from two different sources. We decided to ask Receptiviti what they thought about our commentary on them and here are the three attributes they came back with:
We didn’t see ridiculously good looking in there but aside from that, pretty much spot on. You can test Receptiviti yourself and see what it has to say about what you write. Go ahead and kick the tires here.
brFounded in 2016, New York startup BRAIQ has taken in $120K so far to develop technology that detects how you feel about that autonomous car that’s carting you around. The value proposition here was so compelling that BRAIQ was accepted into the techcenter program of Plug and Play, a global innovation platform that has graduated the likes of PayPal, Dropbox, and Lending Club with an admission rate tougher than Harvard (2%). While most of the design philosophies of self-driving vehicles are centered on how the vehicle interacts with its environment, BRAIQ is looking into how autonomous vehicles should interact with the passengers.
Car manufacturers know 75% of Americans are actually “afraid’ of self-driving cars and don’t trust them. BRAIQ hopes to earn this trust by allowing their technology to intuitively read emotional signals hoping you enjoy the ride more than worrying about getting there. BRAIQ sees $4 billion of funding commitments on autonomous driving research over the next 10 years that they can tap into. With an estimated 20 million fully autonomous vehicles hitting roads by 2025, they hope their technology will be ubiquitous and enabled by hidden sensors so you’ll never even know it’s there.
Founded in 2015, Toronto startup NuraLogix started with a $115K
Idea to Innovation (I2I) grant to develop a technology that can read human emotions and can even tell if you’re lying. The startup developed a technique to “read” human emotional state called Transdermal Optical Imaging™ (TOI™) using a conventional video camera to extract information from the blood flow underneath the human face. Contrary to what Pinocchio was told, it is not the growing nose itself that can determine if you’re actually lying but the blood flow around the area of the nose. The technology has applications in the field of marketing, research, and health with the software expected to be deployed into the Cloud this year. A mobile phone version is still years down the road so you can still safely answer the “honey do I look fat in these” questions without any backlash.
Founded in 2014, New York startup Emoshape has taken in $150k so far to develop what they are calling an Emotion Processing Unit (EPU). The startup aims to provide intelligent objects like robots, self-driving cars, affective toys or even ordinary consumer electronic products with the capability to interact with humans via a microchip they call the Emotion Processing Unit (EPU).
The chip will use algorithms to identify any combination of the twelve primary emotions, allowing an intelligent device to manifest potentially 64 trillion distinct emotional states. The EPU is available for purchase so go buy one and see what it can do yourself.
Founded in 2013, CrowdEmotion is a privately-held London-based startup with a technology to gauge human emotions visually by mapping the movements of 43 muscles on the human face. BBC StoryWorks used this technology in a 2016 study called ‘The
Science of Engagement‘ where they measured the second-by-second facial movements of 5,153 people while viewing English language international news in six of BBC’s key markets. CrowdEmotion currently offers the technology over the cloud and has a fair number of clients already including, of course, the BBC. You can
upload a video to their website and then it will give you some running commentary on the sort of emotions you’re exhibiting.
Our initial thoughts were that because you know you are being observed, it will make you behave differently, so this technology will probably be most effective when applied undetected. That’s not exactly true though.
According to a subject matter expert we spoke to, “micro facial expressions are really difficult to hide consciously. We actually only see impact on the first 1-2 seconds if at all. This has been consistent across billions of facial data points“.
The very discussion though of emotion analysis that is hidden raises privacy concerns.
We discussed this with Matthew Celuszak, CEO of CrowdEmotion, and his comments were that “in most countries, emotions are or will be treated as personal data and it is illegal to capture these without notification
and/or consent. How emotion companies handle CCTV data is a bit of a grey area in how notification is done, but turning on a camera or microphone without consent is a sure fire way to kill your company“. This is certainly something to be aware of for
companies like BRAIQ that want to observe passenger emotions undetected. The next time you communicate with a smart device, think about that fact that you may be talking to something that reads emotions better than you do.
Looking to buy shares in companies before they IPO? A company called Motif Investing lets you buy pre-IPO shares in companies that are led by JP Morgan. You can
open an account with Motif with no deposit required so that you are ready to buy pre-IPO shares when they are offered.
@RebmannResearch (Health Econ)
@Revuze (AI brand mgmt.)
@2breathe4sleep (Sleeptech based on ancient science of Breathe)