Versions Compared

Key

  • This line was added.
  • This line was removed.
  • Formatting was changed.

Using this feature, you can enable your Digital Person to be more alive and engaging by using their hands and bodies to interact with the user. This includes pointing to objects or content in their digital environment they would like the user to focus on.

Info

Note: Human 0S 2-2.2 includes shoulder gesturing, and Human OS 2.2 onwards includes full hand and arm gestures.

As it speaks, the Every Digital Person can see you, hear you, respond to your emotions, and has an inner emotional state. This is their base behavior. With Real-Time Gesturing, your Digital Person analyzes what it’s saying via Natural Language Processing (NLP), and adds emotionally appropriate gesturing and emotionally expressive behavior to their speech in real-time.

Real-Time Gesturing is a powerful tool for conversation writers and designers, as the . The words , they write to bring the digital person’s behavior to life. For example, if a Digital Person’s conversation is sad in tone, their behavior will autonomously express sadness and concern. If they are talking about something surprising and delightful, their behavior will express surprise and happiness

Components of Real-time Gesturing

Mood and inner state

The digital person will always return to their base mood, but their inner state is affected from moment to moment by the content of what the digital person is saying themselves, and the signals they are picking up from the user and the environment.

Emotional gestures

Emotional gestures are the facial expressions the Digital Person performs in accordance with the emotional content of the text they’re speaking. This may include smiles, frowns, concerned expressions and head tilts, and so on. These are associated with the emotional tone of the text being delivered.

Symbolic gestures

Symbolic gestures are gestures that relate closely in meaning to some word or expression being spoken by the Digital Person. For example, a gesture in which the Person opens their arms out wide when they’re speaking a phrase such as “everybody” or “all of them.” These may vary in the degree to which the gesture and the phrase are semantically tied in meaning-- some might be very tightly linked (think of pointing towards yourself when speaking the words “me” or “I”) while others may be much more loosely associated (e.g. a ‘smile shrug’ gesture with words like ‘wonderful’ or ‘warmth’).

Beat gestures

Beat gestures are gestures that convey neither meaning nor emotion, but rather are the natural gestures produced in alignment with the rhythm or cadence of speech. Think of someone delivering a short speech and holding their arm slightly out, “beating” on each emphasized word. Or in more informal, day-to-day communication, the way one’s arms or shoulders might just move rhythmically in sync with what we’re saying, without having any real tie to the meaning or emotion of what we’re communicating

Head & Lip Sync

Lip sync refers to the animations that drive the movement of the Digital Person’s mouth, lips, and related facial muscles when speaking. This is aligned with the text-to-speech audio and phonemes as realistically as possible. Head sync refers to the movement of the Digital Person’s head in alignment with the audio to create realistic motion through the neck and head.

Empathetic Reaction to the User

The digital person also analyses the emotional content of the words the user is saying, and the expressions on their face, and produces a facial expression in reaction. Each behavior style has its own, behavior-appropriate reactive style. Please note: Reacting to User Speech is only available in Human OS 2.3+

Reacting to On Screen Content

When onscreen content is used with Digital People running Human OS 2.2+, the Digital Person will direct their gaze towards it, drawing draw the user’s attention, and gesture towards it depending on the screen layout.

Using Personality Tags with Real-Time Gesturing

Personality Tags can be used in conjunction with Real-Time Gesturing. We recommend they are used for words and phrases specific to your brand, product, or use-case, or where you want a performance that is indifferent to what is delivered by Real-Time Gesturing. For example,  the autonomous gesturing system will decide to associate the word “beautiful” with a happy expression but if you have defined a  thoughtful expression for this word using personality tags, then this is what your Digital Person will express instead.

Personality Tags will temporarily override Real-Time Gesturing to deliver the behavior you have specified. The personality tags Neutral Long (Sentence-Based) and Neutral Short (Word-Based) can be used to define sentences and words where your Digital Person will behave neutrally.

Image Removed Info

Notes:

The Real-Time Gesturing feature is only available in a Soul Machines Digital Person running Human OS 2.0+.

it with their hand and arm it if the frame and screen layout allow.

For more information about this feature visit the Content Awareness section.

Back-channelling behavior

  • When listening to an end user speak, the digital people display a range of nodding behaviors to acknowledge they have heard.

  • While accessing a reply the digital person then displays a range of thinking behaviors before responding.

  • Both are available in Human OS 2.6+.

This back-channeling behaviour is designed to help the end user have confidence they have been heard and will be answered. If the person speaking is not close to the microphone however, neither nodding or thinking will be triggered. The digital person will still hear and respond to speech.

Thinking.mov

Iconic gestures

Iconic gestures are gestures that closely relate to the semantic content of segments of speech. In general they are not linked to a specific word or expression, but to the meaning that is conveyed given the context of the speech. They visually represent a concept or object they are referring to, such as a Heart Sign representing passion or dedication, or a Peace Sign to illustrate taking a selfie during your vacation. This option will allow the digital person to automatically perform these more expressive gestures based on the language and the intention of the digital person. The feature is currently in beta and can be enabled using the toggle under Other behavioral settings.

b79e8b48-1c73-42ec-9ae2-095ead0298f6.pngImage Added

Info

You can temporarily override Real-Time Gesturing for specific sentences or words spoken by your Digital Person via the

Personality

Behavior Tags

facility. The personality tags

feature.

  • You can use the behavior tags to override the autonomous behavior with something more appropriate to your use-case. For example, you might want to add brand and product related words to Smiling Gesture (Word-Based).

  • You can also use Neutral Long (Sentence-Based) and Neutral Short (Word-Based)

can be used

  • to define sentences and words that will

make your Digital Person act neutral, temporarily deactivating the autonomous gesturing feature.
  • In addition, the personality tags you have defined take priority and are used instead of the autonomous gesturing system. For example, the autonomous gesturing system can decide that the word “opulent” is associated with a happy expression but if you have defined a thoughtful expression for this word in the personality tags facility, then this is what your Digital Person will express instead.

  • Related Topics

    Customizing Personality and Behaviors

    • temporarily deactivate the real-time gesturing entirely.

    Related Topics

    Contents

    Table of Contents