Personal tools
You are here: Home SIGs and LIGs Patent SIG Granted Patents

Granted Patents

Granted patents in the area of emotion-related technology

Patent [···]
Patent [···]
A simulator which may be programmed to interactively mimic the effects of an addictive habit as well as the characteristics of a particular personality or trait. A case houses a speaker, microphone, visual display and associated electronic components, including a microprocessor. Numerous messages as stored in a nonvolatile random access memory which are issued to a user of the device to prompt the user to engage in activities that would simulate participation in the addictive behavior being simulated. In one version, the simulator is used in a teaching environment and includes controls that permit a teacher to select certain simulation parameters. In the case of a simulation of the habit of smoking, such parameters would include, for example, the number of packages of cigarettes available for consumption, the number of packages to be consumed each day and the rate at which the user's craving for nicotine will escalate. The device will calculate various physiological parameters and generate information relevant to the elapsed time since the simulation began. The simulator includes a signaling device such as a beeper or vibrator to notify the user that a cigarette must be consumed. The microprocessor is programmed to recognize sounds such as coughing and inhaling received through the microphone to ensure that the user is performing these activities in response to prompts by the device. An orifice is formed within the surface of the case into which a straw or tube is inserted through which the user must occasionally exhale, inhale or puff. A slot is formed within the case to receive currency in order to simulate the purchase of the addictive product.
PROBLEM TO BE SOLVED: To provide a sentence feeling recognition device allowing the recognition of a feeling matching the personality of a user, and allowing feeling recognition reflecting the feeling of the user with respect to an "object" when obtaining a feeling value expressed by a sentence. SOLUTION: A feeling vector generation part 103 calculates a feeling value of a sentence by referring to a basic feeling dictionary 106 storing a word and its feeling value in association with each other and a learning feeling dictionary 107 storing the word to be learned and its feeling value in association with each other. A learning control part 104 calculates a difference of the feeling value between the feeling value found by the feeling vector creation part 103 and that designated by the user, specifies a paragraph to be learned from the sentence serving as a subject of finding the feeling value, and about a word included in the specified paragraph and stored in the learning feeling dictionary, adds the different value for updating the learning feeling dictionary 107.
Patent Method and apparatus for creating and customizing avatars using genetic paradigm
A method and apparatus for creating an avatar are described. A number of user-modifiable genes are maintained in a local processing system, each of which represents a different user-perceivable attribute of the avatar. An encoded representation of the genes is then transmitted to a remote processing system over a network. The genes represent non-geometry parameters, which are used in both the local processing system and the remote processing system to generate geometry for rendering the avatar. One or more of the genes may represent a biomechanical, behavioral, and/or biometric attribute of the avatar. One or more of the genes may be adjusted by a user to allow the user to customize the avatar, which may be performed using a graphical user interface designed for avatar customization. Adjusting one or more of the genes may generate a variation of a previously-generated avatar and may be used to control blending between two or more 3-D models or animation scripts. An avatar may be generated by using a pseudorandom number generator to generate one or more of the genes. One or more of the genes may be associated with a personality trait of the avatar, such as alertness, shiftiness, curiosity, or tendency to daydream.
A speech synthesizer comprising an emotion input section (202) for acquiring the utterance mode of a speech waveform to be subjected to speech synthesis, a prosody creating section (205) for creating a prosody of when a text subjected to a language processing is uttered in the acquired utterance mode, a characteristic sound tone selecting section (203) for selecting a characteristic sound tone observed when the text is uttered in the utterance mode according to the utterance mode, a characteristic sound tone temporal position deducting section (604) for judging from the phoneme sequence of the text, the characteristic sound tone, and the prosody whether or not each phoneme constituting the phoneme sequence is to be uttered in the characteristic sound tone and determining the phoneme at the position where the phoneme is uttered in the characteristic sound tone, and a fragment selecting section (606) and a fragment connecting section (209) for creating a sound waveform with which the text is uttered in the utterance mode and in the characteristic sound tone at the determined utterance position.
PROBLEM TO BE SOLVED: To add feeling to synthesized speech while the prosodic feature of a language is maintained. SOLUTION: In the speech synthesis device 200, a language processing part 201 generates a string of pronunciation marks from a text, and a prosodic data generating part 202 creates prosodic data, expressing parameters such as the time duration of phonemes, pitch and sound volume, on the basis of the string of pronunciation marks. A constraint information generating part 203 is fed with the prosodic data and with the string of pronunciation marks to generate constraint information which limits changes in the parameters, adding the constraint information to the prosodic data so as to maintain the prosodic feature of a speech sentence. An emotion filter 204 is fed with the prosodic data added with the constraint information, changes the parameters of the prosodic data within the constraint in response to given feeling state information, and a waveform generating part 205 synthesizes a speech waveform on the basis of the prosodic data with the changed parameters. COPYRIGHT: (C)2003,JPO
In a robot apparatus and a control method thereof, first, the state space of a motion model is expanded/contracted partially or entirely; second, a transition to a specified node of the motion model is described as transition to a virtual node and the node group assigned to the virtual node is changed sequentially; third, the number of emotions and/or desires used for generation of a motion is increased/decreased gradually; and last, the environment is evaluated and the sensitivity to each emotion and/or desire is updated based on the evaluation results. In a robot apparatus and a method for judging the character thereof, a pet robot comprises means for detecting the output from a second robot, means for judging the character of the second robot based on the detection results, and means for varying the character based on the judgment results.
PROBLEM TO BE SOLVED: To provide a method and equipment for synthesizing a voice in which various voices having different speech characteristics, including human feeling, a speech style, personality, or the like are flexibly easily synthesized. SOLUTION: The method and the equipment for synthesizing the voice comprise: obtaining a first parameter indicative of the rhythmical feature of the voice having standard speech characteristics on the basis of language information obtained by analyzing a text; obtaining a second parameter for correcting the first parameter to indicate the rhythmical feature corresponding to the specified speech characteristics on the basis of the language information and the specified speech characteristics; superimposing at least the first parameter on the second parameter for every rhythm control unit as the unit of the speech for controlling the rhythmical feature of the voice to generate a third parameter indicative of the rhythmical feature corresponding to the specified speech characteristics; and generating the synthesized voice corresponding to the specified speech characteristics on the basis of the third parameter. COPYRIGHT: (C)2004,JPO
PURPOSE: To add interest by outputting a montage while changing a face or body corresponding to the emotion of an operator, etc., by recognizing an input voice, changing parts patterns corresponding to the contents and changing the montage. CONSTITUTION: The voice inputted from a voice input part 3 is recognized by a voice recognition part 25, and the volume is detected by a volume detection part 26. When it is decided the volume is less than a set value, a parts pattern number group corresponding to the voice number ('0' showing a silent state in this case) of the recognized voice (such as 'bark', for example), is read out of a montage RAM 24a, each parts pattern corresponding to this is read out of a basic parts pattern ROM 22a. This parts pattern is synthesized to the montage by a synthetic RAM 29, and the synthesized montage is displayed and outputted to a display part 2. Thus, the output state of the montage can be changed corresponding to the contents or characteristics of the voice, and the face of a human being or an animal can be variously changed. COPYRIGHT: (C)1994,JPO&Japio
PROBLEM TO BE SOLVED: To provide information on attention of a subject in an object space including an arbitrary construction to be an action target of the subject. SOLUTION: In an evaluation device 70, characteristics of the subject from a recognition device 10 are obtained as data, a noticed space is evaluated in a space evaluation part 92 by taking into consideration of the structure of objects present in the space, and the state of attention (emotion) of the space is evaluated by a attention state evaluation part 94. In an attention part evaluation part 96, a portion which the subject has paid attention to is derived and outputted as subject portion data 98. Thus, it is possible to present the attention portion the subject has paid attention to. COPYRIGHT: (C)2005,JPO&NCIPI
PROBLEM TO BE SOLVED: To realize a robot device capable of learning the information, while maintaining the interaction with a user. SOLUTION: This robot device is provided with an outside information detecting unit 2 for detecting the outside information, a speech and image data base unit 3 for storing the outside information detected by the outside information detecting unit 2, a speech generating unit 40 for extracting the speech from the outside information stored in the speech and image data base unit 3 to output it outside, an action control unit 50 for controlling the operation for expressing the emotion in response to a result of a comparison between an image corresponding to the generated speech and an image of the outside information detected in response to the speech, a naming unit 60 for naming the outside information stored in the speech and image data base unit 3 for renew, and a driving unit 6 for driving the autonomous action. This robot device autonomously acts to learn the information, while maintaining the interaction with the user. COPYRIGHT: (C)2002,JPO
Patent Videophone Terminal
The present invention provides a videophone terminal by which the emotions and impressions of a user can be transmitted to and easily understood by a co-communicant. when a user performs a predetermined operation or when an audio processor identifies a predetermined keyword in the speech of the user, a picture of a virtual character is modified, or a completely different picture is displayed. To modify the picture of the virtual character, the size of each part of the face or the whole face is changed, or a pattern expressing an emotion is added thereto. And when a pending mode is released, the virtual character is recovered and displayed on the screen while a melody is played. At the end of a video conversation, in accordance with the user's selection, a picture indicating that the virtual character disappears from the screen is displayed before disconnecting the line.
Patent Method and apparatus for specification, control and modulation of social primitives in animated characters
Non Verbal Behaviors (NVBs) are parameterized and selected for using in enhancing communicative abilities of an animated character. A parameter selection device selects the NVBs from a database of NVBs based on selection criteria consistent with a response or action generated by an animated character program. The selected NVBs are modulated based on an emphasis to be placed on the communication, and other factors, such as a personality of the animated character. The modulated NVBs are placed in a data stream of NVBs and other actions and communications, including, for example, language, that is provided to an animation system for display. The processes of the present invention allow for NVBs to be utilized to more effectively engage users of the animated character. The processes of the present invention may be applied in any animated character system including products by Microsoft (Office Assistant, Agent, etc.).
A video processor (5) receives an image signal from a camera (3), and a skin detection signal from a skin detection unit (1). The video processor (5) processes the image data to produce an output video signal (9) for display on a display means (not shown). The video processor (5) is configured to automatically adapt the image signal, for example change the skin color, based on an input signal (7). The input (7) relates to a characteristic of the image being viewed. For example, the input signal (7) may relate to an emotional characteristic of the person being viewed. The emotion of the person being viewed can be detected from the tone of voice of that person (e.g. the average pitch of the voice), or by means of a separate infrared camera, which detects heat from the face of the person being viewed. Based on the input signal (7) representing a characteristic of the object being viewed, the video processor (5) is configured to adapt the image signal accordingly. In one example, the skin color of the person is changed according to the emotion of the person. For example, the skin color of the person could be changed to red when an angry tone is detected, or grey when a calm tone is detected.
PROBLEM TO BE SOLVED: To provide a communication voice data management system in a call center, enabling improved work efficiency by managing communication voice data in a database, based on a variety of management data generated from the communication voice data, and to provide an operator terminal appropriate for the system. SOLUTION: An operator terminal 3 generates a communication voice data 11 having a recorded communication voice, preserves the communication voice data 11 into a communication recording server 6. Voice recognition of the communication voice data 11 is performed, thereby separates the communication voice data 11 into a customer communication voice data 13 and an operator communication voice data 14. Then, clause data 15, 16 each having a predetermined format are generated from the customer communication voice data 13 and the operator communication voice data 14. Through the classification of the entire clause data 15 corresponding to the customer communication voice data 13 into one of a plurality of emotion states, and then, and after the above classified data being made related to the customer concerned, a customer management data 18 is generated. Thereafter, folders 32-1 to 32-5, being separated for each customer emotion state, in which the corresponding customer management data 18 is collected, are generated in a voice management database 7. COPYRIGHT: (C)2005,JPO&NCIPI
PROBLEM TO BE SOLVED: To increase the precision of emotion detection processing and to reduce the possibility of an error occurrence. SOLUTION: According to a specific input voice, at least some of features of at least 1st and 2nd feature classes are evaluated, derived, and/or extracted and the current emotional state of a speaker and/or parameters of the emotional state are derived from the features. The 1st and 2nd feature classes are allocated to individual dimensions of an underlying emotional manifold or emotional space, such as an awakening degree and comfortableness, and/or with subspaces thereof. COPYRIGHT: (C)2003,JPO
Sensory output devices such as wearable items, three dimensional objects such as pebbles, ornaments, toy characters and the like, include controls responsive to the content of SMS messages or to recognition of spoken words or phrases in a telephone conversation to provide a response such as a thermal change, vibrational or other tactile response, colour change or olfactory output. The output may be intensified in dependence upon the number of times at which a particular word, phrase or emoticon is identified, the control means learning from identity information to associate a current call with an historic personality trait to maintain or adapt the response provided by the sensory output device.
Patent Sign recognition apparatus and method and sign translation system using same.
In continuous sign language recognition, reference sign language patterns of good quality are generated. Continuous sign language patterns and reference sign language patterns are efficiently compressed. The compressed continuous sign language patterns are sequentially and directly matched with the compressed reference sign language patterns to recognize the sign language at high speed and with high precision. A reference sign pattern generating unit (3) generates a reference sign language pattern by normalizing sample patterns while taking into consideration of their nonlinear compression/expansion and by calculating an average of the sample patterns. A continuous sign language recognition unit (5) recognizes a continuous sign language at high speed by sequentially matching the continuous sign language of time sequential patterns with reference sign language patterns while allowing nonlinear expansion and compression in the time domain. A sign language translation system is provided in which a sign language as well as the facial expression or emotion of a sign language user is recognized and converted into a spoken language with emotion, and transferred in the form of texts, voices, sign languages of another type to a particular partner among a plurality of unknown system users.
A robotic cellular phone is disclosed. The robotic cellular phone includes a self-environmental recognition function part, a mobility function part, and an emotion function part. The self-environmental recognition function part is for recognizing an external environment by a photo sensor for a sensing luminosity of a surrounding device; a temperature sensor for sensing temperatures of the surrounding device; and a touch sensor for sensing a touch between a human and the surrounding device. The mobility function part is for moving the antenna to an optimum location to increase a sending or receiving signal sensitivity and for opening and closing a folder by using a plurality of wheels formed on a part of rechargeable battery. The emotion function part is for identifying a sender by comparing data between sender telephone information included a receiving signal and a telephone data stored in an internal memory; and for making a different vibration according to an emotional pattern by modeling a frequency of a vibrator based on a major or minor key harmony or for spraying perfume through a micro-nozzle.
Document Actions
Powered by Plone

Portal usage statistics