US20130185076A1 - Motion analyzer, voice acquisition apparatus, motion analysis system, and motion analysis method - Google Patents
Motion analyzer, voice acquisition apparatus, motion analysis system, and motion analysis method Download PDFInfo
- Publication number
- US20130185076A1 US20130185076A1 US13/565,365 US201213565365A US2013185076A1 US 20130185076 A1 US20130185076 A1 US 20130185076A1 US 201213565365 A US201213565365 A US 201213565365A US 2013185076 A1 US2013185076 A1 US 2013185076A1
- Authority
- US
- United States
- Prior art keywords
- voice
- motion
- speaking
- subject
- unit
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Granted
Links
- 230000033001 locomotion Effects 0.000 title claims abstract description 82
- 238000004458 analytical method Methods 0.000 title claims description 52
- 238000001514 detection method Methods 0.000 claims abstract description 38
- 230000008054 signal transmission Effects 0.000 claims 3
- 230000001133 acceleration Effects 0.000 description 36
- 230000005540 biological transmission Effects 0.000 description 12
- 238000012545 processing Methods 0.000 description 11
- 238000007405 data analysis Methods 0.000 description 10
- 238000013500 data storage Methods 0.000 description 9
- 238000000034 method Methods 0.000 description 9
- 238000004891 communication Methods 0.000 description 8
- 230000006870 function Effects 0.000 description 8
- 238000011156 evaluation Methods 0.000 description 7
- 230000008859 change Effects 0.000 description 5
- 230000000694 effects Effects 0.000 description 5
- 230000003993 interaction Effects 0.000 description 5
- 239000011347 resin Substances 0.000 description 4
- 229920005989 resin Polymers 0.000 description 4
- 238000001914 filtration Methods 0.000 description 3
- 230000007423 decrease Effects 0.000 description 2
- 230000007613 environmental effect Effects 0.000 description 2
- 239000000463 material Substances 0.000 description 2
- 239000002184 metal Substances 0.000 description 2
- 238000012986 modification Methods 0.000 description 2
- 230000004048 modification Effects 0.000 description 2
- 230000008569 process Effects 0.000 description 2
- 229920000742 Cotton Polymers 0.000 description 1
- YCKRFDGAMUMZLT-UHFFFAOYSA-N Fluorine atom Chemical compound [F] YCKRFDGAMUMZLT-UHFFFAOYSA-N 0.000 description 1
- XUIMIQQOPSSXEZ-UHFFFAOYSA-N Silicon Chemical compound [Si] XUIMIQQOPSSXEZ-UHFFFAOYSA-N 0.000 description 1
- 210000001015 abdomen Anatomy 0.000 description 1
- 239000003990 capacitor Substances 0.000 description 1
- 238000006243 chemical reaction Methods 0.000 description 1
- 239000011248 coating agent Substances 0.000 description 1
- 238000000576 coating method Methods 0.000 description 1
- 238000011109 contamination Methods 0.000 description 1
- 239000000284 extract Substances 0.000 description 1
- 239000000835 fiber Substances 0.000 description 1
- 229910052731 fluorine Inorganic materials 0.000 description 1
- 239000011737 fluorine Substances 0.000 description 1
- 230000010365 information processing Effects 0.000 description 1
- 239000010985 leather Substances 0.000 description 1
- 239000002649 leather substitute Substances 0.000 description 1
- 230000000877 morphologic effect Effects 0.000 description 1
- 229910052710 silicon Inorganic materials 0.000 description 1
- 239000010703 silicon Substances 0.000 description 1
- 239000012209 synthetic fiber Substances 0.000 description 1
- 229920002994 synthetic fiber Polymers 0.000 description 1
Images
Classifications
-
- G—PHYSICS
- G01—MEASURING; TESTING
- G01S—RADIO DIRECTION-FINDING; RADIO NAVIGATION; DETERMINING DISTANCE OR VELOCITY BY USE OF RADIO WAVES; LOCATING OR PRESENCE-DETECTING BY USE OF THE REFLECTION OR RERADIATION OF RADIO WAVES; ANALOGOUS ARRANGEMENTS USING OTHER WAVES
- G01S11/00—Systems for determining distance or velocity not using reflection or reradiation
- G01S11/14—Systems for determining distance or velocity not using reflection or reradiation using ultrasonic, sonic, or infrasonic waves
Definitions
- the present invention relates to a motion analyzer, a voice acquisition apparatus, a motion analysis system, and a motion analysis method.
- a motion analyzer including: a motion detection unit that detects motion of a part of a body of a subject; a speaking detection unit that detects speaking of the subject; and a determination unit that determines that the subject has performed predetermined motion when motion of a part of the body is detected by the motion detection unit and speaking of the subject is detected by the speaking detection unit.
- FIG. 1 is a view showing an example of the configuration of a motion analysis system
- FIG. 2 is a view showing an example of the configuration of a terminal apparatus
- FIG. 3 is a view showing a positional relationship between mouths (speaking portions) of a wearer and another person and a microphone;
- FIG. 4 is a view showing a relationship of sound pressure (input volume) and a distance between a microphone and a sound source;
- FIG. 5 is a view showing a method of identifying a voice of a wearer and a voice of another person
- FIG. 6 is a flow chart showing the basic operation of a terminal apparatus
- FIG. 7 is a view showing an output from an acceleration sensor
- FIG. 8 is a view showing an output of an acceleration sensor when a wearer is in a conversation and the speaking information of the wearer.
- FIG. 9 is a view showing a situation where plural wearers who wear the terminal apparatus according to the exemplary embodiment have a conversation with each other;
- FIG. 10 is a view showing an example of the speaking information of each terminal apparatus in the conversation situation shown in FIG. 9 ;
- FIG. 11 is a view showing an example of the functional configuration of a host apparatus in an application example.
- FIG. 1 is a view showing an example of the configuration of a motion analysis system of the present exemplary embodiment.
- the system according to the present exemplary embodiment is configured to include a terminal apparatus 10 and a host apparatus 20 .
- the terminal apparatus 10 and the host apparatus 20 are connected to each other through a radio communication line.
- lines based on the known methods such as Wireless Fidelity (Wi-Fi) (registered trademark), Bluetooth (registered trademark), ZigBee (registered trademark), and Ultra Wideband (UWB), may be used.
- Wi-Fi Wireless Fidelity
- Bluetooth registered trademark
- ZigBee registered trademark
- UWB Ultra Wideband
- a wearer the user who wears the terminal apparatus 10 is called a wearer.
- the terminal apparatus 10 as an example of a voice acquisition apparatus includes at least two microphones (a first microphone 111 and a second microphone 112 ) and amplifiers (a first amplifier 131 and a second amplifier 132 ).
- the terminal apparatus 10 includes a voice analysis unit 15 that analyzes the recorded voice and a data transmission unit 16 that transmits an analysis result to the host apparatus 20 , and further includes a power supply unit 17 .
- the terminal apparatus 10 includes a timepiece unit 18 and an acceleration sensor 19 .
- the timepiece unit 18 may be configured using a clock IC (Integrated Circuit), such as a real-time clock, and counts the passage of time and generates calendar information including current year, month, and date and time information (hereinafter, “calendar information and time information” is called “date information”).
- the timepiece unit 18 outputs the generated date information to the voice analysis unit 15 and the data transmission unit 16 .
- the first and second microphones 111 and 112 are disposed at different positions, at which distances of sound wave propagation paths from the mouth (speaking portion) of the wearer (hereinafter, simply referred to as “distances”) are different.
- the first microphone 111 as an example of a first voice acquisition unit is disposed at the position (for example, about 35 cm) far from the mouth (speaking portion) of the wearer.
- the second microphone 112 as an example of a second voice acquisition unit is disposed at the position (for example, about 10 cm) near the mouth (speaking portion) of the wearer.
- various types of known microphones such as a dynamic type microphone and a capacitor type microphone, may be used as the first and second microphones 111 and 112 in the present exemplary embodiment.
- the first and second amplifiers 131 and 132 amplify electric signals (voice signals) that the first and second microphones 111 and 112 output according to the acquired voice.
- Known operational amplifiers or the like may be used as the first and second amplifiers 131 and 132 in the present exemplary embodiment.
- the voice analysis unit 15 analyzes the voice signals output from the first and second amplifiers 131 and 132 . In addition, the voice analysis unit 15 identifies whether the voices acquired by the first and second microphones 111 and 112 are a voice from the wearer, who wears the terminal apparatus 10 , or voices from others. That is, the voice analysis unit 15 functions as an identification unit that identifies a speaker of the voice on the basis of the voices acquired by the first and second microphones 111 and 112 . Details of specific processing for identification of a speaker will be described later.
- the data transmission unit 16 transmits the acquired data including the analysis result of the voice analysis unit 15 , an ID of the terminal apparatus 10 , the output from the acceleration sensor 19 , and the date information generated by the timepiece unit 18 to the host apparatus 20 through the radio communication line.
- information transmitted to the host apparatus 20 for example, information regarding sound pressure of the acquired voice and the like may be included in addition to the analysis result according to processing performed in the host apparatus 20 .
- the analysis result of the voice analysis unit 15 may be stored in the terminal apparatus 10 , and data stored for a certain period of time may be collectively transmitted.
- the data may be transmitted through a cable line.
- the power supply unit 17 supplies electric power to the first and second microphones 111 and 112 , the first and second amplifiers 131 and 132 , the voice analysis unit 15 , the data transmission unit 16 , the timepiece unit 18 , and the acceleration sensor 19 .
- the power supply it is possible to use known power supplies, such as a dry battery or a rechargeable battery, for example.
- the power supply unit 17 includes known circuits, such as a voltage conversion circuit and a charging control circuit, when necessary.
- the host apparatus 20 includes a data receiving unit 21 that receives the data transmitted from the terminal apparatus 10 , a data storage unit 22 that stores the received data, a data analysis unit 23 that analyzes the stored data, and an output unit 24 that outputs the analysis result.
- the host apparatus 20 is realized by an information processing apparatus, such as a personal computer, for example.
- the plural terminal apparatuses 10 are used as described above, and the host apparatus 20 receives the data from each of the plural terminal apparatuses 10 .
- the data receiving unit 21 corresponds to the radio communication line described above, and receives the data from each terminal apparatus 10 and transmits the received data to the data storage unit 22 .
- the data storage unit 22 is realized by a memory of a magnetic disk device of a personal computer, for example, and stores the received data acquired from the data receiving unit 21 for each speaker.
- identification of a speaker is performed on the basis of a terminal ID transmitted from the terminal apparatus 10 and a combination of a speaker name and a terminal ID registered in the host apparatus 20 in advance.
- a wearer's name may be transmitted from the terminal apparatus 10 .
- the data analysis unit 23 is realized by a CPU program-controlled by a personal computer, for example, and analyzes the data stored in the data storage unit 22 .
- a specific analysis content and analysis method various kinds of content and methods may be adopted depending on the purpose or aspect of use of the system according to the present exemplary embodiment. For example, the frequency of conversation between wearers of the terminal apparatuses 10 or the tendencies of a conversation partner of each wearer are analyzed, or the relationship between speakers in a conversation is estimated from the information regarding the length or sound pressure of the voice in the conversation.
- the output unit 24 outputs the analysis result of the data analysis unit 23 or performs output based on the analysis result.
- various kinds of units including display of a display device, printout using a printer, and voice output may be adopted according to the purpose or aspect of use of the system, the content or format of an analysis result, and the like.
- FIG. 2 is a view showing an example of the configuration of the terminal apparatus 10 .
- the terminal apparatus 10 is used in a state worn by each user.
- the terminal apparatus 10 in the present exemplary embodiment is configured to include an apparatus body 30 and a strap 40 connected to the apparatus body 30 so that the user may wear the terminal apparatus 10 , as shown in FIG. 2 .
- the user wears the strap 40 on the neck so that the apparatus body 30 is hung from the neck.
- the apparatus body 30 is configured such that at least circuits for realizing the first and second amplifiers 131 and 132 , the voice analysis unit 15 , the data transmission unit 16 , the power supply unit 17 , and the timepiece unit 18 and a power supply (battery) of the power supply unit 17 are housed in a rectangular parallelepiped thin case 31 formed of metal, resin, or the like.
- a pocket through which an ID card, on which ID information such as the name or affiliation of the wearer is displayed, is inserted may be provided in the case 31 .
- ID information or the like may be printed on the surface of the case 31 , or a seal on which the ID information or the like is described may be attached to the surface of the case 31 .
- the first and second microphones 111 and 112 are provided in the strap 40 (hereinafter, referred to as microphones 111 and 112 when the first and second microphones 111 and 112 are not distinguished from each other).
- the acceleration sensor 19 functioning as a motion detection unit is provided in the strap 40 .
- the microphones 111 and 112 are connected to the first and second amplifiers 131 and 132 by cables (electric wires or the like) passing through the inside of the strap 40 .
- the acceleration sensor 19 is connected to the data transmission unit 16 by cables (electric wires or the like) passing through the inside of the strap 40 .
- the strap 40 As materials of the strap 40 , it is possible to use various known materials, such as leather, synthetic leather, cotton, other natural fibers, synthetic fiber using resin, and metal. In addition, coating processing using silicon resin, fluorine resin, or the like may be performed.
- the strap 40 has a cylindrical structure, and the microphones 111 and 112 and the acceleration sensor 19 are housed inside the strap 40 .
- the microphones 111 and 112 are provided inside the strap 40 , it is possible to prevent damage or contamination of the microphones 111 and 112 , and it is suppressed that a speaker in a conversation is aware of the existence of the microphones 111 and 112 .
- the first microphone 111 disposed at the position far from the mouth (speaking portion) of the wearer may be provided in the apparatus body 30 so as to be housed in the case 31 .
- the first microphone 111 is provided in the strap 40 as an example.
- the microphones 111 and 112 are provided in portions of the strap 40 where the abdomen of the wearer is located.
- the acceleration sensor 19 is provided in a portion of the strap 40 on the back of the wearer side. More specifically, the acceleration sensor 19 is provided in a portion, which is located behind the neck of the wearer, of the strap 40 .
- the first microphone 111 is provided at the end (for example, a position within 10 cm from a connection portion) of the strap 40 connected to the apparatus body 30 . Accordingly, the first microphone 111 is disposed at the position distant from the mouth (speaking portion) of the wearer by about 30 cm to 40 cm in a state where the wearer wears the strap 40 on the neck so that the apparatus body 30 is hung from the neck. In addition, also when the first microphone 111 is provided in the apparatus body 30 , the distance from the mouth (speaking portion) of the wearer to the first microphone 111 is almost the same as above.
- the second microphone 112 is provided at the position (for example, a position of about 20 cm to 30 cm from a connection portion) distant from the end of the strap 40 connected to the apparatus body 30 . Accordingly, in a state where the wearer wears the strap 40 on the neck so that the apparatus body 30 is hung from the neck, the second microphone 112 is located on the neck (for example, a position equivalent to the collarbone) of the wearer and is disposed at the position distant from the mouth (speaking portion) of the wearer by about 10 cm to 20 cm.
- the strap 40 formed in an annular shape is hung on the neck of the wearer, and this strap 40 is pulled downward by the apparatus body 30 .
- a first portion 41 which starts from a portion connected to the apparatus body 30 and extends toward the back of the neck of the wearer through the right shoulder side of the wearer
- a second portion 42 which starts from a portion connected to the apparatus body 30 and extends toward the back of the neck of the wearer through the left shoulder side of the wearer, are provided in the strap 40 .
- the first and second microphones 111 and 112 are attached to the first portion 41 .
- the first microphone 111 may be provided in one of the first and second portions 41 and 42
- the second microphone 112 may be provided in the other portion of the first and second portions 41 and 42 .
- the microphones 111 and 112 and the acceleration sensor 19 are not limited to being provided in the strap 40 , and may be worn on the wearer by various methods.
- each of the first and second microphones 111 and 112 and the acceleration sensor 19 may be separately fixed to clothes using a pin or the like.
- the configuration of the apparatus body 30 is not limited to the configuration shown in FIG. 2 in which the apparatus body 30 is connected to the strap 40 so as to be hung from the neck of the wearer, and the apparatus body 30 may be formed as a device which may be easily carried.
- the apparatus body 30 may be mounted on clothes or the body using a clip or a belt instead of the strap in the present exemplary embodiment, or the apparatus body 30 may simply be carried in a pocket or the like.
- mobile phones or known portable electronic information terminals may be configured to realize a function of receiving a voice signal from the microphones 111 and 112 and amplifying and analyzing the voice signal.
- the first microphone 111 is provided in the apparatus body 30 , it is necessary to maintain the positional relationship between the first and second microphones 111 and 112 as described above. Accordingly, the position of the apparatus body 30 at the time of carrying is specified.
- the microphones 111 and 112 and the apparatus body 30 may be wirelessly connected to each other instead of being connected using a cable.
- the acceleration sensor 19 and the apparatus body 30 (or the data transmission unit 16 ) may be connected to each other through radio communication instead of being connected using a cable.
- the first and second amplifiers 131 and 132 , the voice analysis unit 15 , the data transmission unit 16 , the power supply unit 17 , and the timepiece unit 18 are housed in the single case 31 in the above example of the configuration, they may be grouped into plural parts.
- the power supply unit 17 may be connected to an external power supply without being housed in the case 31 .
- speaker identification is performed on the basis of non-linguistic information, such as sound pressure (volume input to the microphones 111 and 112 ), instead of linguistic information acquired using morphological analysis or dictionary information, among information items of the recorded voice. That is, a speaker of the voice is identified from the speaking situation specified by the non-linguistic information instead of the content of speaking specified by the linguistic information.
- non-linguistic information such as sound pressure (volume input to the microphones 111 and 112 )
- the first microphone 111 of the terminal apparatus 10 is disposed at the position far from the mouth (speaking portion) of the wearer and the second microphone 112 is disposed at the position near the mouth (speaking portion) of the wearer. That is, assuming that the mouth (speaking portion) of the wearer is a sound source, the distance between the first microphone 111 and the sound source is greatly different from the distance between the second microphone 112 and the sound source.
- the distance between the first microphone 111 and the sound source is about 1.5 to 4 times the distance between the second microphone 112 and the sound source.
- sound pressure of the recorded voice in each of the microphones 111 and 112 decreases (distance-decreases) as the distance between each of the microphones 111 and 112 and the sound source increases. Therefore, for the voice of the wearer, the sound pressure of the recorded voice in the first microphone 111 is greatly different from the sound pressure of the recorded voice in the second microphone 112 .
- the distance between the first microphone 111 and the sound source and the distance between the second microphone 112 and the sound source do not change greatly since the other person is away from the wearer.
- the distance between the first microphone 111 and the sound source is not several times the distance between the second microphone 112 and the sound source unlike the case when the mouth (speaking portion) of the wearer is a sound source. Therefore, for the voice of another person, the sound pressure of the recorded voice in the first microphone 111 is not largely different from the sound pressure of the recorded voice in the second microphone 112 unlike the case of the voice of the wearer.
- FIG. 3 is a view showing the positional relationship between mouths (speaking portions) of a wearer and another person and the microphones 111 and 112 .
- the distance between a sound source a, which is a mouth (speaking portion) of the wearer, and the first microphone 111 is La 1
- the distance between the sound source a and the second microphone 112 is La 2
- the distance between a sound source b, which is a mouth (speaking portion) of another person, and the first microphone 111 is Lb 1
- the distance between the sound source b and the second microphone 112 is Lb 2 .
- FIG. 4 is a view showing the relationship of the distance between each of the microphones 111 and 112 and the sound source and the sound pressure (input volume)
- the sound pressure Ga 2 is about 4 times the sound pressure Ga 1 .
- the distances Lb 1 and Lb 2 are approximately equal, sound pressure Gb 1 in the case of the distance Lb 1 and sound pressure Gb 2 in the case of the distance Lb 2 are almost equal.
- a voice of the wearer and voices of others in the recorded voices are identified using the difference in sound pressure ratio.
- the distances Lb 1 and Lb 2 are 60 cm in the example shown in FIG. 4 , it is meaningful that the sound pressure Gb 1 and the sound pressure Gb 2 are almost equal, and the distance Lb 1 and Lb 2 are not limited to the values shown in FIG. 4 .
- FIG. 5 is a view showing a method of identifying a voice of a wearer and a voice of another person.
- the sound pressure Ga 2 of the second microphone 112 is several times (for example, about 4 times) the sound pressure Ga 1 of the first microphone 111 .
- the sound pressure Gb 2 of the second microphone 112 is almost equal to (about 1 times) the sound pressure Gb 1 of the first microphone 111 .
- a threshold value of the ratio between the sound pressure of the second microphone 112 and the sound pressure of the first microphone 111 is set.
- a voice with a larger sound pressure ratio than the threshold value is determined to be the voice of the wearer, and a voice with a smaller sound pressure ratio than the threshold value is determined to be the voice of another person.
- the threshold value is 2. Since the sound pressure ratio Ga 2 /Ga 1 exceeds the threshold value 2, the voice is determined to be the voice of the wearer. In addition, since the sound pressure ratio Gb 2 /Gb 1 is smaller than the threshold value 2, the voice is determined to be the voice of another person.
- the voice not only the voice but also so-called noise, such as environmental sounds, is included in the voice recorded by the microphones 111 and 112 .
- the relationship of the distance between the sound source of noise and each of the microphones 111 and 112 is similar to that in the case of voices of others. That is, according to the example shown in FIGS. 4 and 5 , assuming that the distance between a sound source c of noise and the first microphone 111 is Lc 1 and the distance between the sound source c of noise and the second microphone 112 is Lc 2 , the distances Lc 1 and Lc 2 are approximately equal.
- the sound pressure ratio Gc 2 /Gc 1 in recorded voices of the microphones 111 and 112 is smaller than the threshold value 2.
- such noise is separated and removed from the voice by performing filtering based on the known technique using a band pass filter, a gain filter, or the like.
- FIG. 6 is a flow chart showing the basic operation of the terminal apparatus 10 .
- step 601 when the microphones 111 and 112 of the terminal apparatus 10 acquire voices, electric signals (voice signals) corresponding to the acquired voices are transmitted from the microphones 111 and 112 to the first and second amplifiers 131 and 132 , respectively (step 601 ).
- the first and second amplifiers 131 and 132 amplify the signals and transmit the amplified signals to the voice analysis unit 15 (step 602 ).
- the voice analysis unit 15 receives the signals amplified by the first and second amplifiers 131 and 132 and also performs filtering processing on the received signals to remove noise components, such as environmental sounds, from the signals (step 603 ). Then, the voice analysis unit 15 calculates the average sound pressure in the recorded voice of each of the microphones 111 and 112 for the signals, from which noise components are removed, every fixed time unit (for example, few tenths of a second to few hundredths of a second) (step 604 ).
- the voice analysis unit 15 determines that there is a voice (speaking has been done). Then, the voice analysis unit 15 calculates the ratio (sound pressure ratio) between the average sound pressure in the first microphone 111 and the average sound pressure in the second microphone 112 (step 606 ). Then, when the sound pressure ratio calculated in step 606 is larger than the threshold value (Yes in step 607 ), the voice analysis unit 15 that functions as a voice detection unit determines that the voice is from the wearer (step 608 ) , and detects the speaking of the wearer. In addition, when the sound pressure ratio calculated in step 606 is smaller than the threshold value (No in step 607 ), the voice analysis unit 15 determines that the voice is from another person (step 609 ).
- the voice analysis unit 15 determines that there is no voice (speaking has not been performed) (step 610 ).
- the data transmission unit 16 transmits the information obtained by the processing in steps 604 to 610 , as an analysis result, to the host apparatus 20 (step 611 ). Specifically, the data transmission unit 16 transmits to the host apparatus 20 information regarding whether or not there has been speaking of the wearer, information of date and time of speaking of the wearer when the wearer has spoken, information regarding whether or not there has been speaking of a person other than the wearer (others), information of date and time of speaking of others when a person other than the wearer (others) has spoken, and the like.
- the date and time of speaking of the wearer and the date and time of speaking of others are checked by the voice analysis unit 15 on the basis of the date information from the timepiece unit 18 , the outputs from the microphones 111 and 112 , and the determination result of steps 608 to 610 .
- the voice analysis unit 15 may be transmitted to the host apparatus 20 together with the analysis result.
- the data transmission unit 16 also transmits the output of the acceleration sensor 19 to the host apparatus 20 when necessary, so that the detection result of the acceleration sensor 19 is acquired by the host apparatus 20 .
- the data transmission unit 16 in transmitting the output of the acceleration sensor 19 to the host apparatus 20 , the data transmission unit 16 also transmits the date information generated by the timepiece unit 18 together.
- the data storage unit 22 of the host apparatus 20 stores the output from the acceleration sensor 19 and this date information so as to match each other.
- determination regarding whether the voice is a voice from the wearer or a voice from another person is performed by comparing the sound pressure of the first microphone 111 with the sound pressure of the second microphone 112 .
- any kind of speaker identification may be performed if it is performed on the basis of the non-linguistic information extracted from the voice signals themselves acquired by the microphones 111 and 112 , without being limited to the comparison of sound pressure.
- the voice acquisition time (output time of a voice signal) in the first microphone 111 with the voice acquisition time in the second microphone 112 .
- a difference (time difference) in voice acquisition time occurs to some extent for the voice of the wearer.
- speaking information information regarding speaking obtained as described above by the plural terminal apparatuses 10 are collected in the host apparatus 20 .
- the host apparatus 20 uses the information acquired from the terminal apparatus 10 to perform various kinds of processing according to the purpose or aspect of use of the system. For example, the host apparatus 20 (data analysis unit 23 of the host apparatus 20 ) analyzes the information acquired from the terminal apparatus 10 and determines whether or not the wearer has made a “nodding” motion which is to shake the head back and forth.
- the acceleration sensor 19 is located behind the wearer's neck, and the “nodding” motion may be detected using the output from the acceleration sensor 19 . More specifically, in the present exemplary embodiment, not only the identification of a speaker but also detection regarding whether or not a wearer has made an affirmative answer may be performed.
- the present exemplary embodiment in order to improve the detection accuracy of the “nodding” motion, not only the output from the acceleration sensor 19 but also speaking information of the wearer is used. More specifically, in the present exemplary embodiment, when there is wearer's speaking and there is an output from the acceleration sensor 19 which satisfies the predetermined conditions, it is determined that there is a “nodding” motion.
- FIG. 8 is a view showing an output of the acceleration sensor 19 when a wearer is in a conversation and the speaking information of the wearer.
- “nodding” of the wearer is actually performed in places indicated by the reference numerals 8 A and 8 B.
- the data analysis unit analyzes the output of the acceleration sensor 19 transmitted from the terminal apparatus 10 first and determines whether or not a signal waveform (output) satisfying the predetermined conditions is included in the output of the acceleration sensor 19 .
- time information (date information) when the signal waveform is output is checked.
- the data analysis unit 23 checks the date and time of wearer's speaking when the wearer has spoken on the basis of the information transmitted from the terminal apparatus 10 . In addition, the data analysis unit 23 determines that a “nodding” motion has been made by the wearer when the time interval between the date and time when the signal waveform satisfying the predetermined conditions is output and the date and time specified by the above date and time of wearer's speaking is smaller than the predetermined time interval. On the other hand, when the time interval between the date and time when the signal waveform satisfying the predetermined conditions is output and the date and time specified by the above date and time of wearer's speaking is larger than the predetermined time interval, the data analysis unit 23 determines that no “nodding” motion has been made by the wearer.
- determination regarding whether or not a “nodding” motion has been made may be made by performing video photographing of persons in a conversation and by observing the video image after photographing, for example. In this case, however, it is necessary to construct a large-scale system. In the present exemplary embodiment, the determination regarding whether or not “nodding” has been made is realized with a simple configuration.
- the determination regarding whether or not a “nodding” motion has been made is performed by the host apparatus 20 has been described above, the determination regarding whether or not a “nodding” motion has been made may also be performed by the terminal apparatus 10 .
- the acceleration sensor 19 is disposed so as to be located behind the neck in the above explanation, the acceleration sensor 19 may also be provided in other locations, such as the back of the head, as long as the location is a place linked to the swing of the wearer's neck (movement of the wearer's head).
- the acceleration sensor 19 may also be provided in a hat, an earphone, a headphone, or the like.
- a swing (swing in the left and right direction) motion performed when giving a negative answer to the question may also be detected by the same process as described above.
- a different signal from the signal described above is output from the acceleration sensor 19 .
- the acceleration sensor 19 when there is wearer's speaking for a shorter time than the predetermined time and a signal (output) satisfying the predetermined conditions has been output from the acceleration sensor 19 , it may be determined that there is a “nodding” motion. More specifically, when a time indicated by the reference numeral T in FIG. 8 (period for which wearer's speaking continues) is shorter than the predetermined time and a signal satisfying the predetermined conditions has been output from the acceleration sensor 19 , it may be determined that there is a “nodding” motion. This is because speaking made with a “nodding” motion ends in a short time in many cases and accordingly, a possibility of speaking made at the time of motion other than the “nodding” motion is high when the speaking time is long.
- FIG. 9 is a view showing a situation where plural wearers who wear the terminal apparatuses 10 according to the present exemplary embodiment have a conversation with each other.
- FIG. 10 is a view showing an example of the speaking information of terminal apparatuses 10 A and 10 B in the conversation situation shown in FIG. 9 .
- a case is considered in which two wearers A and B each of whom wears the terminal apparatus 10 have a conversation with each other.
- a voice recognized as wearer's speaking in the terminal apparatus 10 A of the wearer A is recognized as another person's speaking in the terminal apparatus 10 B of the wearer B.
- a voice recognized as wearer's speaking in the terminal apparatus 10 B is recognized as another person's speaking in the terminal apparatus 10 A.
- the speaking information is separately transmitted from the terminal apparatuses 10 A and 10 B to the host apparatus 20 .
- identification results of speakers (wearer and another person) in the speaking information acquired from the terminal apparatus 10 A and the speaking information acquired from the terminal apparatus 10 B are opposite as shown in FIG. 10 , but the information showing the speaking situation, such as the length of speaking time or the speaker change timing, is similar.
- the host apparatus 20 in this application example determines that such information items indicate the same speaking situation by comparing the information acquired from the terminal apparatus 10 A with the information acquired from the terminal apparatus 10 B, and recognizes that the wearers A and B have a conversation.
- at least the time information regarding the speaking such as the length of speaking time in each speaking of each speaker, start time and end time of each speaking, and speaker change time (timing), is used as the information showing the speaking situation.
- the time information regarding the speaking may be used, or other information may be additionally used.
- FIG. 11 is a view showing an example of the functional configuration of the host apparatus 20 in this application example.
- the host apparatus 20 includes a conversation information detecting section 201 that detects the speaking information (hereinafter, referred to as conversation information) from the terminal apparatus 10 of the wearer in conversation, among the speaking information items acquired from the terminal apparatus 10 , and a conversation information analysis section 202 that analyzes the detected conversation information.
- the conversation information detecting section 201 and the conversation information analysis section 202 are realized as a function of the data analysis unit 23 .
- the speaking information is transmitted to the host apparatus 20 .
- the speaking information from each terminal apparatus 10 which is received by the data receiving unit 21 is stored in the data storage unit 22 .
- the conversation information detecting section 201 of the data analysis unit 23 reads the speaking information of each terminal apparatus 10 stored in the data storage unit 22 and detects the conversation information which is the speaking information related to a specific conversation.
- the conversation information detecting section 201 compares the speaking information that is acquired from each terminal apparatus 10 and is stored in the data storage unit 22 , detects the speaking information with the above-described correspondence relationship from the speaking information acquired from the plural terminal apparatuses 10 , and identifies the speaking information as conversation information related to the same conversation.
- the speaking information from the plural terminal apparatuses 10 is transmitted to the host apparatus 20 when necessary. Accordingly, the conversation information detecting section 201 determines whether or not the conversation information related to the specific conversation is included by performing the above-described processing while sequentially separating the speaking information in a fixed period, for example.
- the conditions required when the conversation information detecting section 201 detects the conversation information related to a specific conversation from the speaking information of the plural terminal apparatuses 10 are not limited to the correspondence relationship shown in FIG. 10 described above. Any detection method may be used as long as it is possible to identify the conversation information related to a specific conversation from plural speaking information items.
- the number of persons participating in a conversation is not limited to two persons.
- the terminal apparatus 10 worn by each wearer recognizes the voice of the wearer of the host apparatus as the voice of the wearer and distinguishes the voice of the wearer from the voices of others (two or more persons).
- the information showing the speaking situation such as speaking time or speaker change timing, is similar among the acquired information items in the respective terminal apparatuses 10 .
- the conversation information detecting section 201 detects the speaking information acquired from the terminal apparatuses 10 of wearers participating in the same conversation and distinguishes the acquired speaking information from the speaking information acquired from the terminal apparatuses 10 of wearers not participating in the conversation.
- the conversation information analysis section 202 analyzes the conversation information detected by the conversation information detecting section 201 and extracts the features of the conversation.
- the features of the conversation are extracted on the basis of three kinds of criteria for evaluation, that is, the degree of interaction, the degree of listening, and the degree of conversation activity.
- the degree of interaction is assumed to indicate the balance of the speaking frequency of conversation participant.
- the degree of listening is assumed to indicate the degree of listening to others in each conversation participant.
- the degree of conversation activity is assumed to indicate the density of speaking in the entire conversation.
- the degree of interaction is specified by the number of times of speaker changes during the conversation and a variation in time taken until the speaker is changed (time for which one speaker speaks continuously). This may be obtained from the number of times of speaker changes and the time when the speaker change occurs in the conversation information in a fixed period. In addition, it is assumed that the value (level) of the degree of interaction increases as the number of times of speaker changes increases, that is, as a variation in continuous speaking time of each speaker decreases. This criterion for evaluation is common in all conversation information items (speaking information of each terminal apparatus 10 ) related to the same conversation.
- the degree of listening is specified by the ratio of speaking time of each conversation participant and speaking time of others in conversation information. For example, in the following expression, it is assumed that the value (level) of the degree of listening increases as the value of speaking time of others increases.
- This criterion for evaluation differs with the speaking information acquired from the terminal apparatus 10 of each conversation participant even in the conversation information related to the same conversation.
- the degree of conversation activity is an index showing the so-called excitement of the conversation, and is specified by the ratio of silence time (time for which none of conversation participants speak) to the total conversation time. It is assumed that the value (level) of the degree of conversation activity increases as the total silence time becomes short (which means that any one of conversation participants speaks in the conversation). This criterion for evaluation is common in all conversation information items (speaking information of each terminal apparatus 10 ) related to the same conversation.
- the features of a conversation related to the conversation information are extracted by the conversation information analysis of the conversation information analysis section 202 .
- the way of participation of each participant in the conversation is specified by performing the analysis as described above.
- the above criteria for evaluation are just examples of information indicating the features of a conversation, and criteria for evaluation according to the purpose or aspect of use of the system of the present exemplary embodiment may be set by adopting other evaluation criteria or giving a weighting to each item.
- the tendency of communication in a group to which the wearer of the terminal apparatus 10 belongs may be analyzed by performing the above analysis for various kinds of conversation information detected by the conversation information detecting section 201 among the speaking information items stored in the data storage unit 22 .
- the tendencies of conversation in a wearer's group may be determined by checking the number of conversation participants, conversation time, correlation between the values, such as the degree of interaction and the degree of conversation activity, and the occurrence frequency of conversation, for example.
- the communication tendency of each wearer may be analyzed by performing the analysis as described above for plural conversation information items of a specific wearer.
- the way of participation of a specific wearer in the conversation may have a certain tendency according to the conditions, such as a conversation partner or the number of conversation participants. Therefore, it may be expected that the features, such as “the speaking level increases in a conversation with a specific partner” or “the degree of listening becomes large if the number of conversation participants increases”, are detected by checking the plural conversation information items in a specific wearer.
- the speaking information identification processing and the conversation information analysis processing described above just show the application example of the system according to the present exemplary embodiment, and do not limit the purpose or aspect of use of the system according to the present exemplary embodiment, the function of the host apparatus 20 , and the like.
- the processing function for executing various kinds of analyses and examinations for the speaking information acquired by the terminal apparatus 10 according to the present exemplary embodiment may be realized as a function of the host apparatus 20 .
Landscapes
- Physics & Mathematics (AREA)
- Engineering & Computer Science (AREA)
- General Physics & Mathematics (AREA)
- Radar, Positioning & Navigation (AREA)
- Remote Sensing (AREA)
- User Interface Of Digital Computer (AREA)
- Measurement Of The Respiration, Hearing Ability, Form, And Blood Characteristics Of Living Organisms (AREA)
Abstract
Description
- This application is based on and claims priority under 35 USC 119 from Japanese Patent Application No. 2012-003805 filed Jan. 12, 2012.
- The present invention relates to a motion analyzer, a voice acquisition apparatus, a motion analysis system, and a motion analysis method.
- According to an aspect of the invention, there is provided a motion analyzer including: a motion detection unit that detects motion of a part of a body of a subject; a speaking detection unit that detects speaking of the subject; and a determination unit that determines that the subject has performed predetermined motion when motion of a part of the body is detected by the motion detection unit and speaking of the subject is detected by the speaking detection unit.
- Exemplary embodiments of the present invention will be described in detail based on the following figures, wherein:
-
FIG. 1 is a view showing an example of the configuration of a motion analysis system; -
FIG. 2 is a view showing an example of the configuration of a terminal apparatus; -
FIG. 3 is a view showing a positional relationship between mouths (speaking portions) of a wearer and another person and a microphone; -
FIG. 4 is a view showing a relationship of sound pressure (input volume) and a distance between a microphone and a sound source; -
FIG. 5 is a view showing a method of identifying a voice of a wearer and a voice of another person; -
FIG. 6 is a flow chart showing the basic operation of a terminal apparatus; -
FIG. 7 is a view showing an output from an acceleration sensor; -
FIG. 8 is a view showing an output of an acceleration sensor when a wearer is in a conversation and the speaking information of the wearer. -
FIG. 9 is a view showing a situation where plural wearers who wear the terminal apparatus according to the exemplary embodiment have a conversation with each other; -
FIG. 10 is a view showing an example of the speaking information of each terminal apparatus in the conversation situation shown inFIG. 9 ; and -
FIG. 11 is a view showing an example of the functional configuration of a host apparatus in an application example. - Hereinafter, an exemplary embodiment of the invention will be described in detail with reference to the accompanying drawings.
-
FIG. 1 is a view showing an example of the configuration of a motion analysis system of the present exemplary embodiment. - As shown in
FIG. 1 , the system according to the present exemplary embodiment is configured to include aterminal apparatus 10 and ahost apparatus 20. Theterminal apparatus 10 and thehost apparatus 20 are connected to each other through a radio communication line. As the radio communication line, lines based on the known methods, such as Wireless Fidelity (Wi-Fi) (registered trademark), Bluetooth (registered trademark), ZigBee (registered trademark), and Ultra Wideband (UWB), may be used. In addition, although only oneterminal apparatus 10 is described in the example shown inFIG. 1 , a number ofterminal apparatuses 10 equal to the number of users are prepared in practice since theterminal apparatus 10 is used while worn by the user. This will be described in detail later. Hereinafter, the user who wears theterminal apparatus 10 is called a wearer. - The
terminal apparatus 10 as an example of a voice acquisition apparatus includes at least two microphones (afirst microphone 111 and a second microphone 112) and amplifiers (afirst amplifier 131 and a second amplifier 132). In addition, theterminal apparatus 10 includes avoice analysis unit 15 that analyzes the recorded voice and adata transmission unit 16 that transmits an analysis result to thehost apparatus 20, and further includes apower supply unit 17. - In addition, the
terminal apparatus 10 includes atimepiece unit 18 and anacceleration sensor 19. Here, thetimepiece unit 18 may be configured using a clock IC (Integrated Circuit), such as a real-time clock, and counts the passage of time and generates calendar information including current year, month, and date and time information (hereinafter, “calendar information and time information” is called “date information”). In addition, thetimepiece unit 18 outputs the generated date information to thevoice analysis unit 15 and thedata transmission unit 16. - The first and
second microphones - Here, the
first microphone 111 as an example of a first voice acquisition unit is disposed at the position (for example, about 35 cm) far from the mouth (speaking portion) of the wearer. In addition, thesecond microphone 112 as an example of a second voice acquisition unit is disposed at the position (for example, about 10 cm) near the mouth (speaking portion) of the wearer. In addition, various types of known microphones, such as a dynamic type microphone and a capacitor type microphone, may be used as the first andsecond microphones - The first and
second amplifiers second microphones second amplifiers - The
voice analysis unit 15 analyzes the voice signals output from the first andsecond amplifiers voice analysis unit 15 identifies whether the voices acquired by the first andsecond microphones terminal apparatus 10, or voices from others. That is, thevoice analysis unit 15 functions as an identification unit that identifies a speaker of the voice on the basis of the voices acquired by the first andsecond microphones - The
data transmission unit 16 transmits the acquired data including the analysis result of thevoice analysis unit 15, an ID of theterminal apparatus 10, the output from theacceleration sensor 19, and the date information generated by thetimepiece unit 18 to thehost apparatus 20 through the radio communication line. As the information transmitted to thehost apparatus 20, for example, information regarding sound pressure of the acquired voice and the like may be included in addition to the analysis result according to processing performed in thehost apparatus 20. In addition, the analysis result of thevoice analysis unit 15 may be stored in theterminal apparatus 10, and data stored for a certain period of time may be collectively transmitted. In addition, the data may be transmitted through a cable line. - The
power supply unit 17 supplies electric power to the first andsecond microphones second amplifiers voice analysis unit 15, thedata transmission unit 16, thetimepiece unit 18, and theacceleration sensor 19. As the power supply, it is possible to use known power supplies, such as a dry battery or a rechargeable battery, for example. In addition, thepower supply unit 17 includes known circuits, such as a voltage conversion circuit and a charging control circuit, when necessary. - The
host apparatus 20 includes adata receiving unit 21 that receives the data transmitted from theterminal apparatus 10, adata storage unit 22 that stores the received data, adata analysis unit 23 that analyzes the stored data, and anoutput unit 24 that outputs the analysis result. Thehost apparatus 20 is realized by an information processing apparatus, such as a personal computer, for example. Moreover, in the present exemplary embodiment, theplural terminal apparatuses 10 are used as described above, and thehost apparatus 20 receives the data from each of theplural terminal apparatuses 10. - The
data receiving unit 21 corresponds to the radio communication line described above, and receives the data from eachterminal apparatus 10 and transmits the received data to thedata storage unit 22. Thedata storage unit 22 is realized by a memory of a magnetic disk device of a personal computer, for example, and stores the received data acquired from thedata receiving unit 21 for each speaker. Here, identification of a speaker is performed on the basis of a terminal ID transmitted from theterminal apparatus 10 and a combination of a speaker name and a terminal ID registered in thehost apparatus 20 in advance. In addition, instead of the terminal ID, a wearer's name may be transmitted from theterminal apparatus 10. - The
data analysis unit 23 is realized by a CPU program-controlled by a personal computer, for example, and analyzes the data stored in thedata storage unit 22. As the specific analysis content and analysis method, various kinds of content and methods may be adopted depending on the purpose or aspect of use of the system according to the present exemplary embodiment. For example, the frequency of conversation between wearers of theterminal apparatuses 10 or the tendencies of a conversation partner of each wearer are analyzed, or the relationship between speakers in a conversation is estimated from the information regarding the length or sound pressure of the voice in the conversation. - The
output unit 24 outputs the analysis result of thedata analysis unit 23 or performs output based on the analysis result. As the output unit, various kinds of units including display of a display device, printout using a printer, and voice output may be adopted according to the purpose or aspect of use of the system, the content or format of an analysis result, and the like. - <Example of the Configuration of a Terminal Apparatus>
-
FIG. 2 is a view showing an example of the configuration of theterminal apparatus 10. - As described above, the
terminal apparatus 10 is used in a state worn by each user. Theterminal apparatus 10 in the present exemplary embodiment is configured to include anapparatus body 30 and astrap 40 connected to theapparatus body 30 so that the user may wear theterminal apparatus 10, as shown inFIG. 2 . In the configuration shown inFIG. 2 , the user wears thestrap 40 on the neck so that theapparatus body 30 is hung from the neck. - The
apparatus body 30 is configured such that at least circuits for realizing the first andsecond amplifiers voice analysis unit 15, thedata transmission unit 16, thepower supply unit 17, and thetimepiece unit 18 and a power supply (battery) of thepower supply unit 17 are housed in a rectangular parallelepipedthin case 31 formed of metal, resin, or the like. A pocket through which an ID card, on which ID information such as the name or affiliation of the wearer is displayed, is inserted may be provided in thecase 31. In addition, such ID information or the like may be printed on the surface of thecase 31, or a seal on which the ID information or the like is described may be attached to the surface of thecase 31. - The first and
second microphones microphones second microphones acceleration sensor 19 functioning as a motion detection unit is provided in thestrap 40. Here, themicrophones second amplifiers strap 40. In addition, theacceleration sensor 19 is connected to thedata transmission unit 16 by cables (electric wires or the like) passing through the inside of thestrap 40. - Here, as materials of the
strap 40, it is possible to use various known materials, such as leather, synthetic leather, cotton, other natural fibers, synthetic fiber using resin, and metal. In addition, coating processing using silicon resin, fluorine resin, or the like may be performed. In addition, thestrap 40 has a cylindrical structure, and themicrophones acceleration sensor 19 are housed inside thestrap 40. Here, when themicrophones strap 40, it is possible to prevent damage or contamination of themicrophones microphones first microphone 111 disposed at the position far from the mouth (speaking portion) of the wearer may be provided in theapparatus body 30 so as to be housed in thecase 31. In the present exemplary embodiment, a case where thefirst microphone 111 is provided in thestrap 40 will be described as an example. - In addition, the
microphones strap 40 where the abdomen of the wearer is located. In addition, theacceleration sensor 19 is provided in a portion of thestrap 40 on the back of the wearer side. More specifically, theacceleration sensor 19 is provided in a portion, which is located behind the neck of the wearer, of thestrap 40. - In addition, the
first microphone 111 is provided at the end (for example, a position within 10 cm from a connection portion) of thestrap 40 connected to theapparatus body 30. Accordingly, thefirst microphone 111 is disposed at the position distant from the mouth (speaking portion) of the wearer by about 30 cm to 40 cm in a state where the wearer wears thestrap 40 on the neck so that theapparatus body 30 is hung from the neck. In addition, also when thefirst microphone 111 is provided in theapparatus body 30, the distance from the mouth (speaking portion) of the wearer to thefirst microphone 111 is almost the same as above. - The
second microphone 112 is provided at the position (for example, a position of about 20 cm to 30 cm from a connection portion) distant from the end of thestrap 40 connected to theapparatus body 30. Accordingly, in a state where the wearer wears thestrap 40 on the neck so that theapparatus body 30 is hung from the neck, thesecond microphone 112 is located on the neck (for example, a position equivalent to the collarbone) of the wearer and is disposed at the position distant from the mouth (speaking portion) of the wearer by about 10 cm to 20 cm. - More specifically, in the present exemplary embodiment, the
strap 40 formed in an annular shape is hung on the neck of the wearer, and thisstrap 40 is pulled downward by theapparatus body 30. As a result, afirst portion 41, which starts from a portion connected to theapparatus body 30 and extends toward the back of the neck of the wearer through the right shoulder side of the wearer, and asecond portion 42, which starts from a portion connected to theapparatus body 30 and extends toward the back of the neck of the wearer through the left shoulder side of the wearer, are provided in thestrap 40. In addition, in the present exemplary embodiment, the first andsecond microphones first portion 41. In addition, thefirst microphone 111 may be provided in one of the first andsecond portions second microphone 112 may be provided in the other portion of the first andsecond portions - Here, the
microphones acceleration sensor 19 are not limited to being provided in thestrap 40, and may be worn on the wearer by various methods. For example, each of the first andsecond microphones acceleration sensor 19 may be separately fixed to clothes using a pin or the like. In addition, it is also possible to prepare a dedicated mounting fixture designed so that each of the first andsecond microphones acceleration sensor 19 is fixed at a desired position. - In addition, the configuration of the
apparatus body 30 is not limited to the configuration shown inFIG. 2 in which theapparatus body 30 is connected to thestrap 40 so as to be hung from the neck of the wearer, and theapparatus body 30 may be formed as a device which may be easily carried. For example, theapparatus body 30 may be mounted on clothes or the body using a clip or a belt instead of the strap in the present exemplary embodiment, or theapparatus body 30 may simply be carried in a pocket or the like. - In addition, mobile phones or known portable electronic information terminals may be configured to realize a function of receiving a voice signal from the
microphones first microphone 111 is provided in theapparatus body 30, it is necessary to maintain the positional relationship between the first andsecond microphones apparatus body 30 at the time of carrying is specified. - In addition, the
microphones acceleration sensor 19 and the apparatus body 30 (or the data transmission unit 16) may be connected to each other through radio communication instead of being connected using a cable. In addition, although the first andsecond amplifiers voice analysis unit 15, thedata transmission unit 16, thepower supply unit 17, and thetimepiece unit 18 are housed in thesingle case 31 in the above example of the configuration, they may be grouped into plural parts. For example, thepower supply unit 17 may be connected to an external power supply without being housed in thecase 31. - <Identification of a Speaker (Wearer and Others) Based on Non-Linguistic Information of Recorded Voice>
- Next, a method of identifying a speaker in the present exemplary embodiment will be described.
- In the present exemplary embodiment, it is identified whether the speaker of the recorded voice is the wearer or not. In addition, in the present exemplary embodiment, speaker identification is performed on the basis of non-linguistic information, such as sound pressure (volume input to the
microphones 111 and 112), instead of linguistic information acquired using morphological analysis or dictionary information, among information items of the recorded voice. That is, a speaker of the voice is identified from the speaking situation specified by the non-linguistic information instead of the content of speaking specified by the linguistic information. - As described with reference to
FIGS. 1 and 2 , in the present exemplary embodiment, thefirst microphone 111 of theterminal apparatus 10 is disposed at the position far from the mouth (speaking portion) of the wearer and thesecond microphone 112 is disposed at the position near the mouth (speaking portion) of the wearer. That is, assuming that the mouth (speaking portion) of the wearer is a sound source, the distance between thefirst microphone 111 and the sound source is greatly different from the distance between thesecond microphone 112 and the sound source. - Specifically, the distance between the
first microphone 111 and the sound source is about 1.5 to 4 times the distance between thesecond microphone 112 and the sound source. Here, sound pressure of the recorded voice in each of themicrophones microphones first microphone 111 is greatly different from the sound pressure of the recorded voice in thesecond microphone 112. - On the other hand, assuming that the mouth (speaking portion) of a person other than the wearer (another person) is a sound source, the distance between the
first microphone 111 and the sound source and the distance between thesecond microphone 112 and the sound source do not change greatly since the other person is away from the wearer. Although there may be a difference between both the distances depending on the position of another person with respect to the wearer, the distance between thefirst microphone 111 and the sound source is not several times the distance between thesecond microphone 112 and the sound source unlike the case when the mouth (speaking portion) of the wearer is a sound source. Therefore, for the voice of another person, the sound pressure of the recorded voice in thefirst microphone 111 is not largely different from the sound pressure of the recorded voice in thesecond microphone 112 unlike the case of the voice of the wearer. -
FIG. 3 is a view showing the positional relationship between mouths (speaking portions) of a wearer and another person and themicrophones - In the relationship shown in
FIG. 3 , the distance between a sound source a, which is a mouth (speaking portion) of the wearer, and thefirst microphone 111 is La1, and the distance between the sound source a and thesecond microphone 112 is La2. In addition, the distance between a sound source b, which is a mouth (speaking portion) of another person, and thefirst microphone 111 is Lb1, and the distance between the sound source b and thesecond microphone 112 is Lb2. In this case, the following relationship is satisfied. -
La1>La2(La1≅1.5×La2 to 4×La2) -
Lb1≅Lb2 -
FIG. 4 is a view showing the relationship of the distance between each of themicrophones - As described above, the sound pressure distance-decreases with the distance between each of the
microphones FIG. 4 , when sound pressure Ga1 in the case of the distance La1 is compared with sound pressure Ga2 in the case of the distance La2, the sound pressure Ga2 is about 4 times the sound pressure Ga1. On the other hand, since the distances Lb1 and Lb2 are approximately equal, sound pressure Gb1 in the case of the distance Lb1 and sound pressure Gb2 in the case of the distance Lb2 are almost equal. - In the present exemplary embodiment, therefore, a voice of the wearer and voices of others in the recorded voices are identified using the difference in sound pressure ratio. In addition, although the distances Lb1 and Lb2 are 60 cm in the example shown in
FIG. 4 , it is meaningful that the sound pressure Gb1 and the sound pressure Gb2 are almost equal, and the distance Lb1 and Lb2 are not limited to the values shown inFIG. 4 . -
FIG. 5 is a view showing a method of identifying a voice of a wearer and a voice of another person. - As described with reference to
FIG. 4 , for the voice of the wearer, the sound pressure Ga2 of thesecond microphone 112 is several times (for example, about 4 times) the sound pressure Ga1 of thefirst microphone 111. In addition, for the voice of another person, the sound pressure Gb2 of thesecond microphone 112 is almost equal to (about 1 times) the sound pressure Gb1 of thefirst microphone 111. - In the present exemplary embodiment, therefore, a threshold value of the ratio between the sound pressure of the
second microphone 112 and the sound pressure of thefirst microphone 111 is set. In addition, a voice with a larger sound pressure ratio than the threshold value is determined to be the voice of the wearer, and a voice with a smaller sound pressure ratio than the threshold value is determined to be the voice of another person. In the example shown inFIG. 5 , the threshold value is 2. Since the sound pressure ratio Ga2/Ga1 exceeds thethreshold value 2, the voice is determined to be the voice of the wearer. In addition, since the sound pressure ratio Gb2/Gb1 is smaller than thethreshold value 2, the voice is determined to be the voice of another person. - Incidentally, not only the voice but also so-called noise, such as environmental sounds, is included in the voice recorded by the
microphones microphones FIGS. 4 and 5 , assuming that the distance between a sound source c of noise and thefirst microphone 111 is Lc1 and the distance between the sound source c of noise and thesecond microphone 112 is Lc2, the distances Lc1 and Lc2 are approximately equal. In addition, the sound pressure ratio Gc2/Gc1 in recorded voices of themicrophones threshold value 2. However, such noise is separated and removed from the voice by performing filtering based on the known technique using a band pass filter, a gain filter, or the like. - <Example of an Operation of a Terminal Apparatus>
-
FIG. 6 is a flow chart showing the basic operation of theterminal apparatus 10. - As shown in
FIG. 6 , when themicrophones terminal apparatus 10 acquire voices, electric signals (voice signals) corresponding to the acquired voices are transmitted from themicrophones second amplifiers microphones second amplifiers - The
voice analysis unit 15 receives the signals amplified by the first andsecond amplifiers voice analysis unit 15 calculates the average sound pressure in the recorded voice of each of themicrophones - Then, when there is a gain of the average sound pressure in each of the
microphones voice analysis unit 15 determines that there is a voice (speaking has been done). Then, thevoice analysis unit 15 calculates the ratio (sound pressure ratio) between the average sound pressure in thefirst microphone 111 and the average sound pressure in the second microphone 112 (step 606). Then, when the sound pressure ratio calculated instep 606 is larger than the threshold value (Yes in step 607), thevoice analysis unit 15 that functions as a voice detection unit determines that the voice is from the wearer (step 608) , and detects the speaking of the wearer. In addition, when the sound pressure ratio calculated instep 606 is smaller than the threshold value (No in step 607), thevoice analysis unit 15 determines that the voice is from another person (step 609). - On the other hand, when there is no gain of the average sound pressure in each of the
microphones voice analysis unit 15 determines that there is no voice (speaking has not been performed) (step 610). In addition, regarding the determination in step 605, it may be determined that there is a gain when the value of the gain of average sound pressure is equal to or larger than the predetermined value, in consideration of a case where noise which is not removed by filtering processing in step 603 remains in signals. - Then, the
data transmission unit 16 transmits the information obtained by the processing in steps 604 to 610, as an analysis result, to the host apparatus 20 (step 611). Specifically, thedata transmission unit 16 transmits to thehost apparatus 20 information regarding whether or not there has been speaking of the wearer, information of date and time of speaking of the wearer when the wearer has spoken, information regarding whether or not there has been speaking of a person other than the wearer (others), information of date and time of speaking of others when a person other than the wearer (others) has spoken, and the like. In addition, the date and time of speaking of the wearer and the date and time of speaking of others are checked by thevoice analysis unit 15 on the basis of the date information from thetimepiece unit 18, the outputs from themicrophones host apparatus 20 together with the analysis result. - In addition, although the explanation is omitted, the
data transmission unit 16 also transmits the output of theacceleration sensor 19 to thehost apparatus 20 when necessary, so that the detection result of theacceleration sensor 19 is acquired by thehost apparatus 20. In addition, in transmitting the output of theacceleration sensor 19 to thehost apparatus 20, thedata transmission unit 16 also transmits the date information generated by thetimepiece unit 18 together. In addition, in the present exemplary embodiment, thedata storage unit 22 of thehost apparatus 20 stores the output from theacceleration sensor 19 and this date information so as to match each other. - In addition, in the present exemplary embodiment, determination regarding whether the voice is a voice from the wearer or a voice from another person is performed by comparing the sound pressure of the
first microphone 111 with the sound pressure of thesecond microphone 112. In the present exemplary embodiment, any kind of speaker identification may be performed if it is performed on the basis of the non-linguistic information extracted from the voice signals themselves acquired by themicrophones - For example, it is also possible to compare the voice acquisition time (output time of a voice signal) in the
first microphone 111 with the voice acquisition time in thesecond microphone 112. In this case, since there is a large difference between the distance from the mouth (speaking portion) of the wearer to thefirst microphone 111 and the distance from the mouth (speaking portion) of the wearer to thesecond microphone 112, a difference (time difference) in voice acquisition time occurs to some extent for the voice of the wearer. - On the other hand, since there is a small difference between the distance from the mouth (speaking portion) to the
first microphone 111 and the distance from the mouth (speaking portion) to thesecond microphone 112, a difference in voice acquisition time for the voice of another person is smaller than that in the case of the voice of the wearer. Therefore, it is possible to set a threshold value of the time difference in voice acquisition time and to determine that the voice is from the wearer when the time difference in voice acquisition time is larger than the threshold value and determine that the voice is from another person when the time difference in voice acquisition time is smaller than the threshold value. - <Application Example of a System and Functions of a Host Apparatus>
- In the system according to the present exemplary embodiment, information regarding speaking (hereinafter, referred to as speaking information) obtained as described above by the plural
terminal apparatuses 10 are collected in thehost apparatus 20. In addition, using the information acquired from theterminal apparatus 10, thehost apparatus 20 performs various kinds of processing according to the purpose or aspect of use of the system. For example, the host apparatus 20 (data analysis unit 23 of the host apparatus 20) analyzes the information acquired from theterminal apparatus 10 and determines whether or not the wearer has made a “nodding” motion which is to shake the head back and forth. - Here, in people's conversation, when an affirmative answer or the like is made, the head and neck are shaken back and forth to make a “nodding” motion in many cases. In the system according to the present exemplary embodiment, the
acceleration sensor 19 is located behind the wearer's neck, and the “nodding” motion may be detected using the output from theacceleration sensor 19. More specifically, in the present exemplary embodiment, not only the identification of a speaker but also detection regarding whether or not a wearer has made an affirmative answer may be performed. - However, similar motions to the motion made at the time of “nodding” are made in many cases by the wearer. When the “nodding” motion is detected only on the basis of the output from the
acceleration sensor 19, incorrect detection easily occurs. This will be described in detail with reference toFIG. 7 which shows the output from theacceleration sensor 19. InFIG. 7 , even though the “nodding” motion is made in practice only in places indicated by thereference numerals reference numerals - In the present exemplary embodiment, in order to improve the detection accuracy of the “nodding” motion, not only the output from the
acceleration sensor 19 but also speaking information of the wearer is used. More specifically, in the present exemplary embodiment, when there is wearer's speaking and there is an output from theacceleration sensor 19 which satisfies the predetermined conditions, it is determined that there is a “nodding” motion. -
FIG. 8 is a view showing an output of theacceleration sensor 19 when a wearer is in a conversation and the speaking information of the wearer. In addition, inFIG. 8 , “nodding” of the wearer is actually performed in places indicated by thereference numerals - Here, as shown in
FIG. 8 , wearer's speaking is detected in the places where the “nodding” is actually performed. On the other hand, when motions other than “nodding” motion (in places other than the places indicated by thereference numerals acceleration sensor 19 which satisfies the predetermined conditions, the accuracy of determination regarding whether or not there has been “nodding” is improved. - Here, the process of determining whether or not there has been a “nodding” motion will be specifically described. In the present exemplary embodiment, when determining whether or not a “nodding” motion has been made, the data analysis unit analyzes the output of the
acceleration sensor 19 transmitted from theterminal apparatus 10 first and determines whether or not a signal waveform (output) satisfying the predetermined conditions is included in the output of theacceleration sensor 19. In addition, when it is determined that the signal waveform satisfying the predetermined conditions is included in the output of theacceleration sensor 19, time information (date information) when the signal waveform is output is checked. - In addition, the
data analysis unit 23 checks the date and time of wearer's speaking when the wearer has spoken on the basis of the information transmitted from theterminal apparatus 10. In addition, thedata analysis unit 23 determines that a “nodding” motion has been made by the wearer when the time interval between the date and time when the signal waveform satisfying the predetermined conditions is output and the date and time specified by the above date and time of wearer's speaking is smaller than the predetermined time interval. On the other hand, when the time interval between the date and time when the signal waveform satisfying the predetermined conditions is output and the date and time specified by the above date and time of wearer's speaking is larger than the predetermined time interval, thedata analysis unit 23 determines that no “nodding” motion has been made by the wearer. - In addition, determination regarding whether or not a “nodding” motion has been made may be made by performing video photographing of persons in a conversation and by observing the video image after photographing, for example. In this case, however, it is necessary to construct a large-scale system. In the present exemplary embodiment, the determination regarding whether or not “nodding” has been made is realized with a simple configuration.
- In addition, although the case where the determination regarding whether or not a “nodding” motion has been made is performed by the
host apparatus 20 has been described above, the determination regarding whether or not a “nodding” motion has been made may also be performed by theterminal apparatus 10. In addition, although theacceleration sensor 19 is disposed so as to be located behind the neck in the above explanation, theacceleration sensor 19 may also be provided in other locations, such as the back of the head, as long as the location is a place linked to the swing of the wearer's neck (movement of the wearer's head). In addition, although the case where theacceleration sensor 19 is attached to thestrap 40 has been described above, theacceleration sensor 19 may also be provided in a hat, an earphone, a headphone, or the like. - In addition, although the case of detecting the “nodding” motion performed when giving an affirmative answer to the question has been described as an example, for example, a swing (swing in the left and right direction) motion performed when giving a negative answer to the question may also be detected by the same process as described above. In addition, when the swing in the left and right direction is performed, a different signal from the signal described above is output from the
acceleration sensor 19. Moreover, in this case, it is determined that there has been a swing according to the negative answer when the different signal is output and there is a wearer's speaking. - Alternatively, when there is wearer's speaking for a shorter time than the predetermined time and a signal (output) satisfying the predetermined conditions has been output from the
acceleration sensor 19, it may be determined that there is a “nodding” motion. More specifically, when a time indicated by the reference numeral T inFIG. 8 (period for which wearer's speaking continues) is shorter than the predetermined time and a signal satisfying the predetermined conditions has been output from theacceleration sensor 19, it may be determined that there is a “nodding” motion. This is because speaking made with a “nodding” motion ends in a short time in many cases and accordingly, a possibility of speaking made at the time of motion other than the “nodding” motion is high when the speaking time is long. - Hereinafter, an example will be described in which the system according to the present exemplary embodiment is used as a system that acquires the information regarding communication of plural wearers.
-
FIG. 9 is a view showing a situation where plural wearers who wear theterminal apparatuses 10 according to the present exemplary embodiment have a conversation with each other.FIG. 10 is a view showing an example of the speaking information ofterminal apparatuses FIG. 9 . - As shown in
FIG. 9 , a case is considered in which two wearers A and B each of whom wears theterminal apparatus 10 have a conversation with each other. In this case, a voice recognized as wearer's speaking in theterminal apparatus 10A of the wearer A is recognized as another person's speaking in theterminal apparatus 10B of the wearer B. On the contrary, a voice recognized as wearer's speaking in theterminal apparatus 10B is recognized as another person's speaking in theterminal apparatus 10A. - The speaking information is separately transmitted from the
terminal apparatuses host apparatus 20. In this case, identification results of speakers (wearer and another person) in the speaking information acquired from theterminal apparatus 10A and the speaking information acquired from theterminal apparatus 10B are opposite as shown inFIG. 10 , but the information showing the speaking situation, such as the length of speaking time or the speaker change timing, is similar. - Therefore, the
host apparatus 20 in this application example determines that such information items indicate the same speaking situation by comparing the information acquired from theterminal apparatus 10A with the information acquired from theterminal apparatus 10B, and recognizes that the wearers A and B have a conversation. Here, at least the time information regarding the speaking, such as the length of speaking time in each speaking of each speaker, start time and end time of each speaking, and speaker change time (timing), is used as the information showing the speaking situation. In addition, in order to determine the speaking situation related to a specific conversation, only some of the time information regarding the speaking may be used, or other information may be additionally used. -
FIG. 11 is a view showing an example of the functional configuration of thehost apparatus 20 in this application example. - In this application example, the
host apparatus 20 includes a conversationinformation detecting section 201 that detects the speaking information (hereinafter, referred to as conversation information) from theterminal apparatus 10 of the wearer in conversation, among the speaking information items acquired from theterminal apparatus 10, and a conversationinformation analysis section 202 that analyzes the detected conversation information. The conversationinformation detecting section 201 and the conversationinformation analysis section 202 are realized as a function of thedata analysis unit 23. - Also from the
terminal apparatus 10 other than theterminal apparatuses host apparatus 20. The speaking information from eachterminal apparatus 10 which is received by thedata receiving unit 21 is stored in thedata storage unit 22. In addition, the conversationinformation detecting section 201 of thedata analysis unit 23 reads the speaking information of eachterminal apparatus 10 stored in thedata storage unit 22 and detects the conversation information which is the speaking information related to a specific conversation. - As shown in
FIG. 10 , from the speaking information of theterminal apparatus 10A and the speaking information of theterminal apparatus 10B, the characteristic correspondence relationship which is different from the speaking information of otherterminal apparatuses 10 is extracted. The conversationinformation detecting section 201 compares the speaking information that is acquired from eachterminal apparatus 10 and is stored in thedata storage unit 22, detects the speaking information with the above-described correspondence relationship from the speaking information acquired from the pluralterminal apparatuses 10, and identifies the speaking information as conversation information related to the same conversation. The speaking information from the pluralterminal apparatuses 10 is transmitted to thehost apparatus 20 when necessary. Accordingly, the conversationinformation detecting section 201 determines whether or not the conversation information related to the specific conversation is included by performing the above-described processing while sequentially separating the speaking information in a fixed period, for example. - In addition, the conditions required when the conversation
information detecting section 201 detects the conversation information related to a specific conversation from the speaking information of the pluralterminal apparatuses 10 are not limited to the correspondence relationship shown inFIG. 10 described above. Any detection method may be used as long as it is possible to identify the conversation information related to a specific conversation from plural speaking information items. - In addition, although the case where two wearers each of whom wears the
terminal apparatus 10 have a conversation is shown in the above example, the number of persons participating in a conversation is not limited to two persons. When three or more wearers have a conversation, theterminal apparatus 10 worn by each wearer recognizes the voice of the wearer of the host apparatus as the voice of the wearer and distinguishes the voice of the wearer from the voices of others (two or more persons). However, the information showing the speaking situation, such as speaking time or speaker change timing, is similar among the acquired information items in the respectiveterminal apparatuses 10. Therefore, similar to the case where two persons have a conversation, the conversationinformation detecting section 201 detects the speaking information acquired from theterminal apparatuses 10 of wearers participating in the same conversation and distinguishes the acquired speaking information from the speaking information acquired from theterminal apparatuses 10 of wearers not participating in the conversation. - Then, the conversation
information analysis section 202 analyzes the conversation information detected by the conversationinformation detecting section 201 and extracts the features of the conversation. In the present exemplary embodiment, as a specific example, the features of the conversation are extracted on the basis of three kinds of criteria for evaluation, that is, the degree of interaction, the degree of listening, and the degree of conversation activity. Here, the degree of interaction is assumed to indicate the balance of the speaking frequency of conversation participant. The degree of listening is assumed to indicate the degree of listening to others in each conversation participant. The degree of conversation activity is assumed to indicate the density of speaking in the entire conversation. - The degree of interaction is specified by the number of times of speaker changes during the conversation and a variation in time taken until the speaker is changed (time for which one speaker speaks continuously). This may be obtained from the number of times of speaker changes and the time when the speaker change occurs in the conversation information in a fixed period. In addition, it is assumed that the value (level) of the degree of interaction increases as the number of times of speaker changes increases, that is, as a variation in continuous speaking time of each speaker decreases. This criterion for evaluation is common in all conversation information items (speaking information of each terminal apparatus 10) related to the same conversation.
- The degree of listening is specified by the ratio of speaking time of each conversation participant and speaking time of others in conversation information. For example, in the following expression, it is assumed that the value (level) of the degree of listening increases as the value of speaking time of others increases.
- Degree of listening=(speaking time of others)/(speaking time of a wearer)
- This criterion for evaluation differs with the speaking information acquired from the
terminal apparatus 10 of each conversation participant even in the conversation information related to the same conversation. - The degree of conversation activity is an index showing the so-called excitement of the conversation, and is specified by the ratio of silence time (time for which none of conversation participants speak) to the total conversation time. It is assumed that the value (level) of the degree of conversation activity increases as the total silence time becomes short (which means that any one of conversation participants speaks in the conversation). This criterion for evaluation is common in all conversation information items (speaking information of each terminal apparatus 10) related to the same conversation.
- As described above, the features of a conversation related to the conversation information are extracted by the conversation information analysis of the conversation
information analysis section 202. In addition, the way of participation of each participant in the conversation is specified by performing the analysis as described above. In addition, the above criteria for evaluation are just examples of information indicating the features of a conversation, and criteria for evaluation according to the purpose or aspect of use of the system of the present exemplary embodiment may be set by adopting other evaluation criteria or giving a weighting to each item. - The tendency of communication in a group to which the wearer of the
terminal apparatus 10 belongs may be analyzed by performing the above analysis for various kinds of conversation information detected by the conversationinformation detecting section 201 among the speaking information items stored in thedata storage unit 22. Specifically, the tendencies of conversation in a wearer's group may be determined by checking the number of conversation participants, conversation time, correlation between the values, such as the degree of interaction and the degree of conversation activity, and the occurrence frequency of conversation, for example. - In addition, the communication tendency of each wearer may be analyzed by performing the analysis as described above for plural conversation information items of a specific wearer. The way of participation of a specific wearer in the conversation may have a certain tendency according to the conditions, such as a conversation partner or the number of conversation participants. Therefore, it may be expected that the features, such as “the speaking level increases in a conversation with a specific partner” or “the degree of listening becomes large if the number of conversation participants increases”, are detected by checking the plural conversation information items in a specific wearer.
- In addition, the speaking information identification processing and the conversation information analysis processing described above just show the application example of the system according to the present exemplary embodiment, and do not limit the purpose or aspect of use of the system according to the present exemplary embodiment, the function of the
host apparatus 20, and the like. The processing function for executing various kinds of analyses and examinations for the speaking information acquired by theterminal apparatus 10 according to the present exemplary embodiment may be realized as a function of thehost apparatus 20. - The foregoing description of the exemplary embodiments of the present invention has been provided for the purpose of illustration and description. It is not intended to be exhaustive or to limit the invention to the precise forms disclosed. Obviously, many modifications and variations will be apparent to practitioners skilled in the art. The embodiments were chosen and described in order to best explain the principles of the invention and its practical applications, thereby enabling others skilled in the art to understand the invention for various embodiments and with the various modifications as are suited to the particular use contemplated. It is intended that the scope of the invention be defined by the following claims and their equivalents.
Claims (11)
Applications Claiming Priority (2)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
JP2012-003805 | 2012-01-12 | ||
JP2012003805A JP2013142843A (en) | 2012-01-12 | 2012-01-12 | Operation analyzer, voice acquisition device, and operation analysis system |
Publications (2)
Publication Number | Publication Date |
---|---|
US20130185076A1 true US20130185076A1 (en) | 2013-07-18 |
US8983843B2 US8983843B2 (en) | 2015-03-17 |
Family
ID=48780611
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
US13/565,365 Active 2033-07-03 US8983843B2 (en) | 2012-01-12 | 2012-08-02 | Motion analyzer having voice acquisition unit, voice acquisition apparatus, motion analysis system having voice acquisition unit, and motion analysis method with voice acquisition |
Country Status (2)
Country | Link |
---|---|
US (1) | US8983843B2 (en) |
JP (1) | JP2013142843A (en) |
Cited By (3)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US20130080169A1 (en) * | 2011-09-27 | 2013-03-28 | Fuji Xerox Co., Ltd. | Audio analysis system, audio analysis apparatus, audio analysis terminal |
US20150302867A1 (en) * | 2014-04-17 | 2015-10-22 | Arthur Charles Tomlin | Conversation detection |
US9922667B2 (en) | 2014-04-17 | 2018-03-20 | Microsoft Technology Licensing, Llc | Conversation, presence and context detection for hologram suppression |
Families Citing this family (3)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN102647525A (en) * | 2012-04-16 | 2012-08-22 | 中兴通讯股份有限公司 | Mobile terminal and processing method on abnormal communication of mobile terminal |
TWI502487B (en) * | 2013-10-24 | 2015-10-01 | Hooloop Corp | Methods for voice management, and related devices and computer program prodcuts |
US10438577B2 (en) * | 2015-10-16 | 2019-10-08 | Sony Corporation | Information processing device and information processing system |
Citations (10)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US6263836B1 (en) * | 1998-11-27 | 2001-07-24 | Robert L. Hollis | Dog behavior monitoring and training apparatus |
US20010034250A1 (en) * | 2000-01-24 | 2001-10-25 | Sanjay Chadha | Hand-held personal computing device with microdisplay |
US6513532B2 (en) * | 2000-01-19 | 2003-02-04 | Healthetech, Inc. | Diet and activity-monitoring device |
US6549792B1 (en) * | 1999-06-25 | 2003-04-15 | Agere Systems Inc. | Accelerometer influenced communication device |
US6956564B1 (en) * | 1997-10-28 | 2005-10-18 | British Telecommunications Public Limited Company | Portable computers |
US20070136064A1 (en) * | 2003-04-16 | 2007-06-14 | Carroll David W | Mobile personal computer with movement sensor |
US7246033B1 (en) * | 2006-03-13 | 2007-07-17 | Susan Leeds Kudo | Pedometer for pets |
US7312981B2 (en) * | 2003-04-16 | 2007-12-25 | Carroll David W | Mobile, hand-held personal computer |
US20090076827A1 (en) * | 2007-09-19 | 2009-03-19 | Clemens Bulitta | Control of plurality of target systems |
US20090164219A1 (en) * | 2007-12-19 | 2009-06-25 | Enbiomedic | Accelerometer-Based Control of Wearable Devices |
Family Cites Families (17)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
JPS57177200A (en) * | 1981-04-25 | 1982-10-30 | Tokyo Shibaura Electric Co | Selector for data |
JPH1124694A (en) * | 1997-07-04 | 1999-01-29 | Sanyo Electric Co Ltd | Instruction recognition device |
JP2000338987A (en) | 1999-05-28 | 2000-12-08 | Mitsubishi Electric Corp | Utterance start monitor, speaker identification device, voice input system, speaker identification system and communication system |
JP3652961B2 (en) | 2000-06-27 | 2005-05-25 | シャープ株式会社 | Audio processing apparatus, audio / video processing apparatus, and recording medium recording audio / video processing program |
JP2002149191A (en) | 2000-11-09 | 2002-05-24 | Toyota Central Res & Dev Lab Inc | Voice input device |
JP2003044075A (en) * | 2001-07-30 | 2003-02-14 | Matsushita Electric Ind Co Ltd | Electronic equipment with speech recognizing function |
JP2003228449A (en) * | 2002-02-01 | 2003-08-15 | Sharp Corp | Interactive apparatus and recording medium with recorded program |
JP2006071936A (en) | 2004-09-01 | 2006-03-16 | Matsushita Electric Works Ltd | Dialogue agent |
JP2006166007A (en) * | 2004-12-07 | 2006-06-22 | Sony Ericsson Mobilecommunications Japan Inc | Method and device for sound source direction detection and imaging device |
JP4992218B2 (en) * | 2005-09-29 | 2012-08-08 | ソニー株式会社 | Information processing apparatus and method, and program |
JP4989871B2 (en) * | 2005-10-05 | 2012-08-01 | 電気化学工業株式会社 | Adhesive tape |
JP2007121576A (en) * | 2005-10-26 | 2007-05-17 | Matsushita Electric Works Ltd | Voice operation device |
JP4669041B2 (en) * | 2006-02-28 | 2011-04-13 | パナソニック株式会社 | Wearable terminal |
JP2007272534A (en) * | 2006-03-31 | 2007-10-18 | Advanced Telecommunication Research Institute International | Abbreviation completion device, abbreviation completion method, and program |
US20080047996A1 (en) * | 2006-07-27 | 2008-02-28 | Frederic Blouin | Intelligent personal carrier |
JP4867798B2 (en) * | 2007-06-05 | 2012-02-01 | ヤマハ株式会社 | Voice detection device, voice conference system, and remote conference system |
JP5143114B2 (en) * | 2009-12-02 | 2013-02-13 | 日本電信電話株式会社 | Preliminary motion detection and transmission method, apparatus and program for speech |
-
2012
- 2012-01-12 JP JP2012003805A patent/JP2013142843A/en active Pending
- 2012-08-02 US US13/565,365 patent/US8983843B2/en active Active
Patent Citations (10)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US6956564B1 (en) * | 1997-10-28 | 2005-10-18 | British Telecommunications Public Limited Company | Portable computers |
US6263836B1 (en) * | 1998-11-27 | 2001-07-24 | Robert L. Hollis | Dog behavior monitoring and training apparatus |
US6549792B1 (en) * | 1999-06-25 | 2003-04-15 | Agere Systems Inc. | Accelerometer influenced communication device |
US6513532B2 (en) * | 2000-01-19 | 2003-02-04 | Healthetech, Inc. | Diet and activity-monitoring device |
US20010034250A1 (en) * | 2000-01-24 | 2001-10-25 | Sanjay Chadha | Hand-held personal computing device with microdisplay |
US20070136064A1 (en) * | 2003-04-16 | 2007-06-14 | Carroll David W | Mobile personal computer with movement sensor |
US7312981B2 (en) * | 2003-04-16 | 2007-12-25 | Carroll David W | Mobile, hand-held personal computer |
US7246033B1 (en) * | 2006-03-13 | 2007-07-17 | Susan Leeds Kudo | Pedometer for pets |
US20090076827A1 (en) * | 2007-09-19 | 2009-03-19 | Clemens Bulitta | Control of plurality of target systems |
US20090164219A1 (en) * | 2007-12-19 | 2009-06-25 | Enbiomedic | Accelerometer-Based Control of Wearable Devices |
Cited By (7)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US20130080169A1 (en) * | 2011-09-27 | 2013-03-28 | Fuji Xerox Co., Ltd. | Audio analysis system, audio analysis apparatus, audio analysis terminal |
US8892424B2 (en) * | 2011-09-27 | 2014-11-18 | Fuji Xerox Co., Ltd. | Audio analysis terminal and system for emotion estimation of a conversation that discriminates utterance of a user and another person |
US20150302867A1 (en) * | 2014-04-17 | 2015-10-22 | Arthur Charles Tomlin | Conversation detection |
US9922667B2 (en) | 2014-04-17 | 2018-03-20 | Microsoft Technology Licensing, Llc | Conversation, presence and context detection for hologram suppression |
US20180137879A1 (en) * | 2014-04-17 | 2018-05-17 | Microsoft Technology Licensing, Llc | Conversation, presence and context detection for hologram suppression |
US10529359B2 (en) * | 2014-04-17 | 2020-01-07 | Microsoft Technology Licensing, Llc | Conversation detection |
US10679648B2 (en) * | 2014-04-17 | 2020-06-09 | Microsoft Technology Licensing, Llc | Conversation, presence and context detection for hologram suppression |
Also Published As
Publication number | Publication date |
---|---|
JP2013142843A (en) | 2013-07-22 |
US8983843B2 (en) | 2015-03-17 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
US9129611B2 (en) | Voice analyzer and voice analysis system | |
US8855331B2 (en) | Audio analysis apparatus | |
US8892424B2 (en) | Audio analysis terminal and system for emotion estimation of a conversation that discriminates utterance of a user and another person | |
US8983843B2 (en) | Motion analyzer having voice acquisition unit, voice acquisition apparatus, motion analysis system having voice acquisition unit, and motion analysis method with voice acquisition | |
US9153244B2 (en) | Voice analyzer | |
US8731213B2 (en) | Voice analyzer for recognizing an arrangement of acquisition units | |
JP6003472B2 (en) | Speech analysis apparatus, speech analysis system and program | |
JP2014077736A (en) | Voice analysis device, voice analysis system, and program | |
JP6160042B2 (en) | Positioning system | |
JP6007487B2 (en) | Information processing apparatus, information processing system, and program | |
JP6191747B2 (en) | Speech analysis apparatus and speech analysis system | |
US9140774B2 (en) | Voice analyzer, voice analysis system, and non-transitory computer readable medium storing a program | |
JP5862318B2 (en) | Sound analysis apparatus, sound analysis system and program | |
US9014395B2 (en) | Audio analysis apparatus and audio analysis system | |
US12100420B2 (en) | Speech detection using multiple acoustic sensors | |
JP2016144134A (en) | Voice analysis device, voice analysis system and program | |
JP2013164468A (en) | Voice analysis device, voice analysis system, and program | |
JP2013140534A (en) | Voice analysis device, voice analysis system, and program | |
JP5929810B2 (en) | Voice analysis system, voice terminal apparatus and program | |
JP6056544B2 (en) | Speech analysis device, signal analysis device, speech analysis system, and program | |
JP2014050048A (en) | Voice analysis device | |
JP2014048618A (en) | Speech analyzing device, speech analyzing system, and program | |
JP2015127768A (en) | Signal analysis apparatus, signal analysis system, and program |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
AS | Assignment |
Owner name: FUJI XEROX CO., LTD., JAPAN Free format text: ASSIGNMENT OF ASSIGNORS INTEREST;ASSIGNORS:YONEYAMA, HIROHITO;SHIMOTANI, KEI;IIDA, KIYOSHI;AND OTHERS;REEL/FRAME:028724/0633 Effective date: 20120726 |
|
STCF | Information on status: patent grant |
Free format text: PATENTED CASE |
|
MAFP | Maintenance fee payment |
Free format text: PAYMENT OF MAINTENANCE FEE, 4TH YEAR, LARGE ENTITY (ORIGINAL EVENT CODE: M1551); ENTITY STATUS OF PATENT OWNER: LARGE ENTITY Year of fee payment: 4 |
|
AS | Assignment |
Owner name: FUJIFILM BUSINESS INNOVATION CORP., JAPAN Free format text: CHANGE OF NAME;ASSIGNOR:FUJI XEROX CO., LTD.;REEL/FRAME:058287/0056 Effective date: 20210401 |
|
MAFP | Maintenance fee payment |
Free format text: PAYMENT OF MAINTENANCE FEE, 8TH YEAR, LARGE ENTITY (ORIGINAL EVENT CODE: M1552); ENTITY STATUS OF PATENT OWNER: LARGE ENTITY Year of fee payment: 8 |