US20190311174A1 - Method and system for 2d and 3d biometric sensing using a same sensor - Google Patents
Method and system for 2d and 3d biometric sensing using a same sensor Download PDFInfo
- Publication number
- US20190311174A1 US20190311174A1 US16/385,437 US201916385437A US2019311174A1 US 20190311174 A1 US20190311174 A1 US 20190311174A1 US 201916385437 A US201916385437 A US 201916385437A US 2019311174 A1 US2019311174 A1 US 2019311174A1
- Authority
- US
- United States
- Prior art keywords
- sensor
- finger
- processor
- elements
- sensor array
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Abandoned
Links
- 238000000034 method Methods 0.000 title abstract description 9
- 238000007920 subcutaneous administration Methods 0.000 claims abstract description 22
- 210000000988 bone and bone Anatomy 0.000 claims description 18
- 230000002792 vascular Effects 0.000 claims description 15
- 230000017531 blood circulation Effects 0.000 claims description 9
- 230000001934 delay Effects 0.000 claims 1
- 238000002604 ultrasonography Methods 0.000 description 27
- 238000003384 imaging method Methods 0.000 description 24
- 210000001519 tissue Anatomy 0.000 description 21
- 206010033675 panniculitis Diseases 0.000 description 17
- 210000004304 subcutaneous tissue Anatomy 0.000 description 17
- 230000005540 biological transmission Effects 0.000 description 7
- 239000000090 biomarker Substances 0.000 description 6
- 210000004204 blood vessel Anatomy 0.000 description 6
- 239000000463 material Substances 0.000 description 5
- 230000008054 signal transmission Effects 0.000 description 5
- 238000012285 ultrasound imaging Methods 0.000 description 5
- 239000000919 ceramic Substances 0.000 description 4
- 238000001514 detection method Methods 0.000 description 4
- 238000013507 mapping Methods 0.000 description 4
- 230000008569 process Effects 0.000 description 4
- 238000005070 sampling Methods 0.000 description 4
- 210000005166 vasculature Anatomy 0.000 description 4
- 238000012795 verification Methods 0.000 description 4
- 238000010586 diagram Methods 0.000 description 3
- 239000000945 filler Substances 0.000 description 3
- 239000011521 glass Substances 0.000 description 3
- 229910052451 lead zirconate titanate Inorganic materials 0.000 description 3
- 238000012545 processing Methods 0.000 description 3
- 230000001681 protective effect Effects 0.000 description 3
- 230000003068 static effect Effects 0.000 description 3
- 230000008901 benefit Effects 0.000 description 2
- 230000015572 biosynthetic process Effects 0.000 description 2
- 239000002131 composite material Substances 0.000 description 2
- 239000004020 conductor Substances 0.000 description 2
- 230000000694 effects Effects 0.000 description 2
- 239000010410 layer Substances 0.000 description 2
- 239000003550 marker Substances 0.000 description 2
- 238000005259 measurement Methods 0.000 description 2
- 230000000877 morphologic effect Effects 0.000 description 2
- 230000010355 oscillation Effects 0.000 description 2
- 230000004044 response Effects 0.000 description 2
- 239000004593 Epoxy Substances 0.000 description 1
- JOYRKODLDBILNP-UHFFFAOYSA-N Ethyl urethane Chemical compound CCOC(N)=O JOYRKODLDBILNP-UHFFFAOYSA-N 0.000 description 1
- 230000008321 arterial blood flow Effects 0.000 description 1
- 239000008280 blood Substances 0.000 description 1
- 210000004369 blood Anatomy 0.000 description 1
- 230000008859 change Effects 0.000 description 1
- 238000004891 communication Methods 0.000 description 1
- 230000008602 contraction Effects 0.000 description 1
- 238000013523 data management Methods 0.000 description 1
- 238000009826 distribution Methods 0.000 description 1
- 230000005684 electric field Effects 0.000 description 1
- 230000005284 excitation Effects 0.000 description 1
- 230000001815 facial effect Effects 0.000 description 1
- 238000001914 filtration Methods 0.000 description 1
- 239000006261 foam material Substances 0.000 description 1
- 230000006870 function Effects 0.000 description 1
- 238000002847 impedance measurement Methods 0.000 description 1
- 230000010354 integration Effects 0.000 description 1
- HFGPZNIAWCZYJU-UHFFFAOYSA-N lead zirconate titanate Chemical compound [O-2].[O-2].[O-2].[O-2].[O-2].[Ti+4].[Zr+4].[Pb+2] HFGPZNIAWCZYJU-UHFFFAOYSA-N 0.000 description 1
- 239000011159 matrix material Substances 0.000 description 1
- 239000004005 microsphere Substances 0.000 description 1
- 238000012986 modification Methods 0.000 description 1
- 230000004048 modification Effects 0.000 description 1
- 238000012544 monitoring process Methods 0.000 description 1
- 230000001766 physiological effect Effects 0.000 description 1
- 229920000642 polymer Polymers 0.000 description 1
- 238000003825 pressing Methods 0.000 description 1
- 239000011253 protective coating Substances 0.000 description 1
- 230000010349 pulsation Effects 0.000 description 1
- 238000001228 spectrum Methods 0.000 description 1
- 239000000758 substrate Substances 0.000 description 1
- 230000002123 temporal effect Effects 0.000 description 1
- 238000002113 ultrasound elastography Methods 0.000 description 1
Images
Classifications
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06V—IMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
- G06V40/00—Recognition of biometric, human-related or animal-related patterns in image or video data
- G06V40/10—Human or animal bodies, e.g. vehicle occupants or pedestrians; Body parts, e.g. hands
- G06V40/12—Fingerprints or palmprints
- G06V40/13—Sensors therefor
- G06V40/1306—Sensors therefor non-optical, e.g. ultrasonic or capacitive sensing
-
- G06K9/0002—
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06V—IMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
- G06V40/00—Recognition of biometric, human-related or animal-related patterns in image or video data
- G06V40/10—Human or animal bodies, e.g. vehicle occupants or pedestrians; Body parts, e.g. hands
- G06V40/12—Fingerprints or palmprints
- G06V40/1382—Detecting the live character of the finger, i.e. distinguishing from a fake or cadaver finger
-
- G06K9/0008—
-
- G06K9/00201—
-
- G06K9/00885—
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06V—IMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
- G06V20/00—Scenes; Scene-specific elements
- G06V20/60—Type of objects
- G06V20/64—Three-dimensional objects
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06V—IMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
- G06V40/00—Recognition of biometric, human-related or animal-related patterns in image or video data
- G06V40/10—Human or animal bodies, e.g. vehicle occupants or pedestrians; Body parts, e.g. hands
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06V—IMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
- G06V40/00—Recognition of biometric, human-related or animal-related patterns in image or video data
- G06V40/10—Human or animal bodies, e.g. vehicle occupants or pedestrians; Body parts, e.g. hands
- G06V40/12—Fingerprints or palmprints
- G06V40/1347—Preprocessing; Feature extraction
- G06V40/1359—Extracting features related to ridge properties; Determining the fingerprint type, e.g. whorl or loop
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06V—IMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
- G06V40/00—Recognition of biometric, human-related or animal-related patterns in image or video data
- G06V40/10—Human or animal bodies, e.g. vehicle occupants or pedestrians; Body parts, e.g. hands
- G06V40/12—Fingerprints or palmprints
- G06V40/1365—Matching; Classification
-
- G06K2009/00932—
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06V—IMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
- G06V40/00—Recognition of biometric, human-related or animal-related patterns in image or video data
- G06V40/10—Human or animal bodies, e.g. vehicle occupants or pedestrians; Body parts, e.g. hands
- G06V40/14—Vascular patterns
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06V—IMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
- G06V40/00—Recognition of biometric, human-related or animal-related patterns in image or video data
- G06V40/10—Human or animal bodies, e.g. vehicle occupants or pedestrians; Body parts, e.g. hands
- G06V40/15—Biometric patterns based on physiological signals, e.g. heartbeat, blood flow
Definitions
- Fingerprinting is one of the most widely used biometric for human identification. Identification is obtained by analyzing a given fingerprint image obtained by a fingerprint sensor for the relative locations and orientations of structural elements such as branching or ending of ridges and valleys known as minutia. These characteristics are obtained in the enrollment mode of a person's finger or multiple fingers. In the verification mode a second fingerprint is obtained and analyzed for similarity based on minutia or other previously defined fingerprint characteristics. This minutia is also referred to herein as a type of biometric marker.
- the probability for false identification either a false acceptance or false rejection depends on the number of minutia identified in the fingerprint.
- the number of minutia increases with the fingertip area being scanned.
- a small area fingerprint sensor is very desirable for integration of fingerprint sensors into mobile devices for access control, such as cell phone.
- Sonavation, Inc. of Palm Beach Gardens, Fla., USA manufactures biometric sensing devices having a ceramic Micro-Electro Mechanical System (MEMS) piezoelectric array that is made from a ceramic composite material.
- MEMS Micro-Electro Mechanical System
- This piezoelectric material is formed into a pillar 1/10th the diameter of a human hair, it has a unique set of properties that enable it to mechanically oscillate when an electric field is applied or create an electrical voltage when mechanically vibrated.
- the piezoelectric pillar is electrically vibrated at its natural ultrasonic resonant frequency. If a fingerprint ridge is directly above the pillar, much of the ultrasonic energy is absorbed by the skin and the signal impedance of the pillar is very high.
- An imaging ASIC electrically controls the pillar oscillation, imaging of the fingerprint and data management of the fingerprint information.
- U.S. Pat. No. 7,141,918 describes an biometric sensing device having the above piezoelectric array operable for fingerprint imaging. It has been found as also described in this patent that the piezoelectric array can be operated in non-fingerprint imaging modes to obtain other biometric information, such as in an echo mode to provide imaging, such as bone, or a Doppler-shift mode to detect blood flow velocity and blood flow patterns. Although the sensor described in this patent is useful, it would be desirable to also operate the sensing device in a three-dimension ultrasound imaging mode to provide improved imaging of subcutaneous structures for use in biometric identification (or medical applications) that does not rely on echo mode imaging as described in U.S. Pat. No. 7,141,918.
- a biometric sensing device having a piezoelectric array providing improved three-dimension imaging of subcutaneous tissue structures of a finger, such as bone or vasculature, utilizing pitch/catch ultrasonically formed images.
- the present invention embodies a biometric sensing device having an array of piezoelectric ceramic elements operable in a first mode for producing first data representative of a fingerprint image, and a second mode for producing second data representative of least one three-dimensional image of subcutaneous tissue structure(s), such as or bone or vascular, formed by pitch-n-catch ultrasound imaging.
- the images provided from operating the sensing device in the first and second modes provide anatomical and morphological biometrics (biometric data) for use in biometric identification.
- the second data representative of least one three-dimensional image of subcutaneous tissue structure may also be used for determining elastic properties of tissue, and vital or proof of life parameters, i.e. physiological information, such as heart beat, blood flow velocities, and pulse wave pattern, or other parameters which can be used to determine if the finger disposed upon the sensor array is fake or dead.
- vital or proof of life parameters i.e. physiological information, such as heart beat, blood flow velocities, and pulse wave pattern, or other parameters which can be used to determine if the finger disposed upon the sensor array is fake or dead.
- the elastic properties of tissue which may, like captured fingerprint image and the one or more images of subcutaneous tissue structure(s), provide biometrics (biometric data) for use in biometric identification.
- biometrics biometric data
- multiple types of biometric data can obtained with a single application of a finger to the sensor array, which can be done in real time and simultaneously.
- the architecture of the identification device is similar to what is described in U.S. Pat. No. 7,141,918, also referred to herein as the '918 patent.
- Embodiments of the subject invention include various improvements over the '918 patent that are described herein. These improvements include those relating to electronic control and data acquisition.
- U.S. Pat. No. 7,141,918 is incorporated herein by reference.
- U.S. Pat. Nos. 7,844,660, and 6,720,712 which are related to U.S. Pat. No. 7,141,918 are also incorporated herein by reference.
- FIG. 1A is a schematic diagram of a top down view of a piezoelectric identification device according to the present invention
- FIG. 1B is a side perspective view of the piezoelectric identification device in which pillars (elements) are evenly spaced and are separated by filler material.
- FIG. 2 a schematic illustration of the sensor array addressing scheme is shown, where there are M and N number of elements 11 , disposed in M rows along an x axis and in N columns along a y axis.
- FIG. 3 illustrates an identification device that is coupled to a computer system.
- FIG. 4 illustrates a cross-section of sensor array 10 and of a finger 30 placed proximate to an upper surface of the sensor array so that the finger's surface 31 is in direct contact with its protective shield 23 .
- FIG. 5 illustrates a scan aperture 40 that is scanned by processor 13 with respect to an x axis (M rows) and y axis (N columns) dimensions of the scan aperture, like shown in FIG. 2 .
- FIG. 6A illustrates a transit aperture 41 that is selected by processor 13 to form a transmit beam or pulse 46 (shown as a translucent volume) having an hour-glass shape, having 6 transmit channels.
- FIG. 6B illustrates transmit signals traveling via channels A-F.
- FIG. 7A illustrates a receive aperture 42 selected by processor 13 to receive acoustic energy (beam or pulse 49 ).
- FIG. 7B illustrates transmit signals received via channels A-F.
- FIG. 8 illustrates a combined output signal 54 .
- FIG. 9 illustrates ultrasound scanning of a blood vessel 50 within a finger via a sensor array.
- FIG. 10 illustrates ultrasound scanning of a bone structure within a finger via a sensor array.
- FIG. 11 illustrates operation of the identification device 9 of FIG. 3 while scanning a finger via a sensor array, like shown in FIGS. 4, 9 and 10 .
- FIG. 12 is a simplified illustration of sensor control hardware and software.
- FIG. 13 illustrates an alternative hardware for transmitting and receiving signals to and from the sensor array.
- Identification device 9 has a sensor array 10 of piezo-electric elements 11 providing mechanical resonator sensing, a sensor input signal generator 12 , and a processor 13 . Under control of processor 13 , the input signal generated by input signal generator 12 is coupled to sensor array 10 by a signal multiplexer 15 a , and output signal of sensor 10 is coupled to processor 13 by a multiplexer 15 b.
- Processor 13 processes the output signals from select element(s) via multiplexor 15 b to obtain biometric data which may then be stored in a memory 14 .
- Biometric data can include one or more fingerprint images, and/or one or more ultrasound images of subcutaneous structures of the finger, subcutaneous tissue parameter(s) such as of tissue elasticity, and/or detected proof of life parameters, as described later below.
- Addressing of elements 11 , via multiplexers 15 a and 15 b is enabled via a mux controller 16 in accordance with user specified imaging modes and/or in detection of proof of life parameters.
- each multiplexor 15 a and 156 is shown singularly, each multiplexor's function may alternatively be designed to be provided by two or more multiplexors as desired.
- Sensor array elements 11 may be of lead zirconate titanate (PZT) or other material having similar properties, preferably, PZT 1-3 composite.
- the piezo-ceramic elements 11 can have shapes other than rectangular, such as circular as shown in FIG. 2 .
- Sensor area 10 is preferably provided on a support layer, such as described in the above incorporated patent.
- Other ASIC chips may provide the electronics shown in FIG. 1A to which the sensor is connected.
- FIG. 1B A more detailed view of sensor array 10 is shown in FIG. 1B in which elements 11 represent evenly spaced pillars with filler material 17 there between.
- Filler material 17 may be an epoxy or polymer having micro-spheres.
- Transmit electrodes (or lines) 19 and receive electrodes (or lines) 20 are provided above and below, respectively, along sensor array 10 .
- a shield or protective coating layer 22 Over the transmit electrodes 19 is a shield or protective coating layer 22 , such as urethane providing a surface 23 upon which a fingertip may be placed.
- a support substrate 24 such as of foam material.
- FIG. 2 a schematic illustration of the addressing scheme is shown, where there are M by N number of elements 11 , disposed in M rows along a horizontal x axis as shown and in N columns along a vertical y axis as shown.
- M and N may equal 368 and 256, respectively, but another number of elements, and row and column grouping thereof, can be employed in accordance with other embodiments of the invention.
- Transmit electrodes 19 are parallel conductors that connect to the top of elements 11 in rows 1 to M
- receive electrodes 20 are parallel conductors that connect to bottom of elements 11 in columns 1 to N.
- each element 11 is 40 microns square by 100 microns deep, thereby yielding a dense sensor array 10 having a 20 MHz fundamental frequency sonic wave elements 11 .
- a spacing of 10 microns is used between elements 11 and is preferred in order to provide a 50-micron total pitch between elements.
- Other geometries may be used, such as for example, a pitch of greater or lower than 50 microns.
- a sensor with 368 rows and 256 columns may be 18.7 by 13 mm, or other size according to the maximum imaging size desired for the particular application.
- sensor array 10 may be manufactured as described in U.S. Pat. No. 7,489,066, which is herein incorporated by reference.
- elements 11 are individually addressable for application of an input signal by row, and then addressable for reading out an output signal by column, by selection of electrodes 19 and 20 , via multiplexors 15 a and 15 b , respectively.
- a ground switch 26 is provided coupled to all transmit electrodes 19 between edge connector 19 a and multiplexor 15 a enabling processor 13 to connect electrodes 19 to ground when needed.
- a ground switch 27 is provided coupled to all receive electrodes 20 between edge connector 20 a and multiplexor 15 b enabling processor 13 enabling processor 13 to connect electrodes 20 to ground when needed.
- the benefit of ground switches 26 and 27 is that it avoids additional switching of ground and signal electrodes as described in U.S. Pat. No. 7,141,918, thereby avoiding unwanted additional capacitive loads parallel to the transmitting and receiving elements 11 .
- processor 13 is programmed within its embedded memory (or memory 14 ) to enable all sensing operations by sensor array 10 as described herein, including at least fingerprint imaging, and three-dimensional ultrasound imaging. Further, processor 13 may provide other activities commonly implemented in an ultrasonic imaging system as part of electronic beam formation including synthetic aperture imaging.
- identification device 9 is coupled to a computer system 28 for data communication to send commands and receive biometric data from identification device 9 .
- Computer system 28 may be any microprocessor-based device operating in accordance with a program or instructions which utilizes identification device 9 to obtain biometric data via sensor array 10 as needed for identification (e.g., verification and/or enrollment) of biometric data.
- Such computer system 28 uses biometric data connected from the sensor to enable biometric feature or identifier matching in accordance with a database or memory within system 28 , or accessible to system 28 .
- computer system 28 may be part of portable identification device, point of transaction system, or any other device requiring secure biometric sensing to enable access to physical and/or electronic resources.
- Identification device 9 is operable in a fingerprint imaging mode, and a three-dimensional subcutaneous tissue structure imaging mode, as described below.
- FIG. 4 illustrates a cross-section of sensor array 10 of the device 9 , and a finger 30 placed proximate to the sensor array so that the finger's surface 31 is in direct contact with its protective shield 23 .
- the finger's Such surface 31 has ridges 32 , which make such direct contact, and valleys 33 which do not make such direct contact with the protective shield 23 .
- an element 11 is driven at a resonant frequency by the input signal generator 12 , via one of transmit electrodes 19 selected by mux controller 16 via multiplexor 15 a , its oscillation is directed or transmitted towards the finger's surface 31 and produces a sonic wave either towards ridges 32 or valleys 33 as graphically illustrated by sonic waves 35 and 36 , respectively.
- the input signal excites elements 11 within the sensor array 10 and causes these elements to oscillate, preferably at a resonant frequency.
- Phenomena outside of the sensor array 10 such as a presence of acoustic impedance or a reflection of acoustic energy off of outside entities (external to the sensor), further influences the motion of these elements 11 (apart from the influence of input signal excitation) within the sensor array 10 .
- Such outside influenced motion is monitored, measured and analyzed by electronic hardware in combination with software that controls operation of the electronic hardware. (See FIG. 12 )
- the surface 31 response to sonic wave differs due to contact to tissue of a ridge versus non-contact of valley difference in impedance (or attenuation/voltage) which is detectable by the same element 11 which transmitted the sonic waves or beam, via one of receive electrodes 20 selected by mux controller 16 via multiplexor 15 b , thereby providing a measure as to whether the element is facing a ridge or a valley.
- the processor 12 builds a map in memory 14 where each element response (output signal) detected by processor 12 represents one pixel of the two-dimensional fingerprint image in memory 14 , where each pixel may be represented as a black or white value to represent a ridge or valley, respectively, or vice versa.
- identification device 9 Such operation of identification device 9 to obtain a fingerprint image is described in more detail in connection with FIGS. 17-22 of in incorporated U.S. Pat. No. 7,141,918 which is included in Appendix A of the prior filed provisional patent application, or other U.S. Pat. Nos. 7,489,066, 7,514,842, 8,331,633, and 8,335,356 which are also all herein incorporated by reference.
- sensor array 10 operates to obtain a fingerprint by detecting the impedance at a resonant frequency of an applied input signal generated 12 , via multiplexor 15 a , where upon soon after a driving input signal of each element 11 ceases in time, and an output signal is read from that same pixel. That output signal is representative of impedance.
- the element 11 ring (vibration) characteristic causes an electrical output signal to be produced by the element that when sampled, via multiplexor 15 b , provides a measure of impedance.
- two impedance measurements can take place at two different frequencies (e.g., 19.8 MHz and 20.2. MHz) for each element 11 , where the difference of measured impedance at each frequency is used to determine whether the element 11 is facing and adjacent to a ridge or a valley as described in incorporated U.S. Pat. No. 7,141,918.
- Identification device 9 can also operate sensor 10 in a pitch/catch imaging mode to obtain three-dimensional ultrasound images within a finger presented to sensor array 10 .
- a sensor principally described for fingerprint image capture can enable viewing of structures within the same tissue that provided a fingerprint image, such as vascular structures (venous and arterial vessels), or bone structure.
- processor 13 operates the elements 11 of the sensor array 10 in this pitch/catch mode by connecting the transmitter and receiver in series, rather than in parallel as in echo imaging of the prior incorporated U.S. Pat. No. 7,141,918.
- an scan aperture 40 is scanned by processor 13 along in x and y axes or dimensions, where the x dimension corresponds to the 1 to M rows, and y dimension corresponds to the 1 to N columns of FIG. 2 .
- the scan aperture 40 is formed along the intersection of a group of “m” number of adjacent rows of elements 11 activated by processor 13 to provide a transmit aperture 41 that produces a transmit beam, and a group of “n” number of adjacent columns of elements 11 activated by processor 13 to provide a receive aperture 42 that receives a receive beam, where the time delay of signals received is representative of structure(s) in the volume of tissue into which the transmit beam is focused and the receive beam is received above scan aperture 40 .
- processor 13 moves the scan aperture 40 along the x and y dimensions by selecting different groups of “m” rows and “n” columns in which to overlap and form different scan apertures 40 .
- the transmit electrodes 19 to the “m” rows of elements 11 are divided equally into “p” number channels, where the number of transmit channels equals “m” divided by “p”.
- the receive electrodes 20 the “n” columns of elements 11 are divided equally into “r” number of receive channels, where the number of receive channels equals “n” divided by “r”.
- FIGS. 6A and 6B An example for one of the multiple scan apertures 40 that may take place during scanning of vasculature within the tissue above sensor array 10 during scanning of multiple different scan apertures is shown in FIGS. 6A and 6B , where FIG. 6A represents a transmit cycle and FIG. 6B represents a receive cycle.
- a transmit aperture 41 is selected by processor 13 to form a transmit beam or pulse 46 (shown as a translucent volume) having an hour-glass shape.
- “m” equals 12 and “p” equals 6, resulting in 6 transmit channels A-F each with two rows.
- the transmit (input) signal is applied by transmit channels A-F and offset in time in accordance with distance of rows from a location in volume 48 .
- the transmit signal is first applied to rows of the most outer transmit channels A and F first, then the second most outer transmit channels B and E, and then the central most transmit channels C and D last, as illustrated in FIG. 6B .
- the transmit aperture 41 forms a transmit beam 46 which will arrive at approximately the same time thereby focusing transmit beam 46 at locations in the intended volume 48 that may contain the object or structure of interest, such as a blood vessel 50 .
- transmit beam 46 all other rows of elements 11 which are not used in the transmit aperture 41 are inactive.
- a blood vessel may or may not be fully included in the transmit beam 46 .
- switch 27 is switched to ground by processor 13 to ground the receive electrodes 20 , while switch 26 is not set to ground.
- transmit beam 46 is launched into the tissue of the finger 30 and an additional period for ring down of the transmit electrodes 19 transmitting elements 11 along the “m” rows (i.e., their electrodes 19 ) are switched to ground by processor 13 via switch 26 , and while switch 27 is not set to ground.
- the receive cycle can then begin.
- FIG. 7A shows an example of the receive aperture 42 selected by processor 13 to receive beam or pulse 49 having an hour-glass shape.
- “n” equals 12 and “r” equals 6, resulting in 6 receive channels A-F each with two columns.
- the receive (output) signal is read (or detected) from receive channels A-F, and read offset in time in accordance with distance of columns from a location in volume 52 .
- the receive signal is first read from columns of the most outer receive channels A and F first, then the second most outer receive channels B and E, and then the central most receive channels C and D last, as illustrated in FIG. 7B .
- the beam received by elements 11 of the receive aperture 42 will arrive at approximately the same time from the intended volume 52 , which in this example includes part of blood vessel 50 .
- the signals from all the receive channels A-F are aligned in accordance with the time offset of reception shown in FIG. 7B and combined by a beam former 53 in processor 13 to form a combined output signal for scan aperture 40 as shown in FIG. 8 .
- Receive beam 49 is orthogonal to the transmit beam 46 , and it is their intersection along transmit aperture 41 and receiver aperture 42 which defines the effective pitch/catch scan aperture 40 .
- the processor 13 receives signal from the “n” column of elements 11 during the sampling interval associated with the round trip time after the ceased transmit beam is backscatter reflected towards the sensor 10 , from the objects or structures desired to be imaged.
- the delay in time of the combined output signal from beam former 53 over the sampling interval represents distance from the sensor array 10 , and the amplitude or value 54 of the signal at different depths along the z dimension sampled during the sampling interval is recorded by processor 13 in memory 14 at x,y,z coordinates associated with that scan aperture 40 .
- the processor 13 may receive combined output signal over the entire depth of the scan aperture 40 , but records information in memory 14 over a desired range of volume's depth of intersecting volumes 48 and 51 of scan aperture 40 to provide a three-dimensional ultrasound image indicating structures of interest which can be within that desired depth range from sensor array 10 .
- FIG. 8 shows an effect upon a beam formed signal by a blood vessel 50 along the z axis at a distance from sensor 10 centered along the scan aperture 40 , where +/ ⁇ amplitude value 54 of coordinate along the x axis at a common y position centered about the receive aperture 42 .
- the amplitude of the signal between the outer boundaries of the vessels diameter may also be processed by processor 13 to provide velocity information of the flow there through at that x, y coordinate for use as a vital parameter).
- this transmit beam 46 is steered in two orthogonal axes x and y within a two-dimensional (2D) scan of the sensor array 11 .
- the 3rd axis, defined as the axial or z-axis is obtained by time of arrival analysis of subcutaneous tissue causing backscattered sound waves by processor 13 .
- the processor adds the information at sampled points of amplitude 54 obtained along the z axis from sensor 10 at the x,y coordinate to a map in memory 14 along the x and y dimensions thereby building a three-dimensional ultrasound image of subcutaneous structures.
- a full 20 x,y image along an x,z plane is obtained from time history in z and receive aperture 42 position in y.
- this 20 image provides a slice along the x,z plane of the full 3D volume presentation of backscattered ultrasound for a given scan aperture 40 . Scanning along the x axis while scanning the receive aperture for each new position creates the full volume representation of the fingertip object.
- switch 26 is switched to ground by processor 13 to ground the transmit electrodes 19 , while switch 27 is not set to ground.
- the processor 13 then repeats the process for different scan apertures 40 along the x any y dimensions over the volume of tissue above sensor array 10 providing multiple slices along x,z planes of scan apertures to complete a three-dimensional ultrasound image of subcutaneous structures.
- the sampling interval for the subcutaneous tissue may be set to provide three-dimensional ultrasound image of the vasculature of finger 30 as illustrated in FIG. 9 , or bone structure of finger 30 as illustrated in FIG. 10 , thereby enabling three-dimensional imaging of different types of subcutaneous structures.
- Other structures in the tissue of the finger may similarly be imaged as desired.
- a subgroup of “n” adjacent transmitters is active providing an electronically focused beam 46 in one lateral direction commonly referred to as azimuth axis.
- the receive aperture 49 is selected as a sub-group of “m” electrodes 20 via the multiplexer 15 b , thus the effective aperture for transmit and receive becomes the spatial intersection between transmit and receive apertures 41 and 42 , respectively.
- Only a sub-group “m” of the M receive electrodes 20 are connected via a multiplexer 15 b to a group of “m” receive amplifiers and signal processing chains for beam formation and further backscatter analysis by processor 13 .
- proof of life parameter(s) are detectable, such as velocity or flow of cells through the vessel, heartbeat, or flow patterns, as desired, in a manner as commonly detected in typical ultrasound imaging system.
- an identification device 9 for identification (or verification) or a subject's finger 30 as presented to sensor array as shown in FIGS. 4, 9, and 10 .
- processor 13 to capture an image (two dimensional representation) of a fingerprint along the surface of finger 30 (step 60 ), as described above, which is stored in memory 14 as minutia (biometric fingerprint identifiers) in relative and local sensor x,y coordinates (step 64 ).
- the fingerprint image may be stored in memory 14 , and/or sent to computer system 28 .
- identification device 9 is switched to three-dimensional ultrasound/volumetric imaging mode.
- An image of subcutaneous fingertip vascular structure of finger 30 is then captured in memory 14 (step 61 ), and processed by processor 13 to obtain biometric data of identifiers uniquely characterizing curvature and/or shape of all or major subcutaneous vascular structure of the finger in relative and local sensor x,y,z coordinates (step 65 ).
- Other tissue characteristics from image may also provide biometric identifiers, such as tissue speckle.
- the three-dimensional ultrasound image may be stored in memory 14 , and/or sent to computer system 28 .
- subcutaneous tissue parameters are measured from the ultrasound image stored in memory 14 .
- the ultrasound image may be processed by processor 13 to determine elastic properties of tissue by applying pressure to the fingertip and estimating the strain in the tissue using typical ultrasound elastography. Reversely, with known tissue elasticity applied pressure is estimated from tissue strain.
- the elastic measure represents another biometric identifier stored in memory 14 .
- the processor 13 using the three-dimensional ultrasound image from step 61 stored in memory 14 determines one or more vital parameters which may be used to reduce the risk that the subject's finger in fake or dead, such as blood flow, vessel wall pulse waves and heart rate parameters.
- Each of the one or more vital parameters are compared with one or more thresholds stored in memory 14 (or by computer system 28 if sent thereto) which if not met indicates that the subject's finger 30 may be fake or dead.
- Blood flow may be identified using common procedure of ultrasonic flow detection, such as described in J. A. Jensen, Estimation of Blood Flow using Ultrasound, Cambridge University Press, 1996, or R. S. C. Cobbold, Foundations of Biomedical Ultrasound, Oxford University Press, 2007. In addition to identifying blood flow, blood mean velocity or maximum velocities as well as flow spectra are obtained. Heart rate and vessel wall motion is detected from lower frequency variations of pulsed and continuous wave ultrasound.
- An image of subcutaneous fingertip bone structure is then captured and stored in memory 14 (step 63 ), and processed by processor 13 to obtain biometric data of identifiers uniquely identifying subcutaneous bone structure of the finger in relative and local sensor x,y,z coordinates (step 65 ).
- Finger bone structure is useful as biometric, particularly if bone curvature or other bone shape identifiers.
- Computer system 28 stores a database of security identification information of previously captured identifiers of biometric data of fingers of enrolled subjects, and attempts to map the identifiers of biometric data obtained from the finger at steps 60 - 63 to such security identification information (step 66 ).
- a score is calculated for each attempted mapping (step 67 ) and when one of the mapping store exceeds a threshold level identification may be considered as being confirmed.
- Use of additional biometric data identifier than a finger print for a small area subcutaneous biometric image increases the probability for true acceptance and true rejection.
- processor 13 or computer system 28 ) detects that one or more of the proof of life parameters is outside their respective acceptable threshold values(s) stored in memory 14 , the identification process ends and the operator of computer system 28 notified.
- the fingerprint, and/or one or more of the three-dimensional ultrasound images of vasculature and bone structure may be stored in memory 14 , and/or sent to computer system 28 for storage in its memory, Further, all or part of the processing of image(s) by processor 13 to provide biometric identifiers may be performed by computer system 28 upon such image(s) if so provided to system 28 , which like processor 13 operates in accordance with a program or software in its memory enabling such operations.
- steps 60 - 65 are also performed, and the biometric data from such steps is sent to computer system 29 for storage in a database of security information of computer system 28 along with other inputted identification information related to the subject, e.g., name, facial picture, department, etc., for future use in biometric identification in a manner typical of fingerprint identification systems. If the processor 13 (or computer system 28 ) detects that one or more of the proof of life parameters is outside their respective acceptable threshold values(s) stored in memory 14 , the enrollment process ends and the operator of computer system 28 notified.
- the identification device 9 may provide other imaging or vital parameter detection.
- a very large aperture 40 unfocused beam transmit and received channels are not time shifted
- a wavelet time frequency pattern
- This wavelet is then utilized to identify areas of pulsation associated with arterial blood flow supporting biometric identification by providing temporal filtering.
- parallel overlapping transmit and receiving beams, and non-overlapping parallel transmit and receive beams, rather than orthogonal as described above, may be used, such as useful for detecting and monitoring flow of correlation in three dimensions.
- a search for subcutaneous features using a variable aperture may be used, where areas of subcutaneous biometric is first coarsely scanned using wider beams; only identified areas by processor 13 are scanned using high resolution scanning of smaller scan apertures, such as described above in connection with FIGS. 5 to 8 .
- Identified areas may be identified by have pixel values (or spatial distributions) above threshold value(s) indicative of possible object detection.
- One or multiple ultrasound three dimensional images described herein may be analyzed using any common ultrasound analysis to provide additional biometric or medical information.
- application of biomedical ultrasound to the fingertip may be used for extracting anatomical, morphological and physiological properties of tissue; each one can increases the number of biometrics used for personal identification and proof of life.
- Ultrasound images provided from sensor 10 although described for identification may be used for medical applications in a manner as typical of ultrasound images.
- FIG. 12 is a simplified illustration of sensor control hardware and software.
- a central processing unit (CPU) 13 also referred to herein as the processor 13
- the processor 13 is electronically attached to a system bus 70 , also referred to herein as the bus 70 .
- Memory 14 , a signal generator 12 , a controller 16 and a signal processor 76 are also electronically attached to the bus and addressable by the processor 13 , via the bus 70 .
- the memory 14 represents memory implemented as one or more memory components that are addressable from the processor 13 , via the bus 70 .
- the processor 13 can address other memory components that are not necessarily electrically attached to the bus 70 , and are addressable via means other than the bus 70 .
- Virtual memory 72 represents processor addressable and accessible memory, whether implemented as memory 14 or as other non-bus attached memory.
- the virtual address space 74 stores digital logic expressed as CPU instructions and processor addressable data.
- Sensor control software 74 is stored within the virtual memory 72 , and is configured to control transmission of signals, and configured to control reception of signal from, the sensor array 10 via the processor 13 , the controller 16 , the signal generator 12 and the signal processor 76 .
- the controller 16 interfaces with multiplexors (“muxes”), like the multiplexors 15 a - 15 b shown in FIG. 1A . Because the processor also interfaces with the controller 16 via the bus 70 , the sensor control software 74 via the processor 13 , also exercises control of the multiplexors 15 a - 1 5 b , via the controller 16 .
- multiplexors like the multiplexors 15 a - 15 b shown in FIG. 1A .
- the controller 16 interfaces with non-multiplexor based hardware, to transmit and receive signals to and from the sensor array 10 . Because the processor also interfaces with the controller 16 via the bus 70 , the sensor control software 74 via the processor 13 , exercises control of the non-multiplexor based hardware via the controller 16 .
- the sensor control software 74 is configured to operable in a first mode for obtaining a first set of data encoding at least one two dimensional image of a fingerprint of a finger.
- the software 74 is also configured to be operable in a second mode for obtaining a second set of data encoding at least one three-dimensional representation of one or more subcutaneous tissue structures that are located within tissue that is embedded within a finger.
- biometric information such as biomarkers
- biomarkers include a nearest three dimensional coordinate of a vascular structure, or a bone structure, relative to one selected fingerprint minutia location.
- the relative location between these biomarkers are represented by three dimensional Cartesian coordinates.
- other metrics such as those employing angles and distances, are employed to quantify a relative location between biomarkers within a fingerprint, within subcutaneous tissue and/or between a fingerprint and subcutaneous tissue.
- location coordinates of points along an outer surface and/or a center point along an intersecting plane to the vascular or bone structure are determined and recorded as a biometric marker.
- a second, third and possibly a fourth mapping of one or more biometric markers over time to identify dynamic properties of portions of subcutaneous tissue.
- locations of biomarkers that change over time can be recorded and analyzed to determine a pattern of motion indicative of a presence and/or frequency of a heart beat and to optionally determine an amount of blood flow or a pulse wave pattern through the vascular structure, Such analysis can also determine elastic properties, such as an expansion and contraction measurement of the vascular structure.
- a static representation of less dynamic, and relatively static biometric markers within the finger print and subcutaneous tissue are measured and combined to represent an overall static biometric characteristic of a person, for which is employed for later comparison with biometric information later obtained from an unidentified person, to perform biometric matching.
- biometric matching involves computation of a matching score. If such matching score equals or exceeds a minimum score value, then an identity match has occurred and as a result, it is highly likely that a person currently having an un-proven identity, is a person from which biometric data has been previously obtained from and registered and later matched in association with the system of the invention.
- FIG. 13 illustrates an alternative hardware for transmitting and receiving signals to and from the sensor array 10 .
- a schematic diagram of a top down view of a piezoelectric identification device, according to an alternative embodiment of the present invention is shown.
- multiplexors 15 a - 15 b are replaced with non-multiplexor based electronic hardware components 85 a - 85 b , respectively.
- the component 85 a includes CMOS drivers and is configured for facilitating transmission of signals from the signal generator 12 to the elements 11 of the sensor array 10 .
- Use of multiplexors adds significant and unwanted capacitance, which degrades use of the sensor array 10 when generating ultrasound acoustic energy from the sensor array 10 .
- the non-multiplexor based electronic hardware 85 a instead employs CMOS drivers for periodically switching the transmission of signals to the sensor array 10 , instead to a ground potential, when the component 85 b , is receiving signals from the sensor array 10 .
- the non-multiplexor based electronic hardware 85 b instead employs pre-amplifiers for receiving signals and periodically switching the reception of signals from the sensor array 10 , to a ground potential, when the component 85 a is transmitting signals to the sensor array 10 .
- the receiving (Rx) lines 20 are clamped to ground during signal transmission over the (Tx) lines 19
- the transmitting (Tx) lines 19 are clamped to ground while receiving signals over the (Rx) lines 20 .
- This allows for a ground potential clamping multiplexor (mux) on low impedance receiving (Rx) lines during the signal transmission (Tx) sequence and for controlling the transmission (Tx) driver to clamp the transmission (Tx) tines during the signal receiving sequence.
- clamping multiplexor (mux)
- these components 85 a - 85 b are substantially implemented from non-multiplexor electronic hardware components, and as a result, are referred to herein as non-multiplexor based hardware.
- H-bridge transmission drivers can be employed, by changing the receive (Rx) clamping multiplexor (mux) to an inverse polarity driven transmission (Tx) driver.
- the second transmission (Tx) driver on the (Rx) lines would be placed into a tri-state during signal (Rx) reception, while the opposite transmission (Tx) driver would clamp to ground potential.
Landscapes
- Engineering & Computer Science (AREA)
- Physics & Mathematics (AREA)
- General Physics & Mathematics (AREA)
- Multimedia (AREA)
- Theoretical Computer Science (AREA)
- Human Computer Interaction (AREA)
- Computer Vision & Pattern Recognition (AREA)
- Measurement Of The Respiration, Hearing Ability, Form, And Blood Characteristics Of Living Organisms (AREA)
- Ultra Sonic Daignosis Equipment (AREA)
- Image Input (AREA)
Abstract
A system, apparatus and method for obtaining biometric data from characteristics of a fingerprint and obtaining characteristics of subcutaneous structures that are embedded within finger tissue and located in relation to the fingerprint.
Description
- The present application is a continuation of U.S. patent application Ser. No. 16/245,299, filed Jan. 11, 2019, which is a continuation of U.S. patent application Ser. No. 15/470,465 filed Mar. 27, 2017, which is a continuation of U.S. patent application Ser. No. 14/174,761 filed Feb. 6, 2014 (now U.S. Pat. No. 9,607,206 issued Mar. 28, 2017), which claims benefit of U.S.
Provisional Patent Application 61/761,665 filed Feb. 6, 2013, all of which are incorporated herein by reference in their entirety. - Fingerprinting is one of the most widely used biometric for human identification. Identification is obtained by analyzing a given fingerprint image obtained by a fingerprint sensor for the relative locations and orientations of structural elements such as branching or ending of ridges and valleys known as minutia. These characteristics are obtained in the enrollment mode of a person's finger or multiple fingers. In the verification mode a second fingerprint is obtained and analyzed for similarity based on minutia or other previously defined fingerprint characteristics. This minutia is also referred to herein as a type of biometric marker.
- The probability for false identification either a false acceptance or false rejection depends on the number of minutia identified in the fingerprint. The number of minutia increases with the fingertip area being scanned. However, for integration of fingerprint sensors into mobile devices for access control, such as cell phone a small area fingerprint sensor is very desirable.
- Sonavation, Inc. of Palm Beach Gardens, Fla., USA manufactures biometric sensing devices having a ceramic Micro-Electro Mechanical System (MEMS) piezoelectric array that is made from a ceramic composite material. When this piezoelectric material is formed into a
pillar 1/10th the diameter of a human hair, it has a unique set of properties that enable it to mechanically oscillate when an electric field is applied or create an electrical voltage when mechanically vibrated. The piezoelectric pillar is electrically vibrated at its natural ultrasonic resonant frequency. If a fingerprint ridge is directly above the pillar, much of the ultrasonic energy is absorbed by the skin and the signal impedance of the pillar is very high. If a valley is directly above the pillar, very little energy is absorbed and the impedance is very low. By arranging the pillars in a matrix of several thousand elements a two-dimensional image of a fingerprint can be created. An imaging ASIC electrically controls the pillar oscillation, imaging of the fingerprint and data management of the fingerprint information. - U.S. Pat. No. 7,141,918 describes an biometric sensing device having the above piezoelectric array operable for fingerprint imaging. It has been found as also described in this patent that the piezoelectric array can be operated in non-fingerprint imaging modes to obtain other biometric information, such as in an echo mode to provide imaging, such as bone, or a Doppler-shift mode to detect blood flow velocity and blood flow patterns. Although the sensor described in this patent is useful, it would be desirable to also operate the sensing device in a three-dimension ultrasound imaging mode to provide improved imaging of subcutaneous structures for use in biometric identification (or medical applications) that does not rely on echo mode imaging as described in U.S. Pat. No. 7,141,918.
- The discussion above is merely provided for general background information and is not intended to be used as an aid in determining the scope of the claimed subject matter.
- Accordingly, it is an object of the present invention to provide a biometric sensing device having a piezoelectric array providing improved three-dimension imaging of subcutaneous tissue structures of a finger, such as bone or vasculature, utilizing pitch/catch ultrasonically formed images.
- It is another object of the present invention to provide a biometric sensing device having a piezoelectric sensor array providing improved three-dimension images of subcutaneous tissue structures of a finger where such images are useful for further providing proof of life parameters.
- Briefly described, the present invention embodies a biometric sensing device having an array of piezoelectric ceramic elements operable in a first mode for producing first data representative of a fingerprint image, and a second mode for producing second data representative of least one three-dimensional image of subcutaneous tissue structure(s), such as or bone or vascular, formed by pitch-n-catch ultrasound imaging. The images provided from operating the sensing device in the first and second modes provide anatomical and morphological biometrics (biometric data) for use in biometric identification.
- The second data representative of least one three-dimensional image of subcutaneous tissue structure, may also be used for determining elastic properties of tissue, and vital or proof of life parameters, i.e. physiological information, such as heart beat, blood flow velocities, and pulse wave pattern, or other parameters which can be used to determine if the finger disposed upon the sensor array is fake or dead.
- The elastic properties of tissue which may, like captured fingerprint image and the one or more images of subcutaneous tissue structure(s), provide biometrics (biometric data) for use in biometric identification. Thus, multiple types of biometric data can obtained with a single application of a finger to the sensor array, which can be done in real time and simultaneously.
- The architecture of the identification device is similar to what is described in U.S. Pat. No. 7,141,918, also referred to herein as the '918 patent. Embodiments of the subject invention include various improvements over the '918 patent that are described herein. These improvements include those relating to electronic control and data acquisition. U.S. Pat. No. 7,141,918 is incorporated herein by reference. Further, U.S. Pat. Nos. 7,844,660, and 6,720,712, which are related to U.S. Pat. No. 7,141,918 are also incorporated herein by reference.
- This brief description of the invention is intended only to provide a brief overview of subject matter disclosed herein according to one or more illustrative embodiments, and does not serve as a guide to interpreting the claims or to define or limit the scope of the invention, which is defined only by the appended claims.
- So that the manner in which the features of the invention can be understood, a detailed description of the invention may be had by reference to certain embodiments, some of which are illustrated in the accompanying drawings. It is to be noted, however, that the drawings illustrate only certain embodiments of this invention and are therefore not to be considered limiting of its scope, for the scope of the invention can encompass other equally effective embodiments.
- The drawings are not necessarily to scale. The emphasis of the drawings is generally being placed upon illustrating the features of certain embodiments of the invention. In the drawings, like numerals are used to indicate like parts throughout the various views. Differences between like parts may cause those parts to be indicated with different numerals. Unlike parts are indicated with different numerals. Thus, for further understanding of the invention, reference can be made to the following detailed description, read in connection with the drawings in which:
-
FIG. 1A is a schematic diagram of a top down view of a piezoelectric identification device according to the present invention;FIG. 1B is a side perspective view of the piezoelectric identification device in which pillars (elements) are evenly spaced and are separated by filler material. -
FIG. 2 a schematic illustration of the sensor array addressing scheme is shown, where there are M and N number ofelements 11, disposed in M rows along an x axis and in N columns along a y axis. -
FIG. 3 illustrates an identification device that is coupled to a computer system. -
FIG. 4 illustrates a cross-section ofsensor array 10 and of afinger 30 placed proximate to an upper surface of the sensor array so that the finger'ssurface 31 is in direct contact with itsprotective shield 23. -
FIG. 5 illustrates ascan aperture 40 that is scanned byprocessor 13 with respect to an x axis (M rows) and y axis (N columns) dimensions of the scan aperture, like shown inFIG. 2 . -
FIG. 6A illustrates atransit aperture 41 that is selected byprocessor 13 to form a transmit beam or pulse 46 (shown as a translucent volume) having an hour-glass shape, having 6 transmit channels.FIG. 6B illustrates transmit signals traveling via channels A-F. -
FIG. 7A illustrates a receiveaperture 42 selected byprocessor 13 to receive acoustic energy (beam or pulse 49).FIG. 7B illustrates transmit signals received via channels A-F. -
FIG. 8 illustrates a combinedoutput signal 54. -
FIG. 9 illustrates ultrasound scanning of ablood vessel 50 within a finger via a sensor array. -
FIG. 10 illustrates ultrasound scanning of a bone structure within a finger via a sensor array. -
FIG. 11 illustrates operation of theidentification device 9 ofFIG. 3 while scanning a finger via a sensor array, like shown inFIGS. 4, 9 and 10 . -
FIG. 12 is a simplified illustration of sensor control hardware and software. -
FIG. 13 illustrates an alternative hardware for transmitting and receiving signals to and from the sensor array. - Referring to
FIG. 1A , a schematic diagram of a thepiezoelectric identification device 9 according to the present invention is shown.Identification device 9 has asensor array 10 of piezo-electric elements 11 providing mechanical resonator sensing, a sensorinput signal generator 12, and aprocessor 13. Under control ofprocessor 13, the input signal generated byinput signal generator 12 is coupled tosensor array 10 by asignal multiplexer 15 a, and output signal ofsensor 10 is coupled toprocessor 13 by amultiplexer 15 b. -
Processor 13 processes the output signals from select element(s) viamultiplexor 15 b to obtain biometric data which may then be stored in amemory 14. Biometric data can include one or more fingerprint images, and/or one or more ultrasound images of subcutaneous structures of the finger, subcutaneous tissue parameter(s) such as of tissue elasticity, and/or detected proof of life parameters, as described later below. Addressing ofelements 11, viamultiplexers mux controller 16 in accordance with user specified imaging modes and/or in detection of proof of life parameters. Although each multiplexor 15 a and 156 is shown singularly, each multiplexor's function may alternatively be designed to be provided by two or more multiplexors as desired. -
Sensor array elements 11 may be of lead zirconate titanate (PZT) or other material having similar properties, preferably, PZT 1-3 composite. The piezo-ceramic elements 11 can have shapes other than rectangular, such as circular as shown inFIG. 2 .Sensor area 10 is preferably provided on a support layer, such as described in the above incorporated patent. Other ASIC chips may provide the electronics shown inFIG. 1A to which the sensor is connected. - A more detailed view of
sensor array 10 is shown inFIG. 1B in whichelements 11 represent evenly spaced pillars withfiller material 17 there between.Filler material 17 may be an epoxy or polymer having micro-spheres. Transmit electrodes (or lines) 19 and receive electrodes (or lines) 20 (SeeFIG. 1A ) are provided above and below, respectively, alongsensor array 10. Over the transmitelectrodes 19 is a shield orprotective coating layer 22, such as urethane providing asurface 23 upon which a fingertip may be placed. Below the receiveelectrodes 20 is asupport substrate 24, such as of foam material. - Referring to
FIG. 2 , a schematic illustration of the addressing scheme is shown, where there are M by N number ofelements 11, disposed in M rows along a horizontal x axis as shown and in N columns along a vertical y axis as shown. For example; M and N may equal 368 and 256, respectively, but another number of elements, and row and column grouping thereof, can be employed in accordance with other embodiments of the invention. Transmitelectrodes 19 are parallel conductors that connect to the top ofelements 11 inrows 1 to M, while receiveelectrodes 20 are parallel conductors that connect to bottom ofelements 11 incolumns 1 to N. In accordance with some embodiments, eachelement 11 is 40 microns square by 100 microns deep, thereby yielding adense sensor array 10 having a 20 MHz fundamental frequencysonic wave elements 11. A spacing of 10 microns is used betweenelements 11 and is preferred in order to provide a 50-micron total pitch between elements. Other geometries may be used, such as for example, a pitch of greater or lower than 50 microns. For example, a sensor with 368 rows and 256 columns may be 18.7 by 13 mm, or other size according to the maximum imaging size desired for the particular application. - In some embodiments,
sensor array 10 may be manufactured as described in U.S. Pat. No. 7,489,066, which is herein incorporated by reference. By arrangement of the elements in an array of rows and columns,elements 11 are individually addressable for application of an input signal by row, and then addressable for reading out an output signal by column, by selection ofelectrodes - A
ground switch 26 is provided coupled to all transmitelectrodes 19 betweenedge connector 19 a andmultiplexor 15 a enablingprocessor 13 to connectelectrodes 19 to ground when needed. Similarly, aground switch 27 is provided coupled to all receiveelectrodes 20 betweenedge connector 20 a andmultiplexor 15b enabling processor 13 enablingprocessor 13 to connectelectrodes 20 to ground when needed. The benefit of ground switches 26 and 27 is that it avoids additional switching of ground and signal electrodes as described in U.S. Pat. No. 7,141,918, thereby avoiding unwanted additional capacitive loads parallel to the transmitting and receivingelements 11. - As will be described below,
processor 13 is programmed within its embedded memory (or memory 14) to enable all sensing operations bysensor array 10 as described herein, including at least fingerprint imaging, and three-dimensional ultrasound imaging. Further,processor 13 may provide other activities commonly implemented in an ultrasonic imaging system as part of electronic beam formation including synthetic aperture imaging. - Referring to
FIG. 3 ,identification device 9 is coupled to acomputer system 28 for data communication to send commands and receive biometric data fromidentification device 9.Computer system 28 may be any microprocessor-based device operating in accordance with a program or instructions which utilizesidentification device 9 to obtain biometric data viasensor array 10 as needed for identification (e.g., verification and/or enrollment) of biometric data.Such computer system 28 uses biometric data connected from the sensor to enable biometric feature or identifier matching in accordance with a database or memory withinsystem 28, or accessible tosystem 28. For example,computer system 28 may be part of portable identification device, point of transaction system, or any other device requiring secure biometric sensing to enable access to physical and/or electronic resources. -
Identification device 9 is operable in a fingerprint imaging mode, and a three-dimensional subcutaneous tissue structure imaging mode, as described below. - Fingerprint Imaging Mode.
-
FIG. 4 illustrates a cross-section ofsensor array 10 of thedevice 9, and afinger 30 placed proximate to the sensor array so that the finger'ssurface 31 is in direct contact with itsprotective shield 23. The finger'sSuch surface 31 hasridges 32, which make such direct contact, andvalleys 33 which do not make such direct contact with theprotective shield 23. When anelement 11 is driven at a resonant frequency by theinput signal generator 12, via one of transmitelectrodes 19 selected bymux controller 16 viamultiplexor 15 a, its oscillation is directed or transmitted towards the finger'ssurface 31 and produces a sonic wave either towardsridges 32 orvalleys 33 as graphically illustrated bysonic waves elements 11 within thesensor array 10 and causes these elements to oscillate, preferably at a resonant frequency. Phenomena outside of thesensor array 10, such as a presence of acoustic impedance or a reflection of acoustic energy off of outside entities (external to the sensor), further influences the motion of these elements 11 (apart from the influence of input signal excitation) within thesensor array 10. Such outside influenced motion is monitored, measured and analyzed by electronic hardware in combination with software that controls operation of the electronic hardware. (SeeFIG. 12 ) - The
surface 31 response to sonic wave differs due to contact to tissue of a ridge versus non-contact of valley difference in impedance (or attenuation/voltage) which is detectable by thesame element 11 which transmitted the sonic waves or beam, via one of receiveelectrodes 20 selected bymux controller 16 viamultiplexor 15 b, thereby providing a measure as to whether the element is facing a ridge or a valley. Theprocessor 12 builds a map inmemory 14 where each element response (output signal) detected byprocessor 12 represents one pixel of the two-dimensional fingerprint image inmemory 14, where each pixel may be represented as a black or white value to represent a ridge or valley, respectively, or vice versa. Thus, read out in which of impedance measured is converted into a fingerprint image of ridges and valleys. - Such operation of
identification device 9 to obtain a fingerprint image is described in more detail in connection with FIGS. 17-22 of in incorporated U.S. Pat. No. 7,141,918 which is included in Appendix A of the prior filed provisional patent application, or other U.S. Pat. Nos. 7,489,066, 7,514,842, 8,331,633, and 8,335,356 which are also all herein incorporated by reference. - Preferably,
sensor array 10 operates to obtain a fingerprint by detecting the impedance at a resonant frequency of an applied input signal generated 12, viamultiplexor 15 a, where upon soon after a driving input signal of eachelement 11 ceases in time, and an output signal is read from that same pixel. That output signal is representative of impedance. In other words, theelement 11 ring (vibration) characteristic causes an electrical output signal to be produced by the element that when sampled, viamultiplexor 15 b, provides a measure of impedance. Further, two impedance measurements can take place at two different frequencies (e.g., 19.8 MHz and 20.2. MHz) for eachelement 11, where the difference of measured impedance at each frequency is used to determine whether theelement 11 is facing and adjacent to a ridge or a valley as described in incorporated U.S. Pat. No. 7,141,918. - Ultrasound (Pitch/Catch) Three Dimensional Imaging Mode.
-
Identification device 9 can also operatesensor 10 in a pitch/catch imaging mode to obtain three-dimensional ultrasound images within a finger presented tosensor array 10. Thus, a sensor principally described for fingerprint image capture can enable viewing of structures within the same tissue that provided a fingerprint image, such as vascular structures (venous and arterial vessels), or bone structure. As described in more detail below,processor 13 operates theelements 11 of thesensor array 10 in this pitch/catch mode by connecting the transmitter and receiver in series, rather than in parallel as in echo imaging of the prior incorporated U.S. Pat. No. 7,141,918. - As illustrated in
FIG. 5 , anscan aperture 40 is scanned byprocessor 13 along in x and y axes or dimensions, where the x dimension corresponds to the 1 to M rows, and y dimension corresponds to the 1 to N columns ofFIG. 2 . Thescan aperture 40 is formed along the intersection of a group of “m” number of adjacent rows ofelements 11 activated byprocessor 13 to provide a transmitaperture 41 that produces a transmit beam, and a group of “n” number of adjacent columns ofelements 11 activated byprocessor 13 to provide a receiveaperture 42 that receives a receive beam, where the time delay of signals received is representative of structure(s) in the volume of tissue into which the transmit beam is focused and the receive beam is received abovescan aperture 40. - During scanning,
processor 13 moves thescan aperture 40 along the x and y dimensions by selecting different groups of “m” rows and “n” columns in which to overlap and formdifferent scan apertures 40. For beam focusing, the transmitelectrodes 19 to the “m” rows ofelements 11 are divided equally into “p” number channels, where the number of transmit channels equals “m” divided by “p”. Similarly, the receiveelectrodes 20 the “n” columns ofelements 11 are divided equally into “r” number of receive channels, where the number of receive channels equals “n” divided by “r”. An example for one of themultiple scan apertures 40 that may take place during scanning of vasculature within the tissue abovesensor array 10 during scanning of multiple different scan apertures is shown inFIGS. 6A and 6B , whereFIG. 6A represents a transmit cycle andFIG. 6B represents a receive cycle. - In
FIG. 6A a transmitaperture 41 is selected byprocessor 13 to form a transmit beam or pulse 46 (shown as a translucent volume) having an hour-glass shape. In this example, “m” equals 12 and “p” equals 6, resulting in 6 transmit channels A-F each with two rows. In order to focus the beam along avolume 48 of the transmitbeam 46 at a distance below the tissue surface, the transmit (input) signal is applied by transmit channels A-F and offset in time in accordance with distance of rows from a location involume 48. The transmit signal is first applied to rows of the most outer transmit channels A and F first, then the second most outer transmit channels B and E, and then the central most transmit channels C and D last, as illustrated inFIG. 6B . - Thus the transmit
aperture 41 forms a transmitbeam 46 which will arrive at approximately the same time thereby focusing transmitbeam 46 at locations in the intendedvolume 48 that may contain the object or structure of interest, such as ablood vessel 50. In forming transmitbeam 46 all other rows ofelements 11 which are not used in the transmitaperture 41 are inactive. A blood vessel may or may not be fully included in the transmitbeam 46. During this transmit cycle,switch 27 is switched to ground byprocessor 13 to ground the receiveelectrodes 20, whileswitch 26 is not set to ground. - After transmit
beam 46 is launched into the tissue of thefinger 30 and an additional period for ring down of the transmitelectrodes 19 transmittingelements 11 along the “m” rows (i.e., their electrodes 19) are switched to ground byprocessor 13 viaswitch 26, and whileswitch 27 is not set to ground. The receive cycle can then begin. -
FIG. 7A shows an example of the receiveaperture 42 selected byprocessor 13 to receive beam orpulse 49 having an hour-glass shape. In this example, “n” equals 12 and “r” equals 6, resulting in 6 receive channels A-F each with two columns. In order to focus receiving beam along avolume 52 beam the tissue surface, the receive (output) signal is read (or detected) from receive channels A-F, and read offset in time in accordance with distance of columns from a location involume 52. The receive signal is first read from columns of the most outer receive channels A and F first, then the second most outer receive channels B and E, and then the central most receive channels C and D last, as illustrated inFIG. 7B . - Thus the beam received by
elements 11 of the receiveaperture 42 will arrive at approximately the same time from the intendedvolume 52, which in this example includes part ofblood vessel 50. The signals from all the receive channels A-F are aligned in accordance with the time offset of reception shown inFIG. 7B and combined by a beam former 53 inprocessor 13 to form a combined output signal forscan aperture 40 as shown inFIG. 8 . - In receiving the output signals from receive channels A-F, all other columns of
elements 11 which are not in receiveaperture 42 are inactive. Receivebeam 49 is orthogonal to the transmitbeam 46, and it is their intersection along transmitaperture 41 andreceiver aperture 42 which defines the effective pitch/catch scan aperture 40. - The
processor 13 receives signal from the “n” column ofelements 11 during the sampling interval associated with the round trip time after the ceased transmit beam is backscatter reflected towards thesensor 10, from the objects or structures desired to be imaged. The delay in time of the combined output signal from beam former 53 over the sampling interval represents distance from thesensor array 10, and the amplitude orvalue 54 of the signal at different depths along the z dimension sampled during the sampling interval is recorded byprocessor 13 inmemory 14 at x,y,z coordinates associated with thatscan aperture 40. Theprocessor 13 may receive combined output signal over the entire depth of thescan aperture 40, but records information inmemory 14 over a desired range of volume's depth of intersectingvolumes 48 and 51 ofscan aperture 40 to provide a three-dimensional ultrasound image indicating structures of interest which can be within that desired depth range fromsensor array 10. -
FIG. 8 shows an effect upon a beam formed signal by ablood vessel 50 along the z axis at a distance fromsensor 10 centered along thescan aperture 40, where +/−amplitude value 54 of coordinate along the x axis at a common y position centered about the receiveaperture 42. (The amplitude of the signal between the outer boundaries of the vessels diameter may also be processed byprocessor 13 to provide velocity information of the flow there through at that x, y coordinate for use as a vital parameter). In other words, this transmitbeam 46 is steered in two orthogonal axes x and y within a two-dimensional (2D) scan of thesensor array 11. The 3rd axis, defined as the axial or z-axis is obtained by time of arrival analysis of subcutaneous tissue causing backscattered sound waves byprocessor 13. - The processor adds the information at sampled points of
amplitude 54 obtained along the z axis fromsensor 10 at the x,y coordinate to a map inmemory 14 along the x and y dimensions thereby building a three-dimensional ultrasound image of subcutaneous structures. A full 20 x,y image along an x,z plane is obtained from time history in z and receiveaperture 42 position in y. In other words, this 20 image provides a slice along the x,z plane of the full 3D volume presentation of backscattered ultrasound for a givenscan aperture 40. Scanning along the x axis while scanning the receive aperture for each new position creates the full volume representation of the fingertip object. During this receive cycle,switch 26 is switched to ground byprocessor 13 to ground the transmitelectrodes 19, whileswitch 27 is not set to ground. - The
processor 13 then repeats the process fordifferent scan apertures 40 along the x any y dimensions over the volume of tissue abovesensor array 10 providing multiple slices along x,z planes of scan apertures to complete a three-dimensional ultrasound image of subcutaneous structures. - Three-dimensional beam forming for ultrasonic imaging is described in C. E. Demore et al., Real Time Volume Imaging Using a Crossed Electrode Array, IEEE UFFC Trans vol. 56 (6) 1252-1261, but heretofore has not been provided by a sensor array of piezoelectric elements.
- As describe above, there is grounding of transmit
electrodes 19 and receiveelectrodes 20 alternating with receive and transmit cycles for eachscan aperture 40. As vascular structures and bone structures are at different depths in the tissue with respect tosensor 10, the sampling interval for the subcutaneous tissue may be set to provide three-dimensional ultrasound image of the vasculature offinger 30 as illustrated inFIG. 9 , or bone structure offinger 30 as illustrated inFIG. 10 , thereby enabling three-dimensional imaging of different types of subcutaneous structures. Other structures in the tissue of the finger may similarly be imaged as desired. - Unlike in fingerprint mode where only one transmitting
element 11 is used at a time, in the ultrasound pitch and catch mode a subgroup of “n” adjacent transmitters (transmitting elements 11) is active providing an electronically focusedbeam 46 in one lateral direction commonly referred to as azimuth axis. In the orthogonal direction, commonly referred to as the elevation direction, the receiveaperture 49 is selected as a sub-group of “m”electrodes 20 via themultiplexer 15 b, thus the effective aperture for transmit and receive becomes the spatial intersection between transmit and receiveapertures electrodes 20 are connected via amultiplexer 15 b to a group of “m” receive amplifiers and signal processing chains for beam formation and further backscatter analysis byprocessor 13. - In the fingerprint mode all available M receive channels utilized in parallel providing maximum speed for data acquisition. All electrodes are connected to a programmable signal from
processor 13 to ground switches 26 and 27. Thus in the ultrasound imaging mode the receiveelectrodes 19 are grounded during the transmission cycle or phase, but switched off from ground (unwounded) during the receive phase during which all transmittingelements 11 are grounded. - By analyzing changes in two or more ultrasound images at a x,y,z coordinate(s) in a blood vessel, proof of life parameter(s) are detectable, such as velocity or flow of cells through the vessel, heartbeat, or flow patterns, as desired, in a manner as commonly detected in typical ultrasound imaging system.
- Referring to
FIG. 11 , the operation of anidentification device 9 will now be described for identification (or verification) or a subject'sfinger 30 as presented to sensor array as shown inFIGS. 4, 9, and 10 . First, in fingerprint imagemode sensor array 10 is operatedprocessor 13 to capture an image (two dimensional representation) of a fingerprint along the surface of finger 30 (step 60), as described above, which is stored inmemory 14 as minutia (biometric fingerprint identifiers) in relative and local sensor x,y coordinates (step 64). Optionally, or in addition, the fingerprint image may be stored inmemory 14, and/or sent tocomputer system 28. - Next,
identification device 9 is switched to three-dimensional ultrasound/volumetric imaging mode. An image of subcutaneous fingertip vascular structure offinger 30 is then captured in memory 14 (step 61), and processed byprocessor 13 to obtain biometric data of identifiers uniquely characterizing curvature and/or shape of all or major subcutaneous vascular structure of the finger in relative and local sensor x,y,z coordinates (step 65). Other tissue characteristics from image may also provide biometric identifiers, such as tissue speckle. Optionally, or in addition, the three-dimensional ultrasound image may be stored inmemory 14, and/or sent tocomputer system 28. - At
step 62, subcutaneous tissue parameters are measured from the ultrasound image stored inmemory 14. The ultrasound image may be processed byprocessor 13 to determine elastic properties of tissue by applying pressure to the fingertip and estimating the strain in the tissue using typical ultrasound elastography. Reversely, with known tissue elasticity applied pressure is estimated from tissue strain. The elastic measure represents another biometric identifier stored inmemory 14. - The
processor 13 using the three-dimensional ultrasound image fromstep 61 stored inmemory 14 determines one or more vital parameters which may be used to reduce the risk that the subject's finger in fake or dead, such as blood flow, vessel wall pulse waves and heart rate parameters. Each of the one or more vital parameters are compared with one or more thresholds stored in memory 14 (or bycomputer system 28 if sent thereto) which if not met indicates that the subject'sfinger 30 may be fake or dead. Blood flow may be identified using common procedure of ultrasonic flow detection, such as described in J. A. Jensen, Estimation of Blood Flow using Ultrasound, Cambridge University Press, 1996, or R. S. C. Cobbold, Foundations of Biomedical Ultrasound, Oxford University Press, 2007. In addition to identifying blood flow, blood mean velocity or maximum velocities as well as flow spectra are obtained. Heart rate and vessel wall motion is detected from lower frequency variations of pulsed and continuous wave ultrasound. - An image of subcutaneous fingertip bone structure is then captured and stored in memory 14 (step 63), and processed by
processor 13 to obtain biometric data of identifiers uniquely identifying subcutaneous bone structure of the finger in relative and local sensor x,y,z coordinates (step 65). Finger bone structure is useful as biometric, particularly if bone curvature or other bone shape identifiers. - The identifiers of biometric data from finger print, vascular image, bone structure image, and elastic parameter, and provided along with determine proof of life parameters to
computer system 28 atstep 66.Computer system 28 stores a database of security identification information of previously captured identifiers of biometric data of fingers of enrolled subjects, and attempts to map the identifiers of biometric data obtained from the finger at steps 60-63 to such security identification information (step 66). A score is calculated for each attempted mapping (step 67) and when one of the mapping store exceeds a threshold level identification may be considered as being confirmed. Use of additional biometric data identifier than a finger print for a small area subcutaneous biometric image increases the probability for true acceptance and true rejection. - If the processor 13 (or computer system 28) detects that one or more of the proof of life parameters is outside their respective acceptable threshold values(s) stored in
memory 14, the identification process ends and the operator ofcomputer system 28 notified. - Optionally, or in addition, the fingerprint, and/or one or more of the three-dimensional ultrasound images of vasculature and bone structure may be stored in
memory 14, and/or sent tocomputer system 28 for storage in its memory, Further, all or part of the processing of image(s) byprocessor 13 to provide biometric identifiers may be performed bycomputer system 28 upon such image(s) if so provided tosystem 28, which likeprocessor 13 operates in accordance with a program or software in its memory enabling such operations. - To enroll a subject rather than for verification, steps 60-65 are also performed, and the biometric data from such steps is sent to computer system 29 for storage in a database of security information of
computer system 28 along with other inputted identification information related to the subject, e.g., name, facial picture, department, etc., for future use in biometric identification in a manner typical of fingerprint identification systems. If the processor 13 (or computer system 28) detects that one or more of the proof of life parameters is outside their respective acceptable threshold values(s) stored inmemory 14, the enrollment process ends and the operator ofcomputer system 28 notified. - The
identification device 9 may provide other imaging or vital parameter detection. For example, a verylarge aperture 40 unfocused beam (transmit and received channels are not time shifted) may be utilized for detecting heartbeat. From the heart beat a wavelet (time frequency pattern) may be constructed byprocessor 13. This wavelet is then utilized to identify areas of pulsation associated with arterial blood flow supporting biometric identification by providing temporal filtering. Further, parallel overlapping transmit and receiving beams, and non-overlapping parallel transmit and receive beams, rather than orthogonal as described above, may be used, such as useful for detecting and monitoring flow of correlation in three dimensions. - Although the
scan aperture 40 is described as being fixed in size along x and y dimensions, a search for subcutaneous features using a variable aperture may be used, where areas of subcutaneous biometric is first coarsely scanned using wider beams; only identified areas byprocessor 13 are scanned using high resolution scanning of smaller scan apertures, such as described above in connection withFIGS. 5 to 8 . Identified areas may be identified by have pixel values (or spatial distributions) above threshold value(s) indicative of possible object detection. - One or multiple ultrasound three dimensional images described herein may be analyzed using any common ultrasound analysis to provide additional biometric or medical information. Thus, application of biomedical ultrasound to the fingertip may be used for extracting anatomical, morphological and physiological properties of tissue; each one can increases the number of biometrics used for personal identification and proof of life. Ultrasound images provided from
sensor 10 although described for identification may be used for medical applications in a manner as typical of ultrasound images. -
FIG. 12 is a simplified illustration of sensor control hardware and software. As shown, a central processing unit (CPU) 13, also referred to herein as theprocessor 13, is electronically attached to asystem bus 70, also referred to herein as thebus 70.Memory 14, asignal generator 12, acontroller 16 and asignal processor 76 are also electronically attached to the bus and addressable by theprocessor 13, via thebus 70. Thememory 14, represents memory implemented as one or more memory components that are addressable from theprocessor 13, via thebus 70. Preferably and in some embodiments, theprocessor 13 can address other memory components that are not necessarily electrically attached to thebus 70, and are addressable via means other than thebus 70. -
Virtual memory 72, represents processor addressable and accessible memory, whether implemented asmemory 14 or as other non-bus attached memory. Thevirtual address space 74 stores digital logic expressed as CPU instructions and processor addressable data.Sensor control software 74, is stored within thevirtual memory 72, and is configured to control transmission of signals, and configured to control reception of signal from, thesensor array 10 via theprocessor 13, thecontroller 16, thesignal generator 12 and thesignal processor 76. - In some embodiments, the
controller 16 interfaces with multiplexors (“muxes”), like the multiplexors 15 a-15 b shown inFIG. 1A . Because the processor also interfaces with thecontroller 16 via thebus 70, thesensor control software 74 via theprocessor 13, also exercises control of the multiplexors 15 a-1 5 b, via thecontroller 16. - In other embodiments, as shown in
FIG. 13 ; thecontroller 16 interfaces with non-multiplexor based hardware, to transmit and receive signals to and from thesensor array 10. Because the processor also interfaces with thecontroller 16 via thebus 70, thesensor control software 74 via theprocessor 13, exercises control of the non-multiplexor based hardware via thecontroller 16. - The
sensor control software 74 is configured to operable in a first mode for obtaining a first set of data encoding at least one two dimensional image of a fingerprint of a finger. Thesoftware 74 is also configured to be operable in a second mode for obtaining a second set of data encoding at least one three-dimensional representation of one or more subcutaneous tissue structures that are located within tissue that is embedded within a finger. - Further, the software identifies biometric information, such as biomarkers, within both the fingerprint and subcutaneous tissue that is embedded within the finger. Besides minutia, other biomarkers include a nearest three dimensional coordinate of a vascular structure, or a bone structure, relative to one selected fingerprint minutia location. The relative location between these biomarkers are represented by three dimensional Cartesian coordinates. In other embodiments, other metrics, such as those employing angles and distances, are employed to quantify a relative location between biomarkers within a fingerprint, within subcutaneous tissue and/or between a fingerprint and subcutaneous tissue.
- With respect to vascular and bone structures, location coordinates of points along an outer surface and/or a center point along an intersecting plane to the vascular or bone structure, are determined and recorded as a biometric marker.
- In some embodiments, after an initial mapping of biomarkers within a vascular subcutaneous structure, a second, third and possibly a fourth mapping of one or more biometric markers over time, to identify dynamic properties of portions of subcutaneous tissue.
- For example, locations of biomarkers that change over time, such as those associated with the vascular structure can be recorded and analyzed to determine a pattern of motion indicative of a presence and/or frequency of a heart beat and to optionally determine an amount of blood flow or a pulse wave pattern through the vascular structure, Such analysis can also determine elastic properties, such as an expansion and contraction measurement of the vascular structure.
- Aside from measurement of dynamic properties of biometric markers within subcutaneous tissue, a static representation of less dynamic, and relatively static biometric markers within the finger print and subcutaneous tissue are measured and combined to represent an overall static biometric characteristic of a person, for which is employed for later comparison with biometric information later obtained from an unidentified person, to perform biometric matching.
- In some embodiments, biometric matching involves computation of a matching score. If such matching score equals or exceeds a minimum score value, then an identity match has occurred and as a result, it is highly likely that a person currently having an un-proven identity, is a person from which biometric data has been previously obtained from and registered and later matched in association with the system of the invention.
- Likewise, if such a matching score is less than a minimum score value, then an identity match has not occurred and as a result, it is not likely that a person currently having an non-proven identity, is a person from which biometric data has been previously obtained from and registered in association \with the system of the invention.
-
FIG. 13 illustrates an alternative hardware for transmitting and receiving signals to and from thesensor array 10. A schematic diagram of a top down view of a piezoelectric identification device, according to an alternative embodiment of the present invention is shown. - As shown, multiplexors 15 a-15 b are replaced with non-multiplexor based electronic hardware components 85 a-85 b, respectively. The
component 85 a, includes CMOS drivers and is configured for facilitating transmission of signals from thesignal generator 12 to theelements 11 of thesensor array 10. Use of multiplexors adds significant and unwanted capacitance, which degrades use of thesensor array 10 when generating ultrasound acoustic energy from thesensor array 10. - The non-multiplexor based
electronic hardware 85 a, instead employs CMOS drivers for periodically switching the transmission of signals to thesensor array 10, instead to a ground potential, when thecomponent 85 b, is receiving signals from thesensor array 10. Likewise, the non-multiplexor basedelectronic hardware 85 b, instead employs pre-amplifiers for receiving signals and periodically switching the reception of signals from thesensor array 10, to a ground potential, when thecomponent 85 a is transmitting signals to thesensor array 10. - In other words, the receiving (Rx) lines 20 are clamped to ground during signal transmission over the (Tx) lines 19, and the transmitting (Tx) lines 19 are clamped to ground while receiving signals over the (Rx) lines 20. This allows for a ground potential clamping multiplexor (mux) on low impedance receiving (Rx) lines during the signal transmission (Tx) sequence and for controlling the transmission (Tx) driver to clamp the transmission (Tx) tines during the signal receiving sequence. Hence, although such a clamping multiplexor (mux) can be employed within 85 a-85 b, these components 85 a-85 b are substantially implemented from non-multiplexor electronic hardware components, and as a result, are referred to herein as non-multiplexor based hardware.
- In other embodiments, H-bridge transmission drivers can be employed, by changing the receive (Rx) clamping multiplexor (mux) to an inverse polarity driven transmission (Tx) driver. In this type of configuration, the second transmission (Tx) driver on the (Rx) lines would be placed into a tri-state during signal (Rx) reception, while the opposite transmission (Tx) driver would clamp to ground potential.
- From the foregoing description it will be apparent that there has been provided an improved biometric sensing devices, and systems and methods using same for biometric identification. The illustrated description as a whole is to be taken as illustrative and not as limiting of the scope of the invention. Such variations, modifications and extensions, which are within the scope of the invention, will undoubtedly become apparent to those skilled in the art
- This written description uses examples to disclose the invention, including the best mode, and also to enable any person skilled in the art to practice the invention, including making and using any devices or systems and performing any incorporated methods. The patentable scope of the invention is defined by the claims, and may include other examples that occur to those skilled in the art. Such other examples are intended to be within the scope of the claims if they have structural elements that do not differ from the literal language of the claims, or if they include equivalent structural elements with insubstantial differences from the literal language of the claims.
Claims (13)
1. A sensor comprising:
an ultrasonic sensor array configured to transmit and receive acoustic signals; and
a control unit configured to operate the ultrasonic sensor array in a first mode or a second mode, wherein
in the first mode, the control unit is configured to obtain a two-dimensional representation of a surface of a finger of a person in physical contact with the sensor based on acoustic signals transmitted and received by sensor elements of the ultrasonic sensor array, and
in the second mode, the control unit is configured to obtain a three-dimensional representation of a subcutaneous structure of the finger based on acoustic signals transmitted by a first set of the sensor elements and received by a second set of the sensor elements, wherein the first and second sets are interleaved.
2. The sensor of claim 1 , wherein the first set of sensor elements is orthogonal to the second set of sensor elements in ultrasonic sensor array.
3. The sensor of claim 1 , wherein the control unit is configured to obtain the two-dimensional representation of the surface of the finger based on impedance related to the acoustical signals received by the sensor elements of the ultrasonic sensor array.
4. The sensor of claim 1 , wherein the control unit is configured to obtain the three-dimensional representation of the subcutaneous structure based on time delays of the acoustic signals received by the second set of the sensor elements.
5. The sensor of claim 1 , wherein the control unit is configured to obtain the three-dimensional representation of the subcutaneous structure based on amplitudes of the acoustic signals received by the second set of the sensor elements.
6. The sensor of claim 1 , wherein the control unit is configured to obtain the two-dimensional representation of the surface of the finger based on at least one of the acoustic signals that is transmitted and received by a same sensor element of the ultrasonic sensor array.
7. The sensor of claim 1 , wherein the subcutaneous structure is a bone structure within the finger.
8. The sensor of claim 1 , wherein the subcutaneous structure is a vascular structure within the finger.
9. The sensor of claim 1 , wherein the three-dimensional representations of the subcutaneous structure are used to determine a dynamic property of the subcutaneous structure.
10. The sensor of claim 1 , wherein the dynamic property includes an elastic property of the subcutaneous structure.
11. The sensor of claim 1 , wherein the dynamic property is used to determine liveness of the person.
12. The sensor of claim 11 , wherein the dynamic property is used to measure a heartbeat.
13. The sensor of claim 11 , wherein the dynamic property is used to measure a blood flow through a vascular structure.
Priority Applications (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
US16/385,437 US20190311174A1 (en) | 2013-02-06 | 2019-04-16 | Method and system for 2d and 3d biometric sensing using a same sensor |
Applications Claiming Priority (5)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
US201361761665P | 2013-02-06 | 2013-02-06 | |
US14/174,761 US9607206B2 (en) | 2013-02-06 | 2014-02-06 | Biometric sensing device for three dimensional imaging of subcutaneous structures embedded within finger tissue |
US201715470465A | 2017-03-27 | 2017-03-27 | |
US16/245,299 US10621404B2 (en) | 2013-02-06 | 2019-01-11 | Biometric sensing device for three dimensional imaging of subcutaneous structures embedded within finger tissue |
US16/385,437 US20190311174A1 (en) | 2013-02-06 | 2019-04-16 | Method and system for 2d and 3d biometric sensing using a same sensor |
Related Parent Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
US16/245,299 Continuation US10621404B2 (en) | 2013-02-06 | 2019-01-11 | Biometric sensing device for three dimensional imaging of subcutaneous structures embedded within finger tissue |
Publications (1)
Publication Number | Publication Date |
---|---|
US20190311174A1 true US20190311174A1 (en) | 2019-10-10 |
Family
ID=51259249
Family Applications (5)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
US14/174,761 Active 2034-06-30 US9607206B2 (en) | 2013-02-06 | 2014-02-06 | Biometric sensing device for three dimensional imaging of subcutaneous structures embedded within finger tissue |
US16/245,299 Expired - Fee Related US10621404B2 (en) | 2013-02-06 | 2019-01-11 | Biometric sensing device for three dimensional imaging of subcutaneous structures embedded within finger tissue |
US16/385,437 Abandoned US20190311174A1 (en) | 2013-02-06 | 2019-04-16 | Method and system for 2d and 3d biometric sensing using a same sensor |
US16/385,413 Active US10528785B2 (en) | 2013-02-06 | 2019-04-16 | Method and system for beam control in biometric sensing |
US16/385,339 Abandoned US20190311173A1 (en) | 2013-02-06 | 2019-04-16 | Method and system for biometric sensing and subcutaneous structure detection |
Family Applications Before (2)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
US14/174,761 Active 2034-06-30 US9607206B2 (en) | 2013-02-06 | 2014-02-06 | Biometric sensing device for three dimensional imaging of subcutaneous structures embedded within finger tissue |
US16/245,299 Expired - Fee Related US10621404B2 (en) | 2013-02-06 | 2019-01-11 | Biometric sensing device for three dimensional imaging of subcutaneous structures embedded within finger tissue |
Family Applications After (2)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
US16/385,413 Active US10528785B2 (en) | 2013-02-06 | 2019-04-16 | Method and system for beam control in biometric sensing |
US16/385,339 Abandoned US20190311173A1 (en) | 2013-02-06 | 2019-04-16 | Method and system for biometric sensing and subcutaneous structure detection |
Country Status (7)
Country | Link |
---|---|
US (5) | US9607206B2 (en) |
EP (1) | EP2954458A4 (en) |
JP (3) | JP2016513983A (en) |
KR (1) | KR20150115789A (en) |
CN (1) | CN105264542A (en) |
CA (1) | CA2900479A1 (en) |
WO (1) | WO2014124167A1 (en) |
Cited By (1)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US20220198173A1 (en) * | 2020-12-22 | 2022-06-23 | Qualcomm Incorporated | Apparatus and method for ultrasonic fingerprint and touch sensing |
Families Citing this family (103)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US9396382B2 (en) * | 2012-08-17 | 2016-07-19 | Flashscan3D, Llc | System and method for a biometric image sensor with spoofing detection |
US9511994B2 (en) | 2012-11-28 | 2016-12-06 | Invensense, Inc. | Aluminum nitride (AlN) devices with infrared absorption structural layer |
US9618405B2 (en) | 2014-08-06 | 2017-04-11 | Invensense, Inc. | Piezoelectric acoustic resonator based sensor |
US10497747B2 (en) | 2012-11-28 | 2019-12-03 | Invensense, Inc. | Integrated piezoelectric microelectromechanical ultrasound transducer (PMUT) on integrated circuit (IC) for fingerprint sensing |
US9114977B2 (en) | 2012-11-28 | 2015-08-25 | Invensense, Inc. | MEMS device and process for RF and low resistance applications |
US10726231B2 (en) | 2012-11-28 | 2020-07-28 | Invensense, Inc. | Integrated piezoelectric microelectromechanical ultrasound transducer (PMUT) on integrated circuit (IC) for fingerprint sensing |
EP2954458A4 (en) * | 2013-02-06 | 2016-11-09 | Sonavation Inc | Biometric sensing device for three dimensional imaging of subcutaneous structures embedded within finger tissue |
US9754149B2 (en) * | 2013-04-01 | 2017-09-05 | AMI Research & Development, LLC | Fingerprint based smart phone user verification |
US10121049B2 (en) | 2013-04-01 | 2018-11-06 | AMI Research & Development, LLC | Fingerprint based smart phone user verification |
CN104091107B (en) * | 2014-07-21 | 2018-01-16 | 友达光电股份有限公司 | Identity recognition device and operation method thereof |
KR102356453B1 (en) * | 2014-08-29 | 2022-01-27 | 삼성전자주식회사 | Finger print detection sensor and finger print detection system comprising the same |
US9639765B2 (en) * | 2014-09-05 | 2017-05-02 | Qualcomm Incorporated | Multi-stage liveness determination |
WO2016061412A1 (en) * | 2014-10-15 | 2016-04-21 | Qualcomm Incorporated | Active beam-forming technique for piezoelectric ultrasonic transducer array |
US10139479B2 (en) * | 2014-10-15 | 2018-11-27 | Qualcomm Incorporated | Superpixel array of piezoelectric ultrasonic transducers for 2-D beamforming |
KR102402146B1 (en) * | 2015-04-21 | 2022-05-26 | 삼성전자주식회사 | Method and apparatus for sensing fingerprints |
JP6606854B2 (en) * | 2015-04-30 | 2019-11-20 | セイコーエプソン株式会社 | Ultrasonic device unit and probe, electronic apparatus and ultrasonic diagnostic apparatus |
US9424456B1 (en) | 2015-06-24 | 2016-08-23 | Amazon Technologies, Inc. | Ultrasonic fingerprint authentication based on beam forming |
US10387704B2 (en) * | 2015-06-29 | 2019-08-20 | Qualcomm Incorporated | Method and apparatus for enabling the touchscreen display of a mobile device |
CN106407870A (en) * | 2015-07-28 | 2017-02-15 | 宇龙计算机通信科技(深圳)有限公司 | Fingerprint identification method and user equipment |
US9928398B2 (en) | 2015-08-17 | 2018-03-27 | Invensense, Inc. | Always-on sensor device for human touch |
US9722793B2 (en) | 2015-09-09 | 2017-08-01 | Microsoft Technology Licensing, Llc | Personal identification via acoustically stimulated biospeckles |
US9726755B2 (en) | 2015-09-23 | 2017-08-08 | Qualcomm Incorporated | Spoof detection by ultrasonic subdermal probe |
US10140534B2 (en) * | 2015-09-26 | 2018-11-27 | Qualcomm Incorporated | Ultrasonic imaging devices and methods |
US10497748B2 (en) | 2015-10-14 | 2019-12-03 | Qualcomm Incorporated | Integrated piezoelectric micromechanical ultrasonic transducer pixel and array |
US10282585B2 (en) | 2015-11-13 | 2019-05-07 | Cypress Semiconductor Corporation | Sensor-compatible overlay |
US10832029B2 (en) | 2015-11-13 | 2020-11-10 | Cypress Semiconductor Corporation | Sensor-compatible overlay |
US9639734B1 (en) | 2015-11-13 | 2017-05-02 | Cypress Semiconductor Corporation | Fingerprint sensor-compatible overlay material |
CN105512636A (en) * | 2015-12-19 | 2016-04-20 | 南京福瑞林生物科技有限公司 | Living fingerprint identification device and living fingerprint identification method |
KR101730842B1 (en) * | 2016-02-05 | 2017-04-27 | 주식회사 베프스 | Apparatus for recognizing biometric information and method for activating a plurality of piezoelectric element individually |
KR101678012B1 (en) * | 2016-02-05 | 2016-11-21 | 주식회사 베프스 | Apparatus and method for recognizing biometric information |
US11020057B2 (en) * | 2016-02-12 | 2021-06-01 | Qualcomm Incorporated | Ultrasound devices for estimating blood pressure and other cardiovascular properties |
CN105760851B (en) * | 2016-03-10 | 2018-03-02 | 广东欧珀移动通信有限公司 | Method and terminal for fingerprint identification |
KR101661634B1 (en) * | 2016-03-11 | 2016-09-30 | 주식회사 베프스 | Apparatus for recognizing biometric information and method for activating a plurality of piezoelectric element individually |
KR101661642B1 (en) * | 2016-03-11 | 2016-09-30 | 주식회사 베프스 | Apparatus and method for recognizing biometric information |
FR3049090B1 (en) * | 2016-03-21 | 2021-06-25 | Sebastien Jean Serge Dupont | ADAPTIVE BIOMETRIC AUTHENTICATION DEVICE BY ULTRASOUND, VISIBLE CONTRAST AND INFRARED LIGHT PHOTOGRAPHS, WITHOUT DISCLOSURE, THROUGH A DECENTRALIZED COMPUTER NETWORK |
US10325915B2 (en) | 2016-05-04 | 2019-06-18 | Invensense, Inc. | Two-dimensional array of CMOS control elements |
US10445547B2 (en) | 2016-05-04 | 2019-10-15 | Invensense, Inc. | Device mountable packaging of ultrasonic transducers |
US10656255B2 (en) | 2016-05-04 | 2020-05-19 | Invensense, Inc. | Piezoelectric micromachined ultrasonic transducer (PMUT) |
US10670716B2 (en) | 2016-05-04 | 2020-06-02 | Invensense, Inc. | Operating a two-dimensional array of ultrasonic transducers |
US10315222B2 (en) | 2016-05-04 | 2019-06-11 | Invensense, Inc. | Two-dimensional array of CMOS control elements |
US10235551B2 (en) | 2016-05-06 | 2019-03-19 | Qualcomm Incorporated | Biometric system with photoacoustic imaging |
US10366269B2 (en) * | 2016-05-06 | 2019-07-30 | Qualcomm Incorporated | Biometric system with photoacoustic imaging |
US11673165B2 (en) | 2016-05-10 | 2023-06-13 | Invensense, Inc. | Ultrasonic transducer operable in a surface acoustic wave (SAW) mode |
US10441975B2 (en) | 2016-05-10 | 2019-10-15 | Invensense, Inc. | Supplemental sensor modes and systems for ultrasonic transducers |
US10539539B2 (en) | 2016-05-10 | 2020-01-21 | Invensense, Inc. | Operation of an ultrasonic sensor |
US10600403B2 (en) | 2016-05-10 | 2020-03-24 | Invensense, Inc. | Transmit operation of an ultrasonic sensor |
US10452887B2 (en) | 2016-05-10 | 2019-10-22 | Invensense, Inc. | Operating a fingerprint sensor comprised of ultrasonic transducers |
US10562070B2 (en) | 2016-05-10 | 2020-02-18 | Invensense, Inc. | Receive operation of an ultrasonic sensor |
US10706835B2 (en) * | 2016-05-10 | 2020-07-07 | Invensense, Inc. | Transmit beamforming of a two-dimensional array of ultrasonic transducers |
US10408797B2 (en) | 2016-05-10 | 2019-09-10 | Invensense, Inc. | Sensing device with a temperature sensor |
US10632500B2 (en) | 2016-05-10 | 2020-04-28 | Invensense, Inc. | Ultrasonic transducer with a non-uniform membrane |
US10346663B2 (en) * | 2016-09-05 | 2019-07-09 | Nanchang O-Film Bio-Identification Technology Co., Ltd | Fingerprint sensor and fingerprint identification module comprising the same |
US10192094B2 (en) * | 2016-09-05 | 2019-01-29 | Nanchang O-Film Bio-Identification Technology Co., Ltd | Ultrasonic fingerprint sensor package, ultrasonic fingerprint identification device and electronic device |
US10268866B2 (en) * | 2016-09-05 | 2019-04-23 | Nanchang O-Film Bio-Identification Technology Co., Ltd. | Ultrasonic probe and manufacturing method of the same and ultrasonic fingerprint recognition device |
US10192093B2 (en) * | 2016-09-05 | 2019-01-29 | Nanchang O-Film Bio-Identification Technology Co., Ltd. | Ultrasonic fingerprint sensor package |
US10387706B2 (en) * | 2016-09-05 | 2019-08-20 | Nanchang O-Film Bio-Identification Technology Co., Ltd. | Ultrasonic transducer of ultrasonic fingerprint sensor and manufacturing method thereof |
CN106446817B (en) * | 2016-09-18 | 2018-03-20 | 京东方科技集团股份有限公司 | Fingerprint recognition device, touch display panel and fingerprint recognition device driving method |
CN106529463B (en) * | 2016-10-28 | 2019-04-23 | 京东方科技集团股份有限公司 | A kind of fingerprint recognition display panel, display device and fingerprint identification method |
US11039814B2 (en) * | 2016-12-04 | 2021-06-22 | Exo Imaging, Inc. | Imaging devices having piezoelectric transducers |
US10540044B2 (en) * | 2016-12-14 | 2020-01-21 | Cypress Semiconductor Corporation | Capacitive sensing with multi-pattern scan |
WO2018126081A1 (en) * | 2016-12-28 | 2018-07-05 | Shadd Warren M | Systems and methods for biometrically authenticating a user using authentication data and liveness data |
WO2018151547A1 (en) * | 2017-02-16 | 2018-08-23 | 주식회사 베프스 | Biometric information recognition apparatus and biometric information recognition method using same |
US11374929B2 (en) | 2017-03-21 | 2022-06-28 | Global E-Dentity, Inc. | Biometric authentication for an augmented reality or a virtual reality device |
US10880303B2 (en) | 2017-03-21 | 2020-12-29 | Global E-Dentity, Inc. | Real-time COVID-19 outbreak identification with non-invasive, internal imaging for dual biometric authentication and biometric health monitoring |
US10135822B2 (en) | 2017-03-21 | 2018-11-20 | YouaretheID, LLC | Biometric authentication of individuals utilizing characteristics of bone and blood vessel structures |
US10515255B2 (en) * | 2017-03-24 | 2019-12-24 | Qualcomm Incorporated | Fingerprint sensor with bioimpedance indicator |
US10552658B2 (en) | 2017-03-24 | 2020-02-04 | Qualcomm Incorporated | Biometric sensor with finger-force navigation |
US10438040B2 (en) | 2017-03-24 | 2019-10-08 | Qualcomm Incorporated | Multi-functional ultrasonic fingerprint sensor |
CN106897715A (en) * | 2017-03-31 | 2017-06-27 | 努比亚技术有限公司 | A kind of unlocked by fingerprint processing method and mobile terminal |
US10846501B2 (en) | 2017-04-28 | 2020-11-24 | The Board Of Trustees Of The Leland Stanford Junior University | Acoustic biometric touch scanner |
US10489627B2 (en) | 2017-04-28 | 2019-11-26 | The Board Of Trustees Of The Leland Stanford Junior University | Acoustic biometric touch scanner |
US9953205B1 (en) | 2017-04-28 | 2018-04-24 | The Board Of Trustees Of The Leland Stanford Junior University | Acoustic biometric touch scanner |
US10891461B2 (en) | 2017-05-22 | 2021-01-12 | Invensense, Inc. | Live fingerprint detection utilizing an integrated ultrasound and infrared sensor |
US10474862B2 (en) | 2017-06-01 | 2019-11-12 | Invensense, Inc. | Image generation in an electronic device using ultrasonic transducers |
WO2018221429A1 (en) * | 2017-06-02 | 2018-12-06 | パナソニックIpマネジメント株式会社 | Security system comprising biometric authentication sensor, and electronic device |
US10643052B2 (en) | 2017-06-28 | 2020-05-05 | Invensense, Inc. | Image generation in an electronic device using ultrasonic transducers |
US10592718B2 (en) | 2017-08-09 | 2020-03-17 | The Board Of Trustees Of The Leland Stanford Junior University | Interactive biometric touch scanner |
KR101907324B1 (en) | 2017-11-27 | 2018-10-11 | 재단법인대구경북과학기술원 | Ultrasonic sensor, manufacturing method of the same, and security system |
US10984209B2 (en) | 2017-12-01 | 2021-04-20 | Invensense, Inc. | Darkfield modeling |
WO2019109010A1 (en) | 2017-12-01 | 2019-06-06 | Invensense, Inc. | Darkfield tracking |
US10997388B2 (en) | 2017-12-01 | 2021-05-04 | Invensense, Inc. | Darkfield contamination detection |
US11151355B2 (en) | 2018-01-24 | 2021-10-19 | Invensense, Inc. | Generation of an estimated fingerprint |
US10755067B2 (en) | 2018-03-22 | 2020-08-25 | Invensense, Inc. | Operating a fingerprint sensor comprised of ultrasonic transducers |
US11651610B2 (en) * | 2018-05-31 | 2023-05-16 | Qualcomm Incorporated | Heart rate and respiration rate measurement using a fingerprint sensor |
KR102160716B1 (en) * | 2018-09-04 | 2020-09-28 | 한국세라믹기술원 | Transparent panel biometric system using multiple ultrasonic finger scanners |
US10936843B2 (en) | 2018-12-28 | 2021-03-02 | Invensense, Inc. | Segmented image acquisition |
WO2020188866A1 (en) | 2019-03-18 | 2020-09-24 | 株式会社村田製作所 | Piezoelectric device |
WO2020263875A1 (en) | 2019-06-24 | 2020-12-30 | Invensense, Inc. | Fake finger detection using ridge features |
FR3097668B1 (en) * | 2019-06-24 | 2021-07-02 | Commissariat Energie Atomique | TOUCH SURFACE DEVICE |
US11216681B2 (en) | 2019-06-25 | 2022-01-04 | Invensense, Inc. | Fake finger detection based on transient features |
US11176345B2 (en) | 2019-07-17 | 2021-11-16 | Invensense, Inc. | Ultrasonic fingerprint sensor with a contact layer of non-uniform thickness |
US11216632B2 (en) | 2019-07-17 | 2022-01-04 | Invensense, Inc. | Ultrasonic fingerprint sensor with a contact layer of non-uniform thickness |
WO2021020661A1 (en) * | 2019-07-31 | 2021-02-04 | 주식회사 베프스 | Ultrasound-based dual biometric information recognition device and biometric information data processing method |
US11232549B2 (en) | 2019-08-23 | 2022-01-25 | Invensense, Inc. | Adapting a quality threshold for a fingerprint image |
KR102108405B1 (en) * | 2019-08-23 | 2020-05-08 | 박하연 | System for measuring body information, posture information and range of motion and method thereof |
US11392789B2 (en) | 2019-10-21 | 2022-07-19 | Invensense, Inc. | Fingerprint authentication using a synthetic enrollment image |
WO2021183457A1 (en) | 2020-03-09 | 2021-09-16 | Invensense, Inc. | Ultrasonic fingerprint sensor with a contact layer of non-uniform thickness |
US11243300B2 (en) | 2020-03-10 | 2022-02-08 | Invensense, Inc. | Operating a fingerprint sensor comprised of ultrasonic transducers and a presence sensor |
US11328165B2 (en) | 2020-04-24 | 2022-05-10 | Invensense, Inc. | Pressure-based activation of fingerprint spoof detection |
US11995909B2 (en) | 2020-07-17 | 2024-05-28 | Tdk Corporation | Multipath reflection correction |
US12174295B2 (en) | 2020-08-07 | 2024-12-24 | Tdk Corporation | Acoustic multipath correction |
US11385770B1 (en) | 2021-04-21 | 2022-07-12 | Qualcomm Incorporated | User interfaces for single-handed mobile device control |
US12260050B2 (en) | 2021-08-25 | 2025-03-25 | Tdk Corporation | Differential receive at an ultrasonic transducer |
Citations (3)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US6280387B1 (en) * | 1998-05-06 | 2001-08-28 | Siemens Medical Systems, Inc. | Three-dimensional tissue/flow ultrasound imaging system |
US20030216646A1 (en) * | 2002-03-15 | 2003-11-20 | Angelsen Bjorn A.J. | Multiple scan-plane ultrasound imaging of objects |
US20100113952A1 (en) * | 2008-11-03 | 2010-05-06 | Raguin Daniel H | Apparatus and method for the identification of fake fingerprints |
Family Cites Families (33)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
JP2789234B2 (en) | 1989-10-02 | 1998-08-20 | 株式会社日立メディコ | Ultrasound diagnostic equipment |
AU2781401A (en) * | 2000-01-10 | 2001-07-24 | Tarian, Llc | Device using histological and physiological biometric marker for authentication and activation |
CA2403394C (en) * | 2000-03-23 | 2012-01-03 | Cross Match Technologies, Inc. | Piezoelectric identification device and applications thereof |
US7067962B2 (en) | 2000-03-23 | 2006-06-27 | Cross Match Technologies, Inc. | Multiplexer for a piezo ceramic identification device |
US7141918B2 (en) * | 2000-03-23 | 2006-11-28 | Cross Match Technologies, Inc. | Method for obtaining biometric data for an individual in a secure transaction |
US7536557B2 (en) * | 2001-03-22 | 2009-05-19 | Ensign Holdings | Method for biometric authentication through layering biometric traits |
JP2002175529A (en) * | 2000-12-06 | 2002-06-21 | Matsushita Electric Ind Co Ltd | Individual identifying device |
NL1017856C2 (en) * | 2001-04-17 | 2002-10-18 | Hendricus Hermanus Van Velden | S4U (Safety for you). |
JP2003263640A (en) * | 2002-03-08 | 2003-09-19 | Toyota Motor Corp | Personal identification device using biological information |
US6640636B1 (en) | 2002-05-20 | 2003-11-04 | Kohji Toda | Ultrasound radiating and receiving device |
CN1893878A (en) | 2003-12-16 | 2007-01-10 | 株式会社日立医药 | Ultrasonographic bio-movement detection device, image presentation device using the same, and ultrasonographic curing device |
FR2864871A1 (en) | 2004-01-06 | 2005-07-08 | Thomson Licensing Sa | METHOD OF DISCOVERING A DOMESTIC NETWORK AND APPARATUS IMPLEMENTING THE METHOD |
JP4567479B2 (en) * | 2005-02-04 | 2010-10-20 | セイコーインスツル株式会社 | Biological information measuring device and biological information measuring method |
US7567016B2 (en) | 2005-02-04 | 2009-07-28 | Siemens Medical Solutions Usa, Inc. | Multi-dimensional ultrasound transducer array |
US20060195035A1 (en) * | 2005-02-28 | 2006-08-31 | Dehchuan Sun | Non-invasive radial artery blood pressure waveform measuring apparatus system and uses thereof |
JP2006325704A (en) * | 2005-05-24 | 2006-12-07 | Matsushita Electric Ind Co Ltd | Ultrasonic diagnostic apparatus |
US7837626B2 (en) | 2005-08-05 | 2010-11-23 | Siemens Medical Solutions Usa, Inc. | Contrast agent manipulation with medical ultrasound imaging |
US20070046149A1 (en) | 2005-08-23 | 2007-03-01 | Zipparo Michael J | Ultrasound probe transducer assembly and production method |
JP4886371B2 (en) * | 2006-06-07 | 2012-02-29 | 株式会社日立情報制御ソリューションズ | Biometric authentication method and system |
CN101536545B (en) | 2006-11-08 | 2013-02-06 | 松下电器产业株式会社 | ultrasonic probe |
CN101482916B (en) * | 2008-01-08 | 2012-05-23 | 神盾股份有限公司 | Identification method of finger thermal image |
US8331633B2 (en) | 2008-05-08 | 2012-12-11 | Sonavation, Inc. | Method and system for multi-mode mechanical resonator |
US8335356B2 (en) | 2008-05-08 | 2012-12-18 | Sonavation, Inc. | Mechanical resonator optimization using shear wave damping |
US20110200237A1 (en) * | 2008-10-15 | 2011-08-18 | Nec Corporation | Pattern matching device and pattern matching method |
JP5169940B2 (en) * | 2009-03-26 | 2013-03-27 | セイコーエプソン株式会社 | Imaging device and authentication device |
US9042611B2 (en) * | 2010-01-29 | 2015-05-26 | Mayo Foundation For Medical Education And Research | Automated vascular region separation in medical imaging |
US9841318B1 (en) | 2010-04-30 | 2017-12-12 | Sonavation, Inc. | Apparatus for acoustic sensing |
JPWO2012020591A1 (en) * | 2010-08-09 | 2013-10-28 | 日本電気株式会社 | Individual identification system, feature amount specifying device, feature amount specifying method and program |
DE102011080534A1 (en) | 2011-08-05 | 2013-02-07 | Hesse & Knipps Gmbh | Method for ultrasonic bonding |
US20150190114A1 (en) | 2012-09-28 | 2015-07-09 | Hitachi Aloka Medical, Ltd. | Portable ultrasound imaging apparatus |
EP2954458A4 (en) * | 2013-02-06 | 2016-11-09 | Sonavation Inc | Biometric sensing device for three dimensional imaging of subcutaneous structures embedded within finger tissue |
KR101872300B1 (en) | 2013-07-15 | 2018-06-28 | 퀄컴 인코포레이티드 | Method and integrated circuit for operating a sensor array |
JP6402983B2 (en) | 2014-08-29 | 2018-10-10 | セイコーエプソン株式会社 | Ultrasonic device, method for manufacturing ultrasonic device, ultrasonic probe, ultrasonic measuring device, electronic equipment |
-
2014
- 2014-02-06 EP EP14748559.3A patent/EP2954458A4/en not_active Withdrawn
- 2014-02-06 CN CN201480011088.5A patent/CN105264542A/en active Pending
- 2014-02-06 US US14/174,761 patent/US9607206B2/en active Active
- 2014-02-06 CA CA2900479A patent/CA2900479A1/en not_active Abandoned
- 2014-02-06 JP JP2015557074A patent/JP2016513983A/en not_active Withdrawn
- 2014-02-06 WO PCT/US2014/015145 patent/WO2014124167A1/en active Application Filing
- 2014-02-06 KR KR1020157021371A patent/KR20150115789A/en not_active Ceased
-
2018
- 2018-05-29 JP JP2018102366A patent/JP2018192266A/en active Pending
- 2018-11-29 JP JP2018223622A patent/JP2019069160A/en active Pending
-
2019
- 2019-01-11 US US16/245,299 patent/US10621404B2/en not_active Expired - Fee Related
- 2019-04-16 US US16/385,437 patent/US20190311174A1/en not_active Abandoned
- 2019-04-16 US US16/385,413 patent/US10528785B2/en active Active
- 2019-04-16 US US16/385,339 patent/US20190311173A1/en not_active Abandoned
Patent Citations (3)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US6280387B1 (en) * | 1998-05-06 | 2001-08-28 | Siemens Medical Systems, Inc. | Three-dimensional tissue/flow ultrasound imaging system |
US20030216646A1 (en) * | 2002-03-15 | 2003-11-20 | Angelsen Bjorn A.J. | Multiple scan-plane ultrasound imaging of objects |
US20100113952A1 (en) * | 2008-11-03 | 2010-05-06 | Raguin Daniel H | Apparatus and method for the identification of fake fingerprints |
Cited By (2)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US20220198173A1 (en) * | 2020-12-22 | 2022-06-23 | Qualcomm Incorporated | Apparatus and method for ultrasonic fingerprint and touch sensing |
US11837009B2 (en) * | 2020-12-22 | 2023-12-05 | Qualcomm Incorporated | Apparatus and method for ultrasonic fingerprint and touch sensing |
Also Published As
Publication number | Publication date |
---|---|
US20190251319A1 (en) | 2019-08-15 |
WO2014124167A1 (en) | 2014-08-14 |
JP2018192266A (en) | 2018-12-06 |
US10621404B2 (en) | 2020-04-14 |
US10528785B2 (en) | 2020-01-07 |
US20140219521A1 (en) | 2014-08-07 |
US20190244001A1 (en) | 2019-08-08 |
JP2016513983A (en) | 2016-05-19 |
CA2900479A1 (en) | 2014-08-14 |
EP2954458A4 (en) | 2016-11-09 |
US9607206B2 (en) | 2017-03-28 |
US20190311173A1 (en) | 2019-10-10 |
KR20150115789A (en) | 2015-10-14 |
CN105264542A (en) | 2016-01-20 |
EP2954458A1 (en) | 2015-12-16 |
JP2019069160A (en) | 2019-05-09 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
US10528785B2 (en) | Method and system for beam control in biometric sensing | |
US11017251B2 (en) | Ultrasonic imaging devices and methods | |
US11301552B2 (en) | Medical device with integrated ultrasonic authentication | |
WO2018200642A1 (en) | Acoustic biometric touch scanner | |
KR20130060875A (en) | Sensor and method for detecting fingerprint | |
JP2016513983A5 (en) | ||
CN109788942B (en) | Ultrasonic diagnostic apparatus and method for controlling ultrasonic diagnostic apparatus | |
KR20180061826A (en) | Biometric information recognition device, system and method | |
Iula et al. | An enhanced ultrasound technique for 3D palmprint recognition | |
US20140296713A1 (en) | Ultrasonic measurement apparatus, ultrasonic imaging apparatus, and ultrasonic measurement method | |
EP3659515A1 (en) | Imaging system comprising an ultrasound transducer array and skin contact electrodes, and corresponding imaging method | |
CN105078472B (en) | Physiological characteristic image acquiring device and method | |
US10685204B2 (en) | Biometric age estimation via ultrasonic imaging | |
KR102138358B1 (en) | Ultrasound-based biometric information recognition device and finger vein data build up method | |
US20240374233A1 (en) | System for automated real-time detection, outlining, tracking and characterization of blood vessels in ultrasound imaging | |
Iula et al. | Wide 3D ultrasound palmprint for biometric recognition | |
Iula et al. | Experimental evaluation of the moving linear array technique applied to livescan biometrics | |
JP6744769B2 (en) | Ultrasonic diagnostic apparatus and method for acquiring ultrasonic image inside living body |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
AS | Assignment |
Owner name: SONAVATION, INC., FLORIDA Free format text: ASSIGNMENT OF ASSIGNORS INTEREST;ASSIGNORS:SCHMITT, RAINER M.;CRAIG, RONALD A.;REEL/FRAME:050315/0244 Effective date: 20170202 |
|
STPP | Information on status: patent application and granting procedure in general |
Free format text: NON FINAL ACTION MAILED |
|
STPP | Information on status: patent application and granting procedure in general |
Free format text: RESPONSE TO NON-FINAL OFFICE ACTION ENTERED AND FORWARDED TO EXAMINER |
|
STPP | Information on status: patent application and granting procedure in general |
Free format text: FINAL REJECTION MAILED |
|
STCB | Information on status: application discontinuation |
Free format text: ABANDONED -- FAILURE TO RESPOND TO AN OFFICE ACTION |