US20130123616A1 - Medical Workflow System and Method - Google Patents
Medical Workflow System and Method Download PDFInfo
- Publication number
- US20130123616A1 US20130123616A1 US13/679,795 US201213679795A US2013123616A1 US 20130123616 A1 US20130123616 A1 US 20130123616A1 US 201213679795 A US201213679795 A US 201213679795A US 2013123616 A1 US2013123616 A1 US 2013123616A1
- Authority
- US
- United States
- Prior art keywords
- instructions
- bedside controller
- workflow
- measurement
- images
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Abandoned
Links
Images
Classifications
-
- A—HUMAN NECESSITIES
- A61—MEDICAL OR VETERINARY SCIENCE; HYGIENE
- A61B—DIAGNOSIS; SURGERY; IDENTIFICATION
- A61B5/00—Measuring for diagnostic purposes; Identification of persons
- A61B5/0059—Measuring for diagnostic purposes; Identification of persons using light, e.g. diagnosis by transillumination, diascopy, fluorescence
- A61B5/0082—Measuring for diagnostic purposes; Identification of persons using light, e.g. diagnosis by transillumination, diascopy, fluorescence adapted for particular medical purposes
- A61B5/0084—Measuring for diagnostic purposes; Identification of persons using light, e.g. diagnosis by transillumination, diascopy, fluorescence adapted for particular medical purposes for introduction into the body, e.g. by catheters
-
- A—HUMAN NECESSITIES
- A61—MEDICAL OR VETERINARY SCIENCE; HYGIENE
- A61B—DIAGNOSIS; SURGERY; IDENTIFICATION
- A61B5/00—Measuring for diagnostic purposes; Identification of persons
- A61B5/0059—Measuring for diagnostic purposes; Identification of persons using light, e.g. diagnosis by transillumination, diascopy, fluorescence
-
- A—HUMAN NECESSITIES
- A61—MEDICAL OR VETERINARY SCIENCE; HYGIENE
- A61B—DIAGNOSIS; SURGERY; IDENTIFICATION
- A61B5/00—Measuring for diagnostic purposes; Identification of persons
- A61B5/0059—Measuring for diagnostic purposes; Identification of persons using light, e.g. diagnosis by transillumination, diascopy, fluorescence
- A61B5/0062—Arrangements for scanning
- A61B5/0066—Optical coherence imaging
-
- A—HUMAN NECESSITIES
- A61—MEDICAL OR VETERINARY SCIENCE; HYGIENE
- A61B—DIAGNOSIS; SURGERY; IDENTIFICATION
- A61B5/00—Measuring for diagnostic purposes; Identification of persons
- A61B5/0059—Measuring for diagnostic purposes; Identification of persons using light, e.g. diagnosis by transillumination, diascopy, fluorescence
- A61B5/0077—Devices for viewing the surface of the body, e.g. camera, magnifying lens
-
- A—HUMAN NECESSITIES
- A61—MEDICAL OR VETERINARY SCIENCE; HYGIENE
- A61B—DIAGNOSIS; SURGERY; IDENTIFICATION
- A61B5/00—Measuring for diagnostic purposes; Identification of persons
- A61B5/74—Details of notification to user or communication with user or patient ; user input means
- A61B5/742—Details of notification to user or communication with user or patient ; user input means using visual displays
- A61B5/7435—Displaying user selection data, e.g. icons in a graphical user interface
-
- A—HUMAN NECESSITIES
- A61—MEDICAL OR VETERINARY SCIENCE; HYGIENE
- A61B—DIAGNOSIS; SURGERY; IDENTIFICATION
- A61B5/00—Measuring for diagnostic purposes; Identification of persons
- A61B5/74—Details of notification to user or communication with user or patient ; user input means
- A61B5/742—Details of notification to user or communication with user or patient ; user input means using visual displays
- A61B5/7445—Display arrangements, e.g. multiple display units
-
- A—HUMAN NECESSITIES
- A61—MEDICAL OR VETERINARY SCIENCE; HYGIENE
- A61B—DIAGNOSIS; SURGERY; IDENTIFICATION
- A61B6/00—Apparatus or devices for radiation diagnosis; Apparatus or devices for radiation diagnosis combined with radiation therapy equipment
- A61B6/50—Apparatus or devices for radiation diagnosis; Apparatus or devices for radiation diagnosis combined with radiation therapy equipment specially adapted for specific body parts; specially adapted for specific clinical applications
- A61B6/504—Apparatus or devices for radiation diagnosis; Apparatus or devices for radiation diagnosis combined with radiation therapy equipment specially adapted for specific body parts; specially adapted for specific clinical applications for diagnosis of blood vessels, e.g. by angiography
-
- A—HUMAN NECESSITIES
- A61—MEDICAL OR VETERINARY SCIENCE; HYGIENE
- A61B—DIAGNOSIS; SURGERY; IDENTIFICATION
- A61B8/00—Diagnosis using ultrasonic, sonic or infrasonic waves
- A61B8/12—Diagnosis using ultrasonic, sonic or infrasonic waves in body cavities or body tracts, e.g. by using catheters
-
- A—HUMAN NECESSITIES
- A61—MEDICAL OR VETERINARY SCIENCE; HYGIENE
- A61B—DIAGNOSIS; SURGERY; IDENTIFICATION
- A61B8/00—Diagnosis using ultrasonic, sonic or infrasonic waves
- A61B8/44—Constructional features of the ultrasonic, sonic or infrasonic diagnostic device
- A61B8/4416—Constructional features of the ultrasonic, sonic or infrasonic diagnostic device related to combined acquisition of different diagnostic modalities, e.g. combination of ultrasound and X-ray acquisitions
-
- A—HUMAN NECESSITIES
- A61—MEDICAL OR VETERINARY SCIENCE; HYGIENE
- A61M—DEVICES FOR INTRODUCING MEDIA INTO, OR ONTO, THE BODY; DEVICES FOR TRANSDUCING BODY MEDIA OR FOR TAKING MEDIA FROM THE BODY; DEVICES FOR PRODUCING OR ENDING SLEEP OR STUPOR
- A61M5/00—Devices for bringing media into the body in a subcutaneous, intra-vascular or intramuscular way; Accessories therefor, e.g. filling or cleaning devices, arm-rests
- A61M5/007—Devices for bringing media into the body in a subcutaneous, intra-vascular or intramuscular way; Accessories therefor, e.g. filling or cleaning devices, arm-rests for contrast media
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F3/00—Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
- G06F3/01—Input arrangements or combined input and output arrangements for interaction between user and computer
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F3/00—Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
- G06F3/01—Input arrangements or combined input and output arrangements for interaction between user and computer
- G06F3/03—Arrangements for converting the position or the displacement of a member into a coded form
- G06F3/041—Digitisers, e.g. for touch screens or touch pads, characterised by the transducing means
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F3/00—Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
- G06F3/01—Input arrangements or combined input and output arrangements for interaction between user and computer
- G06F3/048—Interaction techniques based on graphical user interfaces [GUI]
- G06F3/0487—Interaction techniques based on graphical user interfaces [GUI] using specific features provided by the input device, e.g. functions controlled by the rotation of a mouse with dual sensing arrangements, or of the nature of the input device, e.g. tap gestures based on pressure sensed by a digitiser
- G06F3/0488—Interaction techniques based on graphical user interfaces [GUI] using specific features provided by the input device, e.g. functions controlled by the rotation of a mouse with dual sensing arrangements, or of the nature of the input device, e.g. tap gestures based on pressure sensed by a digitiser using a touch-screen or digitiser, e.g. input of commands through traced gestures
- G06F3/04883—Interaction techniques based on graphical user interfaces [GUI] using specific features provided by the input device, e.g. functions controlled by the rotation of a mouse with dual sensing arrangements, or of the nature of the input device, e.g. tap gestures based on pressure sensed by a digitiser using a touch-screen or digitiser, e.g. input of commands through traced gestures for inputting data by handwriting, e.g. gesture or text
-
- G—PHYSICS
- G16—INFORMATION AND COMMUNICATION TECHNOLOGY [ICT] SPECIALLY ADAPTED FOR SPECIFIC APPLICATION FIELDS
- G16H—HEALTHCARE INFORMATICS, i.e. INFORMATION AND COMMUNICATION TECHNOLOGY [ICT] SPECIALLY ADAPTED FOR THE HANDLING OR PROCESSING OF MEDICAL OR HEALTHCARE DATA
- G16H40/00—ICT specially adapted for the management or administration of healthcare resources or facilities; ICT specially adapted for the management or operation of medical equipment or devices
- G16H40/60—ICT specially adapted for the management or administration of healthcare resources or facilities; ICT specially adapted for the management or operation of medical equipment or devices for the operation of medical equipment or devices
- G16H40/63—ICT specially adapted for the management or administration of healthcare resources or facilities; ICT specially adapted for the management or operation of medical equipment or devices for the operation of medical equipment or devices for local operation
-
- A—HUMAN NECESSITIES
- A61—MEDICAL OR VETERINARY SCIENCE; HYGIENE
- A61B—DIAGNOSIS; SURGERY; IDENTIFICATION
- A61B2560/00—Constructional details of operational features of apparatus; Accessories for medical measuring apparatus
- A61B2560/04—Constructional details of apparatus
- A61B2560/0437—Trolley or cart-type apparatus
-
- A—HUMAN NECESSITIES
- A61—MEDICAL OR VETERINARY SCIENCE; HYGIENE
- A61B—DIAGNOSIS; SURGERY; IDENTIFICATION
- A61B2560/00—Constructional details of operational features of apparatus; Accessories for medical measuring apparatus
- A61B2560/04—Constructional details of apparatus
- A61B2560/0487—Special user inputs or interfaces
-
- A—HUMAN NECESSITIES
- A61—MEDICAL OR VETERINARY SCIENCE; HYGIENE
- A61B—DIAGNOSIS; SURGERY; IDENTIFICATION
- A61B5/00—Measuring for diagnostic purposes; Identification of persons
- A61B5/0002—Remote monitoring of patients using telemetry, e.g. transmission of vital signals via a communication network
- A61B5/0004—Remote monitoring of patients using telemetry, e.g. transmission of vital signals via a communication network characterised by the type of physiological signal transmitted
- A61B5/0013—Medical image data
-
- A—HUMAN NECESSITIES
- A61—MEDICAL OR VETERINARY SCIENCE; HYGIENE
- A61B—DIAGNOSIS; SURGERY; IDENTIFICATION
- A61B5/00—Measuring for diagnostic purposes; Identification of persons
- A61B5/0093—Detecting, measuring or recording by applying one single type of energy and measuring its conversion into another type of energy
- A61B5/0095—Detecting, measuring or recording by applying one single type of energy and measuring its conversion into another type of energy by applying light and detecting acoustic waves, i.e. photoacoustic measurements
-
- A—HUMAN NECESSITIES
- A61—MEDICAL OR VETERINARY SCIENCE; HYGIENE
- A61B—DIAGNOSIS; SURGERY; IDENTIFICATION
- A61B5/00—Measuring for diagnostic purposes; Identification of persons
- A61B5/103—Measuring devices for testing the shape, pattern, colour, size or movement of the body or parts thereof, for diagnostic purposes
- A61B5/107—Measuring physical dimensions, e.g. size of the entire body or parts thereof
- A61B5/1076—Measuring physical dimensions, e.g. size of the entire body or parts thereof for measuring dimensions inside body cavities, e.g. using catheters
-
- A—HUMAN NECESSITIES
- A61—MEDICAL OR VETERINARY SCIENCE; HYGIENE
- A61B—DIAGNOSIS; SURGERY; IDENTIFICATION
- A61B6/00—Apparatus or devices for radiation diagnosis; Apparatus or devices for radiation diagnosis combined with radiation therapy equipment
- A61B6/56—Details of data transmission or power supply, e.g. use of slip rings
-
- A—HUMAN NECESSITIES
- A61—MEDICAL OR VETERINARY SCIENCE; HYGIENE
- A61B—DIAGNOSIS; SURGERY; IDENTIFICATION
- A61B8/00—Diagnosis using ultrasonic, sonic or infrasonic waves
- A61B8/46—Ultrasonic, sonic or infrasonic diagnostic devices with special arrangements for interfacing with the operator or the patient
- A61B8/467—Ultrasonic, sonic or infrasonic diagnostic devices with special arrangements for interfacing with the operator or the patient characterised by special input means
-
- A—HUMAN NECESSITIES
- A61—MEDICAL OR VETERINARY SCIENCE; HYGIENE
- A61B—DIAGNOSIS; SURGERY; IDENTIFICATION
- A61B8/00—Diagnosis using ultrasonic, sonic or infrasonic waves
- A61B8/56—Details of data transmission or power supply
-
- A—HUMAN NECESSITIES
- A61—MEDICAL OR VETERINARY SCIENCE; HYGIENE
- A61B—DIAGNOSIS; SURGERY; IDENTIFICATION
- A61B8/00—Diagnosis using ultrasonic, sonic or infrasonic waves
- A61B8/56—Details of data transmission or power supply
- A61B8/565—Details of data transmission or power supply involving data transmission via a network
Definitions
- Embodiments of the present disclosure relate generally to the field of medical devices and, more particularly, to a medical workflow system and associated methods of use.
- IVUS intravascular ultrasound
- FL-IVUS forward looking IVUS
- FFR fractional flow reserve
- CFR coronary flow reserve
- OCT optical coherence tomography
- trans-esophageal echocardiography and image-guided therapy.
- image-guided therapy many of these procedures are carried out by a multitude of physicians and clinicians, where each performs an assigned task. For example, a physician may stand next to a patient in the sterile field and guide the insertion and pull back of an imaging catheter. A clinician near the physician may control the procedure workflow with a controller, for example by starting and stopping the capture of images.
- a second clinician in an adjacent control room working at a desktop computer may select the images of interest and make measurements on them.
- the physician in the catheter lab must instruct the clinician in the control room on how to make such measurements. This may lengthen the time of the procedure, increase the cost of the procedure, and may lead to measurement errors due to miscommunication or clinician inexperience.
- a clinician may typically have to select a measurement mode prior to making any measurements, reducing the efficiency of the medical sensing workflow.
- the present disclosure is directed to a method of conducting a medical workflow with a touch-sensitive bedside controller.
- the method includes initiating a medical workflow using a graphical user interface on the bedside controller, positioning an imaging tool within a patient's body based on images captured by the imaging tools and displayed on the bedside controller, controlling the commencement and termination of a recordation of images captured by the imaging tool using the graphical user interface on the bedside controller, navigating through the recorded images to identify an image of interest using the graphical user interface on the bedside controller, and performing measurements on the image of interest using the graphical user interface on the bedside controller.
- the performing measurements may include touching and releasing portions of the image of interest as it is displayed on the bedside controller to make one of an area measurement and a diameter measurement.
- the present disclosure is directed to a bedside controller.
- the bedside controller include a housing, a touch-sensitive display disposed within a surface of the housing and configured to display images and receive user input on the surface, and a processor disposed within the housing.
- the bedside controller also includes a communication module disposed within the housing, communicatively coupled to the processor, and configured to transmit and receive medical data and a non-transitory computer readable storage module disposed within the housing, communicatively coupled to the processor, and including a plurality of instructions stored therein and executable by the processor.
- the plurality of instructions include instructions for rendering a graphical user interface (GUI) on the touch-sensitive display, instructions for displaying images of a patient as they are being captured by an imaging tool disposed within the patient's body, and instructions for initiating and terminating a recordation of the images based on user input to the GUI.
- the plurality of instructions also include instructions for displaying the recorded images within the GUI so that a user may identify an image of interest and instructions for making a measurement on the image of interest based on a user measurement input to the GUI.
- FIG. 1 is a schematic drawing depicting a medical sensing system including a bedside controller according to one embodiment of the present disclosure.
- FIG. 2 is a schematic drawing depicting a medical sensing system including a wireless bedside controller according to another embodiment of the present disclosure.
- FIG. 3A is a diagrammatic perspective view of a bedside controller.
- FIG. 3B is a diagrammatic rear perspective view of the bedside controller of FIG. 3A .
- FIG. 3C is a diagrammatic perspective view of the bedside controller of FIG. 3A mounted to a bed rail.
- FIG. 4 is a functional block diagram of the bedside controller of FIGS. 3A and 3B according to aspects of the present disclosure.
- FIG. 5 is a diagrammatic perspective view of a multi-modality mobile processing system with the bedside controller of FIGS. 3A and 3B attached thereto.
- FIG. 6 is a diagrammatic perspective view of the bedside controller of FIGS. 3A and 3B releasably mounted on an IV pole.
- FIG. 7 is a high-level flowchart illustrating a method of conducting a medical sensing workflow with a bedside controller according to various aspects of the present disclosure.
- FIG. 8 is high-level flowchart of a method that describes a measurement workflow conducted on a bedside controller according to various aspects of the present disclosure.
- FIGS. 9-11 are partial screen images illustrating various aspects of the method of FIG. 8 .
- FIG. 1 is a schematic drawing depicting a medical sensing system 100 including a bedside controller 102 according to one embodiment of the present disclosure.
- the medical sensing system 100 provides for coherent integration and consolidation of multiple forms of acquisition and processing elements designed to be sensitive to a variety of methods used to acquire and interpret human biological physiology and morphological information.
- the bedside controller 102 is a touch-enabled, integrated computing device for the acquisition, control, interpretation, measurement, and display of multi-modality medical sensing data.
- the bedside controller 102 is a tablet-style touch-sensitive computer that provides user controls and diagnostic images on a single surface.
- the bedside controller 102 is operable to present workflow control options and patient image data via graphical user interfaces (GUIs) corresponding to a plurality of medical sensing modalities.
- GUIs graphical user interfaces
- the bedside controller 102 will be described in greater detail in association with FIGS. 3A , 3 B, and 4 .
- the medical sensing system 100 is deployed in a catheter lab 104 .
- the catheter lab 104 may be used to perform on a patient 106 any number of medical sensing procedures alone or in combination such as, by way of example and not limitation, angiography, intravascular ultrasound (IVUS), virtual histology (VH), forward looking IVUS (FL-IVUS), intravascular photoacoustic (IVPA) imaging, fractional flow reserve (FFR) determination, coronary flow reserve (CFR) determination, optical coherence tomography (OCT), computed tomography, intracardiac echocardiography (ICE), forward-looking ICE (FLICE), intravascular palpography, transesophageal ultrasound, or any other medical sensing modalities known in the art.
- IVUS intravascular ultrasound
- VH virtual histology
- FL-IVUS forward looking IVUS
- IVPA intravascular photoacoustic
- FFR fractional flow reserve
- CFR coronary flow reserve
- OCT optical coherence tomography
- computed tomography
- the bedside controller may be used to cooperate with and control medical treatment systems such as, for example but without limitation, those used for stent placement, coil embolism, ablation therapy, kidney stone treatments, basket placement in a cystoscopy, tumor removal, and chemical therapies.
- the catheter lab 104 further includes a sterile field 105 that encompasses the portions of the catheter lab surrounding the patient 106 on a procedure table 109 and a clinician 107 , who may perform any number of medical sensing procedures or treatments.
- the bedside controller 102 may be positioned within the sterile field 105 and may be utilized by the clinician 107 to control a workflow of a medical sensing procedure or treatment being performed on the patient 106 .
- the clinician 107 may initiate the procedure workflow, watch real-time IVUS images captured during the procedure, and make measurements on the IVUS images all using the bedside controller 102 inside of the sterile field 105 .
- the bedside controller 102 may be utilized outside of the sterile field 105 , for instance, in other locations within the catheter lab 104 or in a control room adjacent to the catheter lab. A method of utilizing the bedside controller 102 to control a medical sensing workflow or treatment workflow will be discussed in greater detail in association with FIGS. 7 and 8 .
- the medical sensing system 100 additionally includes a number of interconnected medical sensing-related tools in the catheter lab 104 to facilitate a multi-modality workflow procedure, such as an IVUS catheter 108 , an IVUS patient isolation module (PIM) 112 , an OCT catheter 110 , and OCT PIM 114 , an electrocardiogram (ECG) device 116 , an angiogram system 117 , a boom display 122 , and a multi-modality processing system 124 .
- the bedside controller 102 , PIMs 112 and 114 , ECG device 116 , angiography system 117 , and boom display 122 are communicatively coupled to the processing system 124 .
- the processing system 124 is a computer workstation with the hardware and software to acquire, process, and display multi-modality medical sensing data, but in other embodiments, the processing system may be any other type of computing system operable to process medical sensing data.
- the processing system 124 is operable to accept raw IVUS data from the IVUS PIM 112 , transform it into IVUS images, and make the images available to the bedside controller 124 , so that they may be displayed to the clinician 107 for analysis.
- the processing system 124 is a computer workstation
- the system includes at least a processor such as a microcontroller or a dedicated central processing unit (CPU), a non-transitory computer-readable storage medium such as a hard drive, random access memory (RAM), and/or compact disk read only memory (CD-ROM), a video controller such as a graphics processing unit (GPU), and a network communication device such as an Ethernet controller.
- the multi-modality processing system 124 is communicatively coupled to a data network 125 .
- the data network 125 is a TCP/IP-based local area network (LAN), however in other embodiments, it may utilize a different protocol such as Synchronous Optical Networking (SONET), or may be a wide area network (WAN).
- SONET Synchronous Optical Networking
- the processing system 124 may connect to various resources via the network 125 , such as a Digital Imaging and Communications in Medicine (DICOM) system, a Picture Archiving and Communication System (PACS), and a Hospital Information System.
- DICOM Digital Imaging and Communications in Medicine
- PES Picture Archiving and Communication System
- U.S. Patent Application No. 61/473,570 entitled “MULTI-MODALITY MEDICAL SENSING SYSTEM AND METHOD” and filed on Apr. 8, 2011, discloses a multi-modality processing system that processes medical sensing data and is hereby incorporated by reference in its entirety.
- the IVUS PIM 112 and OCT PIM 114 are operable to respectively receive medical sensing data collected from the patient 106 by the IVUS catheter 108 and OCT catheter 110 and are operable to transmit the received data to the processing system 124 .
- the IVUS PIM 112 and OCT PIM 114 transmit the medical sensing data over a Peripheral Component Interconnect Express (PCIe) data bus connection, but, in other embodiments, they may transmit data over a USB connection, a Thunderbolt connection, a FireWire connection, or some other high-speed data bus connection.
- the ECG device 116 is operable to transmit electrocardiogram signals or other hemodynamic data from patient 106 to the processing system 124 .
- the bedside controller 102 is operable to display the ECG data along side medical sensing data.
- the processing system 124 may be operable to synchronize data collection with the catheters 108 and 110 using ECG signals from the ECG 116 .
- the angiogram system 117 is operable to collect x-ray, computed tomography (CT), or magnetic resonance images (MRI) of the patient 106 and transmit them to the processing system 124 .
- CT computed tomography
- MRI magnetic resonance images
- the processing system 124 may co-register image data from angiogram system 117 (e.g. x-ray data, MRI data, CT data, etc.) with sensing data from the IVUS and OCT catheters 108 and 110 .
- the co-registration may be performed to generate three-dimensional images with the sensing data.
- Such co-registered 3-D images data may be viewable on the bedside controller 124 .
- a clinician may rotate, zoom, and otherwise manipulate such 3-D images on the bedside controller 102 using simultaneous touch inputs (i.e. multitouch) and gestures.
- medical sensing tools in system 100 are communicatively coupled to the processing system 124 via a wired connection such as a standard copper link or a fiber optic link.
- the bedside controller 124 may be communicatively and/or electrically coupled to the processing system 124 via a Universal Serial Bus (USB) connection, a Power-over-Ethernet connection, a Thunderbolt connection, a FireWire connection, or some other high-speed data bus connection.
- USB Universal Serial Bus
- FIG. 2 is a schematic drawing depicting a medical sensing system 200 including a wireless bedside controller 202 according to another embodiment of the present disclosure.
- the medical sensing system 200 is similar to the system 100 of FIG. 1 but the medical sensing tools including the wireless bedside controller 202 , a wireless IVUS PIM 204 , and a wireless OCT PIM 206 communicate with a wireless network 208 via wireless networking protocols.
- the bedside controller 202 may send and receive workflow control parameters, medical sensing images, and measurement data to and from a remote processing system via IEEE 802.11 Wi-Fi standards, Ultra Wide-Band (UWB) standards, wireless FireWire, wireless USB, Bluetooth, or another high-speed wireless networking standard.
- IEEE 802.11 Wi-Fi standards Ultra Wide-Band (UWB) standards
- wireless FireWire wireless USB
- Bluetooth or another high-speed wireless networking standard.
- wireless capability allows the clinician 107 to more freely position the bedside controller 202 inside or outside of the sterile field 105 for better workflow management.
- FIG. 3A is a diagrammatic perspective view of a bedside controller 300
- FIG. 3B is a diagrammatic rear perspective view of the bedside controller
- FIG. 3C is a diagrammatic perspective view of the bedside controller mounted to a bed rail
- FIG. 4 is a functional block diagram of the bedside controller 300 according to aspects of the present disclosure.
- the bedside controller 300 is similar to the bedside controllers 102 and 202 in medical sensing systems 100 and 200 , and is operable to, among other things, initiate a medical sensing or treatment procedure workflow, display real-time images captured during the procedure, accept measurement input on the images from a clinician.
- the bedside controller 300 generally improves system control available to a clinician working at a patient table. For instance, giving a clinician both workflow control and measurement capability within the sterile field reduces errors and improves workflow efficiency.
- the bedside controller 300 includes an integrally formed housing 302 that is easy to grasp and move around a catheter lab or other medical setting.
- the integrally formed housing 302 may be seamlessly molded from materials such as thermoplastic or thermosetting plastic or moldable metal.
- the integrally formed housing 302 may comprise a plurality of housing portions fixedly bonded in a substantially permanent manner to form an integral housing.
- the housing 302 is resistant to fluids, and, in one embodiment, may have a rating of IPX4 against fluid ingress as defined by the International Electrotechnical Commission (IEC) standard 60529.
- the hub may have a different fluid ingress rating.
- the housing 302 is about 10.5 inches in width, about 8.25 inches in height, and has as thickness of about 2.75 inches.
- the housing may have a different width, height, or thickness that is similarly conducive to portability.
- the housing 302 further includes self-contained mounting structure 303 disposed on the housing.
- the mounting structure is disposed near an outer edge of the housing.
- the mounting structure 303 allows the bedside controller 300 to be releasably mounted in a variety of places in and out of a catheter lab in a self-contained manner. That is, the bedside controller 300 may be directly secured to another object without the use of a separate external mount.
- the mounting structure 303 includes a mounting channel 304 and a retaining clamp 305 that pivots over the mounting channel to secure a mounting platform therewithin.
- the mounting channel 304 is defined by a longer front wall 350 , a top wall 352 , and a shorter back wall 354 , and the retaining clamp includes a slot 356 that extends through the clamp in a manner generally parallel to the mounting channel.
- the front wall 350 and the back wall 354 are generally perpendicular to a touch-sensitive display 307 in the housing 302
- the top wall 352 is generally parallel to the display 307 .
- the retaining clamp is spring-loaded and releasably exerts pressure on objects situated in the mounting channel.
- the retaining clamp may be configured differently and exert force via mechanisms other than springs.
- the bedside controller 300 may be releasably secured to a mounting platform, for example a bed rail 306 , by pivoting the mounting clamp 305 to an open position, positioning the controller such that the rail extends through the length of the channel 304 , and releasing the clamp such that it secures the rail within the channel.
- a mounting platform for example a bed rail 306
- the mounting clamp 305 is pivoting the mounting clamp 305 to an open position, positioning the controller such that the rail extends through the length of the channel 304 , and releasing the clamp such that it secures the rail within the channel.
- the rail 306 is positioned in the mounting channel 304 and the clamp 305 is holding it therein, three surfaces of the rail are respectively engaged by the front wall 350 , the top wall 352 , and the back wall 354 , and a fourth surface of the rail extends through the slot 356 in the clamp 305 .
- the mounting structure 303 may maintain the bedside controller 300 in a position generally parallel to a procedure table 350 associated with the bed rail 306 , as shown in FIG. 3B .
- the mounting structure 303 is a cantilevered mounting structure in that it secures one end of the controller to an object while the majority of the controller extends away from the object in an unsupported manner. Such a cantilevered position allows for a display of the controller to be both readable and at a comfortable input angle for an operator.
- the self-contained mounting structure 303 allows the bedside controller 300 to be quickly released from the bed rail 306 and reattached to an IV pole, a cart on which a processing system is deployed, or other location in or out of the sterile field to allow for convenient workflow control and image analysis.
- the mounting structure 303 of the bedside controller may vary from the design illustrated in FIGS. 3A and 3B and include additional and/or different components to allow for self-contained mounting.
- the touch-sensitive display 307 Embedded into the front of the housing 302 is the touch-sensitive display 307 that comprises both a touch panel 308 and a flat panel display 309 .
- the touch panel 308 overlays the flat panel display 308 and accepts user input via human touch, stylus touch, or some other analogous input method.
- the touch-sensitive display 307 displays images and accepts user input on the same surface.
- the touch panel 308 is a resistive-type panel, but in alternative embodiments it may be a capacitive-type panel, projective-type panel, or some other suitable type of touch enabled input panel.
- the touch panel 308 is operable to accept multiple inputs simultaneously (multitouch), for instance, to enable rotation of a three-dimensional rendering of a vessel along multiple axes.
- the touch panel 308 is capable of receiving input when a sterile drape 301 is covering the bedside controller 300 and also when a user is gloved.
- the touch panel 308 is controlled by a touch controller 310 disposed within the housing 302 .
- the touch panel is operable to provide haptic feedback via a haptics controller 312 and haptics drivers 314 .
- This haptic technology is operable to simulate a plurality of sensations on the touch panel 308 by varying the intensity and frequency of vibrations generated when a user contacts the touch panel.
- the housing 302 may include a sheath configured to store a stylus therein. Thus, a clinician may remove the stylus from the sheath in the housing to make measurements on the bedside controller and store it when the measurements have been completed.
- the flat panel display 309 that presents a graphical user interface (GUI) 316 to a user.
- GUI graphical user interface
- the flat panel display 309 is a LCD display but in alternative embodiments, it may be a different type of display such an LED display or an AMOLED display.
- the flat panel display 309 is illuminated by a LED backlight power inverter 318 .
- the GUI 316 not only allows a clinician to control a medical sensing workflow but also make measurements on images captured from a patient in the sterile field. A method of interacting with the GUI 316 to make vessel measurements will be discussed in greater detail in association with FIGS. 8-11 .
- the bedside controller 300 includes a single board processing platform 320 within the housing 302 that is operable to render the GUI 316 and process user input.
- the processing platform has a pico form factor and includes integrated processing components such as a processor 321 , system memory 322 , graphics processing unit (GPU), communications module 323 , and I/O bus controller.
- the processor 321 may be a low power processor such as an Intel Atom® processor or a ARM-based processor
- the communications module 323 may be a 10/100/1 Gb Ethernet module.
- the I/O bus controller may be a Universal Serial Bus (USB) controller.
- USB Universal Serial Bus
- the bedside controller 300 further includes a storage module 324 that is a non-transitory computer readable storage medium operable to store an operating system (i.e. software to render and control the GUI), image manipulation software, medical sensing data and images received from a processing system, and other medical sensing-related software.
- the processor 321 is configured to execute software and instructions stored on the storage module 324 .
- the storage module 324 is a solid state drive (SSD) hard drive communicatively coupled to the processing platform 320 via a SATA connection, but, in alternative embodiments, it may be any other type of non-volatile or temporary storage module.
- the bedside controller 300 further includes a wireless communications module 326 communicatively coupled to the processing platform 320 .
- the wireless communications module is a IEEE 802.11 Wi-Fi module, but in other may be a Ultra Wide-Band (UWB) wireless module, a wireless FireWire module, a wireless USB module, a Bluetooth module, or another high-speed wireless networking module.
- UWB Ultra Wide-Band
- the bedside controller 300 is powered via both a wired 12 VDC power-over-Ethernet (PoE) connection 328 and a battery 330 disposed within the housing 302 .
- the battery 330 may be sealed within the integrally formed housing 302 and may be recharged through electrical contacts disposed on the exterior of the housing and electrically coupled to the battery.
- the front wall 350 may include one or more electrical contacts 358 through which the battery 330 may be charged when the controller is mounted to objects with compatible charging structure.
- the housing 302 may include a battery compartment with a removable cover to permit battery replacement. Such a battery compartment cover may be resistant to fluid ingress (e.g., with an IPX4 rating).
- the beside controller 300 may be coupled to a processing system in the catheter lab via the PoE connection 328 , over which it receives medical sensing images that have been captured from the patient and rendered on the processing system.
- a processing system in the catheter lab via the PoE connection 328 , over which it receives medical sensing images that have been captured from the patient and rendered on the processing system.
- the bedside controller when the bedside controller is coupled to the PoE connection 328 , it receives power and communications over the same physical wire.
- the bedside controller 300 is disconnected from the PoE connection 328 , it runs on battery power and receives data wirelessly via the wireless communications module 326 .
- the beside controller may directly communicate with a processing system (i.e. in an ad-hoc wireless mode), or, alternatively, it may communicate with a wireless network that serves a plurality of wireless devices.
- the bedside controller 300 may receive power and data through different wired connections, or receive data communications through a wired data connection and power from the battery 330 , or receive data communications through the wireless module 326 and power from a wired electrical connection.
- the bedside controller 300 may be used in a semi-wireless configuration, in which the battery 330 provides backup power to the controller when the controller is temporarily disconnected from a wired power source. For example, if at the beginning of a procedure, the bedside controller 300 is connected to a PoE connection (or other type of wired connection) and during the procedure the controller must be disconnected from the PoE connection to allow for a cabling adjustment, the battery 330 may keep the controller alive until a PoE connection can be re-established. In this manner, a full power-off and reboot of the controller 300 is avoided during a procedure. As shown in FIG. 4 , a DC-DC power converter 332 converts input voltage to a voltage usable by the processing platform 320 .
- the bedside controller 300 in the illustrated embodiments of FIGS. 3 and 4 includes specific components described herein, the bedside controller may include any number of additional components, for example a charge regulator interposed between the electrical contacts and the battery, and may be configured in any number of alternative arrangements in alternative embodiments.
- FIG. 5 is a diagrammatic perspective view of a multi-modality mobile processing system 500 .
- the processing system 500 is disposed on a cart 502 that enables the processing system to be easily moved between different locations such as different catheter labs.
- the bedside controller 300 is mounted to the cart 502 so that it may be transported to catheter labs with the processing system.
- the bedside controller 300 is releasably secured to the cart via the self-contained mounting structure 303 that is built into the housing 302 .
- the cart 502 may include a dock for the bedside controller 300 such that when the controller is docked on the cart its battery is recharged through the electrical contacts 358 disposed on the housing 302 .
- the bedside controller 300 may also releasably attach to an IV pole 600 via the self-contained mounting structure 303 . When so attached, the bedside controller 300 may be rolled next to a patient in the sterile field and thus within reach of a clinician who may operate the controller with a single hand.
- FIG. 7 is a high-level flowchart illustrating a method 700 of conducting a medical sensing workflow with the bedside controller 300 of FIGS. 3-4 according to various aspects of the present disclosure.
- the method 700 will be described in the context of an IVUS procedure but may equally apply to any number of medical sensing or treatment procedures, such as an OCT procedure, a FLIVUS procedure, an ICE procedure, etc.
- the method 700 begins at block 702 where a medical sensing workflow is initiated with the bedside controller 300 .
- a clinician in the sterile field and adjacent a patient may select the “IVUS” option out of a plurality of modes (e.g., OCT, Chromaflow, FLIVUS, etc) on the bedside controller's GUI to begin the IVUS workflow.
- a plurality of modes e.g., OCT, Chromaflow, FLIVUS, etc
- the clinician may select a ‘Live Images’ option on the bedside controller's GUI to receive live images from the catheter.
- the clinician may guide the catheter within the patient to a desired position.
- a processing system may collect raw IVUS data from the catheter and process the data to render IVUS images.
- the bedside controller retrieves the IVUS images from the processing system and displays them to a user in real-time. Then, in block 706 , after the IVUS catheter has been appropriately positioned in the patient using the live images, the clinician selects a ‘Record’ option on the bedside controller GUI and begins the catheter pull back. The processing system responds to the record command and begins rendering and storing IVUS images. The method 700 proceeds to block 708 where, after the IVUS catheter pull back has been completed, the clinician terminates the recording of IVUS images via the bedside controller's GUI. Then, in block 710 , the clinician at the bedside recalls the captured IVUS images on the bedside controller and finds the IVUS images associated with the area of interest.
- the bedside controller may present a condensed view of all captured images and the clinician may navigate through them using gestures on the bedside controller's touch panel to find the target area.
- the clinician performs measurements on the IVUS images directly on the bedside controller.
- the user of the bedside controller creates measurements by interacting with an image through a series of presses, moves and releases using a finger or stylus on the controller's touch-sensitive display. These actions are interpreted by the bedside controller's internal processor and converted to measurements on the display.
- the clinician may annotate the images using a stylus or another tool compatible with the bedside controller's touch panel.
- the clinician may save the images to the processing system by selecting the appropriate options in the bedside controller GUI. A method of performing measurements on the bedside controller will be described below.
- FIG. 8 is high-level flowchart of a method 800 that describes a measurement workflow on the bedside controller 300 of FIGS. 3A-4 .
- the method 800 may be carried out during block 720 of the method 700 in FIG. 7 as part of a medical sensing workflow on intravascular images.
- the method 800 of making measurements on the bedside controller 300 is implemented in measurement software stored in the storage module 324 in the bedside controller.
- a clinician when measuring images, such as intravascular images, a clinician has the option of making different types of measurements such as diameter measurements and area measurements. Typically, when making area measurements, a clinician may either denote the edges of an object by drawings a series of discrete points that are connected in subsequent processing or by drawing a continuous line around the object to the measured.
- the method 800 of performing measurements on images is “smart” in that it does not require a user to select a particular measurement mode prior to interacting with an image on the bedside controller. For instance, when a user performs a series of measurement inputs on the bedside controller, the GUI software interprets the nature (e.g. shape) of a user's measurement inputs, automatically enters either diameter mode, area-point mode or area-draw mode, and outputs the desired measurement on the controller's display.
- the nature e.g. shape
- the method 800 begins at block 802 where an image to be measured is displayed on the bedside controller and a user inputs a measurement start point on the image with an input device.
- the user may use a finger or stylus to indicate a point on a vessel border from which a measurement will commence. Note that prior to selecting the measurement start point, the measurement software did not require the user to select a measurement mode.
- the user without removing the input device from the image after indicating the start point, drags the input device across the image a distance to trace a line.
- the user withdraws the input device from the image at a measurement end point.
- the method 800 proceeds to decision block 808 where the measurement software determines whether the distance between the start point and the end point is less than a threshold value.
- the threshold value is equivalent to 10 pixels, but, in alternative embodiments, the threshold value may be smaller or larger or measured in different units. Further, in some embodiments, the threshold value is adjustable either manually by a user or automatically based on detected error rates. If the distance is less than the threshold value, the method proceeds to block 810 where the measurement software enters area-point mode and draws a point on the image corresponding to the end point (i.e. where the user lifted the input device from the touch-enabled display). This sequence is illustrated in FIG. 9 . Specifically, when a user presses ( 900 ) an input device on an image and immediately lifts ( 902 ) the input device, the input will be interpreted as a point entry and a point 904 will be drawn on the image.
- the method 800 then proceeds to decision block 812 where it is decided whether additional points are needed to make a measurement on the image. If additional points are needed, the method proceeds to block 814 where a user touches and releases the displayed image at a different location. Note that in this branch of method 800 , the measurement software is in area-point mode so that all entries will be interpreted as points and, when an input is detected, a point will be drawn on the image in block 810 regardless of the distance between a start point and end point of the input. If no additional points are needed to make a measurement in decision block 812 , the method 800 proceeds to block 816 , where a user selects a ‘Done’ button in the bedside controller GUI to exit area-point mode.
- the measurement software creates an area measurement using the entered points. For example, in an embodiment directed toward vessel measurement, the measurement software connects the entered points to create a bounding circle at the vessel's outer edge. In one embodiment, the measurement software uses the entered points as seed points to assist edge detection algorithms.
- the method 800 proceeds to decision block 820 where the measurement software determines whether the drawn line is “relatively straight”. That is, it determines whether the user desires to measure a diameter with a line or an area with an enclosed shape. As shown in FIG. 10 , to make such a determination, the measurement software compares intervening points on the traced line between a start point 1000 and an end point 1002 against a boundary threshold 1004 . If all intervening points are within the boundary threshold 1004 , the measurement software determines that the user desires to make a diameter measurement and transforms the traced line into a straight line 1006 extending from the start point to the end point.
- the diameter measurement is thus based on the length of the straight line 1006 .
- the measurement software may employ different methods for determining whether the user desires to make a diameter measurement or an area measurement, such as detecting whether intervening points between start and end points increase in distance from the start point before decreasing in distance from the start point or detecting whether the traced line extending through the start point, at least one intervening point, and the end point is arcuate past a threshold degree.
- the method proceeds to block 822 where the measurement software enters diameter mode and outputs a measurement of the straight line 1006 created between the start and end points.
- the method 800 proceeds to 818 where the measurement software enters area-draw mode. As shown in FIG. 11 , the traced line 1100 between start point 1102 and end point 1104 extends outside of a boundary threshold (not shown) and is thus not relatively straight, prompting the measurement software to enter area-draw mode. Once this determination is made, the software connects the start and ends points to create an unbroken bounding line 1006 from which an area may be calculated. After an area measurement has been made in block 818 (either in area-point mode or area-draw mode), the method proceeds to decision block 824 where it is determined if another measurement needs to be done. If so, the method proceeds back to block 802 where a user selects another start point on the image without first selecting a measurement mode. If all measurements have been completed, the method 800 ends.
- workflows for some medical sensing procedure may allow for additional measurement modes, such as volumetric measurements.
- a user may initiate any such additional measurement modes without first selecting a measurement mode, thus simplifying the workflow.
- the steps in methods 700 and 800 described above may be completed over the course of more than one patient visit to a catheter lab.
- the touch-enabled integrated bedside controllers 102 and 300 may be used to control and measure non-cardiovascular diagnostic data such as data from cranial or peripheral arteries, as well as data from non-vascular body portions. Further, the controllers 102 and 300 may be used to control an MRI workflow and measure MRI image data, or may be utilized in computer assisted surgery (CAS) applications. Further, the modules described above in association with the bedside controller 300 may be implemented in hardware, software, or a combination of both.
- the bedside controller may be designed to enable user control in many different network settings such as ad-hoc networks, local area networks, client-server networks, wide area networks, internets, and the controller may have a number of form factors such as a tablet, a smartphone, a laptop, or any other similar device. It is understood that such variations may be made in the foregoing without departing from the scope of the present disclosure. Accordingly, it is appropriate that the appended claims be construed broadly and in a manner consistent with the scope of the present disclosure.
Landscapes
- Health & Medical Sciences (AREA)
- Life Sciences & Earth Sciences (AREA)
- Engineering & Computer Science (AREA)
- Biomedical Technology (AREA)
- Physics & Mathematics (AREA)
- Public Health (AREA)
- General Health & Medical Sciences (AREA)
- Medical Informatics (AREA)
- Heart & Thoracic Surgery (AREA)
- Animal Behavior & Ethology (AREA)
- Veterinary Medicine (AREA)
- Biophysics (AREA)
- Pathology (AREA)
- Molecular Biology (AREA)
- Surgery (AREA)
- General Engineering & Computer Science (AREA)
- Theoretical Computer Science (AREA)
- Radiology & Medical Imaging (AREA)
- Nuclear Medicine, Radiotherapy & Molecular Imaging (AREA)
- Human Computer Interaction (AREA)
- General Physics & Mathematics (AREA)
- Vascular Medicine (AREA)
- Anesthesiology (AREA)
- Hematology (AREA)
- High Energy & Nuclear Physics (AREA)
- Oral & Maxillofacial Surgery (AREA)
- Dentistry (AREA)
- Optics & Photonics (AREA)
- Business, Economics & Management (AREA)
- General Business, Economics & Management (AREA)
- Epidemiology (AREA)
- Primary Health Care (AREA)
- Measuring And Recording Apparatus For Diagnosis (AREA)
- Ultra Sonic Daignosis Equipment (AREA)
- Acoustics & Sound (AREA)
Abstract
A method of conducting a medical workflow with a touch-sensitive bedside controller is disclosed. The method includes initiating a medical workflow using a graphical user interface on the bedside controller, positioning an imaging tool within a patient's body based on images captured by the imaging tools and displayed on the bedside controller, controlling the commencement and termination of a recordation of images captured by the imaging tool using the graphical user interface on the bedside controller, navigating through the recorded images to identify an image of interest using the graphical user interface on the bedside controller, and performing measurements on the image of interest using the graphical user interface on the bedside controller.
Description
- This application claims the benefit of U.S. provisional patent application 61/560,677, filed Nov. 16, 2011, entitled “MEDICAL SENSING CONTROL SYSTEM AND METHOD,” the entirety of which is incorporated by reference herein.
- Embodiments of the present disclosure relate generally to the field of medical devices and, more particularly, to a medical workflow system and associated methods of use.
- Innovations in diagnosing and verifying the level of success of treatment of disease have progressed from solely external imaging processes to include internal diagnostic processes. In addition to traditional external image techniques such as X-ray, MRI, CT scans, fluoroscopy, and angiography, small sensors may now be placed directly in the body. For example, diagnostic equipment and processes have been developed for diagnosing vasculature blockages and other vasculature disease by means of ultra-miniature sensors placed upon the distal end of a flexible elongate member such as a catheter, or a guide wire used for catheterization procedures. For example, known medical sensing techniques include intravascular ultrasound (IVUS), forward looking IVUS (FL-IVUS), fractional flow reserve (FFR) determination, a coronary flow reserve (CFR) determination, optical coherence tomography (OCT), trans-esophageal echocardiography, and image-guided therapy. Traditionally, many of these procedures are carried out by a multitude of physicians and clinicians, where each performs an assigned task. For example, a physician may stand next to a patient in the sterile field and guide the insertion and pull back of an imaging catheter. A clinician near the physician may control the procedure workflow with a controller, for example by starting and stopping the capture of images. Further, after images have been captured, a second clinician in an adjacent control room working at a desktop computer may select the images of interest and make measurements on them. Typically, the physician in the catheter lab must instruct the clinician in the control room on how to make such measurements. This may lengthen the time of the procedure, increase the cost of the procedure, and may lead to measurement errors due to miscommunication or clinician inexperience. Further, when making measurements on medical sensing images, a clinician may typically have to select a measurement mode prior to making any measurements, reducing the efficiency of the medical sensing workflow.
- Accordingly, while the existing devices and methods for conducting medical sensing workflows have been generally adequate for their intended purposes, they have not been entirely satisfactory in all respects.
- In one exemplary aspect, the present disclosure is directed to a method of conducting a medical workflow with a touch-sensitive bedside controller. The method includes initiating a medical workflow using a graphical user interface on the bedside controller, positioning an imaging tool within a patient's body based on images captured by the imaging tools and displayed on the bedside controller, controlling the commencement and termination of a recordation of images captured by the imaging tool using the graphical user interface on the bedside controller, navigating through the recorded images to identify an image of interest using the graphical user interface on the bedside controller, and performing measurements on the image of interest using the graphical user interface on the bedside controller.
- In some instances, the performing measurements may include touching and releasing portions of the image of interest as it is displayed on the bedside controller to make one of an area measurement and a diameter measurement.
- In another exemplary aspect, the present disclosure is directed to a bedside controller. The bedside controller include a housing, a touch-sensitive display disposed within a surface of the housing and configured to display images and receive user input on the surface, and a processor disposed within the housing. The bedside controller also includes a communication module disposed within the housing, communicatively coupled to the processor, and configured to transmit and receive medical data and a non-transitory computer readable storage module disposed within the housing, communicatively coupled to the processor, and including a plurality of instructions stored therein and executable by the processor. The plurality of instructions include instructions for rendering a graphical user interface (GUI) on the touch-sensitive display, instructions for displaying images of a patient as they are being captured by an imaging tool disposed within the patient's body, and instructions for initiating and terminating a recordation of the images based on user input to the GUI. The plurality of instructions also include instructions for displaying the recorded images within the GUI so that a user may identify an image of interest and instructions for making a measurement on the image of interest based on a user measurement input to the GUI.
-
FIG. 1 is a schematic drawing depicting a medical sensing system including a bedside controller according to one embodiment of the present disclosure. -
FIG. 2 is a schematic drawing depicting a medical sensing system including a wireless bedside controller according to another embodiment of the present disclosure. -
FIG. 3A is a diagrammatic perspective view of a bedside controller. -
FIG. 3B is a diagrammatic rear perspective view of the bedside controller ofFIG. 3A . -
FIG. 3C is a diagrammatic perspective view of the bedside controller ofFIG. 3A mounted to a bed rail. -
FIG. 4 is a functional block diagram of the bedside controller ofFIGS. 3A and 3B according to aspects of the present disclosure. -
FIG. 5 is a diagrammatic perspective view of a multi-modality mobile processing system with the bedside controller ofFIGS. 3A and 3B attached thereto. -
FIG. 6 is a diagrammatic perspective view of the bedside controller ofFIGS. 3A and 3B releasably mounted on an IV pole. -
FIG. 7 is a high-level flowchart illustrating a method of conducting a medical sensing workflow with a bedside controller according to various aspects of the present disclosure. -
FIG. 8 is high-level flowchart of a method that describes a measurement workflow conducted on a bedside controller according to various aspects of the present disclosure. -
FIGS. 9-11 are partial screen images illustrating various aspects of the method ofFIG. 8 . - For the purposes of promoting an understanding of the principles of the present disclosure, reference will now be made to the embodiments illustrated in the drawings, and specific language will be used to describe the same. It will nevertheless be understood that no limitation of the scope of the disclosure is intended. Any alterations and further modifications in the described devices, instruments, methods, and any further application of the principles of the disclosure as described herein are contemplated as would normally occur to one skilled in the art to which the disclosure relates. In particular, it is fully contemplated that the features, components, and/or steps described with respect to one embodiment may be combined with the features, components, and/or steps described with respect to other embodiments of the present disclosure.
-
FIG. 1 is a schematic drawing depicting amedical sensing system 100 including abedside controller 102 according to one embodiment of the present disclosure. In general, themedical sensing system 100 provides for coherent integration and consolidation of multiple forms of acquisition and processing elements designed to be sensitive to a variety of methods used to acquire and interpret human biological physiology and morphological information. More specifically, insystem 100, thebedside controller 102 is a touch-enabled, integrated computing device for the acquisition, control, interpretation, measurement, and display of multi-modality medical sensing data. In the illustrated embodiment, thebedside controller 102 is a tablet-style touch-sensitive computer that provides user controls and diagnostic images on a single surface. In themedical sensing system 100, thebedside controller 102 is operable to present workflow control options and patient image data via graphical user interfaces (GUIs) corresponding to a plurality of medical sensing modalities. Thebedside controller 102 will be described in greater detail in association withFIGS. 3A , 3B, and 4. - In the illustrated embodiment, the
medical sensing system 100 is deployed in acatheter lab 104. Thecatheter lab 104 may be used to perform on apatient 106 any number of medical sensing procedures alone or in combination such as, by way of example and not limitation, angiography, intravascular ultrasound (IVUS), virtual histology (VH), forward looking IVUS (FL-IVUS), intravascular photoacoustic (IVPA) imaging, fractional flow reserve (FFR) determination, coronary flow reserve (CFR) determination, optical coherence tomography (OCT), computed tomography, intracardiac echocardiography (ICE), forward-looking ICE (FLICE), intravascular palpography, transesophageal ultrasound, or any other medical sensing modalities known in the art. In addition to controlling medical sensing systems, the bedside controller may be used to cooperate with and control medical treatment systems such as, for example but without limitation, those used for stent placement, coil embolism, ablation therapy, kidney stone treatments, basket placement in a cystoscopy, tumor removal, and chemical therapies. Thecatheter lab 104 further includes asterile field 105 that encompasses the portions of the catheter lab surrounding thepatient 106 on a procedure table 109 and aclinician 107, who may perform any number of medical sensing procedures or treatments. As shown inFIG. 1 , thebedside controller 102 may be positioned within thesterile field 105 and may be utilized by theclinician 107 to control a workflow of a medical sensing procedure or treatment being performed on thepatient 106. For example, theclinician 107 may initiate the procedure workflow, watch real-time IVUS images captured during the procedure, and make measurements on the IVUS images all using thebedside controller 102 inside of thesterile field 105. In alternative embodiments, thebedside controller 102 may be utilized outside of thesterile field 105, for instance, in other locations within thecatheter lab 104 or in a control room adjacent to the catheter lab. A method of utilizing thebedside controller 102 to control a medical sensing workflow or treatment workflow will be discussed in greater detail in association withFIGS. 7 and 8 . - In the embodiment illustrated in
FIG. 1 , themedical sensing system 100 additionally includes a number of interconnected medical sensing-related tools in thecatheter lab 104 to facilitate a multi-modality workflow procedure, such as anIVUS catheter 108, an IVUS patient isolation module (PIM) 112, anOCT catheter 110, andOCT PIM 114, an electrocardiogram (ECG)device 116, anangiogram system 117, aboom display 122, and amulti-modality processing system 124. Thebedside controller 102,PIMs ECG device 116,angiography system 117, andboom display 122 are communicatively coupled to theprocessing system 124. In one embodiment, theprocessing system 124 is a computer workstation with the hardware and software to acquire, process, and display multi-modality medical sensing data, but in other embodiments, the processing system may be any other type of computing system operable to process medical sensing data. For example, during an IVUS workflow, theprocessing system 124 is operable to accept raw IVUS data from theIVUS PIM 112, transform it into IVUS images, and make the images available to thebedside controller 124, so that they may be displayed to theclinician 107 for analysis. In the embodiments in which theprocessing system 124 is a computer workstation, the system includes at least a processor such as a microcontroller or a dedicated central processing unit (CPU), a non-transitory computer-readable storage medium such as a hard drive, random access memory (RAM), and/or compact disk read only memory (CD-ROM), a video controller such as a graphics processing unit (GPU), and a network communication device such as an Ethernet controller. Further, themulti-modality processing system 124 is communicatively coupled to adata network 125. In the illustrated embodiment, thedata network 125 is a TCP/IP-based local area network (LAN), however in other embodiments, it may utilize a different protocol such as Synchronous Optical Networking (SONET), or may be a wide area network (WAN). Theprocessing system 124 may connect to various resources via thenetwork 125, such as a Digital Imaging and Communications in Medicine (DICOM) system, a Picture Archiving and Communication System (PACS), and a Hospital Information System. U.S. Patent Application No. 61/473,570, entitled “MULTI-MODALITY MEDICAL SENSING SYSTEM AND METHOD” and filed on Apr. 8, 2011, discloses a multi-modality processing system that processes medical sensing data and is hereby incorporated by reference in its entirety. - In the
medical sensing system 100, theIVUS PIM 112 andOCT PIM 114 are operable to respectively receive medical sensing data collected from thepatient 106 by theIVUS catheter 108 andOCT catheter 110 and are operable to transmit the received data to theprocessing system 124. In one embodiment, theIVUS PIM 112 andOCT PIM 114 transmit the medical sensing data over a Peripheral Component Interconnect Express (PCIe) data bus connection, but, in other embodiments, they may transmit data over a USB connection, a Thunderbolt connection, a FireWire connection, or some other high-speed data bus connection. Additionally, theECG device 116 is operable to transmit electrocardiogram signals or other hemodynamic data frompatient 106 to theprocessing system 124. To aid the clinician in data capture, thebedside controller 102 is operable to display the ECG data along side medical sensing data. Further, in some embodiments, theprocessing system 124 may be operable to synchronize data collection with thecatheters ECG 116. Further, theangiogram system 117 is operable to collect x-ray, computed tomography (CT), or magnetic resonance images (MRI) of thepatient 106 and transmit them to theprocessing system 124. After the x-ray, CT, or MRI data has been processed into human-readable images by theprocessing system 124, theclinician 107 may navigate the GUI on thebedside controller 124 to retrieve the images from theprocessing system 124 and display them on the controller. In some embodiments, theprocessing system 124 may co-register image data from angiogram system 117 (e.g. x-ray data, MRI data, CT data, etc.) with sensing data from the IVUS andOCT catheters bedside controller 124. In one embodiment, a clinician may rotate, zoom, and otherwise manipulate such 3-D images on thebedside controller 102 using simultaneous touch inputs (i.e. multitouch) and gestures. - Additionally, in the illustrated embodiment of
FIG. 1 , medical sensing tools insystem 100, are communicatively coupled to theprocessing system 124 via a wired connection such as a standard copper link or a fiber optic link. Specifically, thebedside controller 124 may be communicatively and/or electrically coupled to theprocessing system 124 via a Universal Serial Bus (USB) connection, a Power-over-Ethernet connection, a Thunderbolt connection, a FireWire connection, or some other high-speed data bus connection. - However, in an alternative embodiment, such as that shown in
FIG. 2 , the medical sensing tools may communicate wirelessly. In that regard,FIG. 2 is a schematic drawing depicting amedical sensing system 200 including awireless bedside controller 202 according to another embodiment of the present disclosure. Themedical sensing system 200 is similar to thesystem 100 ofFIG. 1 but the medical sensing tools including thewireless bedside controller 202, awireless IVUS PIM 204, and awireless OCT PIM 206 communicate with awireless network 208 via wireless networking protocols. For example, thebedside controller 202 may send and receive workflow control parameters, medical sensing images, and measurement data to and from a remote processing system via IEEE 802.11 Wi-Fi standards, Ultra Wide-Band (UWB) standards, wireless FireWire, wireless USB, Bluetooth, or another high-speed wireless networking standard. Such wireless capability allows theclinician 107 to more freely position thebedside controller 202 inside or outside of thesterile field 105 for better workflow management. - With reference now to
FIGS. 3A , 3B, 3C and 4,FIG. 3A is a diagrammatic perspective view of abedside controller 300,FIG. 3B is a diagrammatic rear perspective view of the bedside controller,FIG. 3C is a diagrammatic perspective view of the bedside controller mounted to a bed rail, andFIG. 4 is a functional block diagram of thebedside controller 300 according to aspects of the present disclosure. Thebedside controller 300 is similar to thebedside controllers medical sensing systems bedside controller 300 generally improves system control available to a clinician working at a patient table. For instance, giving a clinician both workflow control and measurement capability within the sterile field reduces errors and improves workflow efficiency. - As show in
FIG. 3A , thebedside controller 300 includes an integrally formedhousing 302 that is easy to grasp and move around a catheter lab or other medical setting. In one embodiment, the integrally formedhousing 302 may be seamlessly molded from materials such as thermoplastic or thermosetting plastic or moldable metal. In other embodiments, the integrally formedhousing 302 may comprise a plurality of housing portions fixedly bonded in a substantially permanent manner to form an integral housing. Thehousing 302 is resistant to fluids, and, in one embodiment, may have a rating of IPX4 against fluid ingress as defined by the International Electrotechnical Commission (IEC) standard 60529. In other embodiments in which thehousing 302 may be used in different environments, the hub may have a different fluid ingress rating. In the illustrated embodiment, thehousing 302 is about 10.5 inches in width, about 8.25 inches in height, and has as thickness of about 2.75 inches. In alternative embodiments, the housing may have a different width, height, or thickness that is similarly conducive to portability. - As shown in
FIG. 3B , thehousing 302 further includes self-containedmounting structure 303 disposed on the housing. In the illustrated embodiment, the mounting structure is disposed near an outer edge of the housing. The mountingstructure 303 allows thebedside controller 300 to be releasably mounted in a variety of places in and out of a catheter lab in a self-contained manner. That is, thebedside controller 300 may be directly secured to another object without the use of a separate external mount. In the illustrated embodiment, the mountingstructure 303 includes a mountingchannel 304 and a retainingclamp 305 that pivots over the mounting channel to secure a mounting platform therewithin. The mountingchannel 304 is defined by a longerfront wall 350, atop wall 352, and ashorter back wall 354, and the retaining clamp includes aslot 356 that extends through the clamp in a manner generally parallel to the mounting channel. Thefront wall 350 and theback wall 354 are generally perpendicular to a touch-sensitive display 307 in thehousing 302, and thetop wall 352 is generally parallel to thedisplay 307. In the illustrated embodiment, the retaining clamp is spring-loaded and releasably exerts pressure on objects situated in the mounting channel. In alternative embodiments, the retaining clamp may be configured differently and exert force via mechanisms other than springs. - As shown in
FIG. 3C , in operation, thebedside controller 300 may be releasably secured to a mounting platform, for example abed rail 306, by pivoting the mountingclamp 305 to an open position, positioning the controller such that the rail extends through the length of thechannel 304, and releasing the clamp such that it secures the rail within the channel. When therail 306 is positioned in the mountingchannel 304 and theclamp 305 is holding it therein, three surfaces of the rail are respectively engaged by thefront wall 350, thetop wall 352, and theback wall 354, and a fourth surface of the rail extends through theslot 356 in theclamp 305. In this manner, the mountingstructure 303 may maintain thebedside controller 300 in a position generally parallel to a procedure table 350 associated with thebed rail 306, as shown inFIG. 3B . Described differently, the mountingstructure 303 is a cantilevered mounting structure in that it secures one end of the controller to an object while the majority of the controller extends away from the object in an unsupported manner. Such a cantilevered position allows for a display of the controller to be both readable and at a comfortable input angle for an operator. Further, the self-containedmounting structure 303 allows thebedside controller 300 to be quickly released from thebed rail 306 and reattached to an IV pole, a cart on which a processing system is deployed, or other location in or out of the sterile field to allow for convenient workflow control and image analysis. In alternative embodiments the mountingstructure 303 of the bedside controller may vary from the design illustrated inFIGS. 3A and 3B and include additional and/or different components to allow for self-contained mounting. - Embedded into the front of the
housing 302 is the touch-sensitive display 307 that comprises both atouch panel 308 and aflat panel display 309. Thetouch panel 308 overlays theflat panel display 308 and accepts user input via human touch, stylus touch, or some other analogous input method. In other words, the touch-sensitive display 307 displays images and accepts user input on the same surface. In the current embodiment, thetouch panel 308 is a resistive-type panel, but in alternative embodiments it may be a capacitive-type panel, projective-type panel, or some other suitable type of touch enabled input panel. Further, thetouch panel 308 is operable to accept multiple inputs simultaneously (multitouch), for instance, to enable rotation of a three-dimensional rendering of a vessel along multiple axes. Additionally, thetouch panel 308 is capable of receiving input when asterile drape 301 is covering thebedside controller 300 and also when a user is gloved. Thetouch panel 308 is controlled by atouch controller 310 disposed within thehousing 302. Further, when a clinician makes contact with thetouch panel 308, the touch panel is operable to provide haptic feedback via ahaptics controller 312 andhaptics drivers 314. This haptic technology is operable to simulate a plurality of sensations on thetouch panel 308 by varying the intensity and frequency of vibrations generated when a user contacts the touch panel. In some embodiments, thehousing 302 may include a sheath configured to store a stylus therein. Thus, a clinician may remove the stylus from the sheath in the housing to make measurements on the bedside controller and store it when the measurements have been completed. - Beneath the
touch panel 308 is theflat panel display 309 that presents a graphical user interface (GUI) 316 to a user. In the illustrated embodiment, theflat panel display 309 is a LCD display but in alternative embodiments, it may be a different type of display such an LED display or an AMOLED display. In the illustrated embodiment, theflat panel display 309 is illuminated by a LEDbacklight power inverter 318. As mentioned above, theGUI 316 not only allows a clinician to control a medical sensing workflow but also make measurements on images captured from a patient in the sterile field. A method of interacting with theGUI 316 to make vessel measurements will be discussed in greater detail in association withFIGS. 8-11 . - The
bedside controller 300 includes a singleboard processing platform 320 within thehousing 302 that is operable to render theGUI 316 and process user input. In the illustrated embodiment, the processing platform has a pico form factor and includes integrated processing components such as aprocessor 321,system memory 322, graphics processing unit (GPU),communications module 323, and I/O bus controller. In some embodiments, theprocessor 321 may be a low power processor such as an Intel Atom® processor or a ARM-based processor, and thecommunications module 323 may be a 10/100/1 Gb Ethernet module. And, the I/O bus controller may be a Universal Serial Bus (USB) controller. Thebedside controller 300 further includes astorage module 324 that is a non-transitory computer readable storage medium operable to store an operating system (i.e. software to render and control the GUI), image manipulation software, medical sensing data and images received from a processing system, and other medical sensing-related software. Theprocessor 321 is configured to execute software and instructions stored on thestorage module 324. In the illustrated embodiment, thestorage module 324 is a solid state drive (SSD) hard drive communicatively coupled to theprocessing platform 320 via a SATA connection, but, in alternative embodiments, it may be any other type of non-volatile or temporary storage module. Thebedside controller 300 further includes awireless communications module 326 communicatively coupled to theprocessing platform 320. In some embodiments, the wireless communications module is a IEEE 802.11 Wi-Fi module, but in other may be a Ultra Wide-Band (UWB) wireless module, a wireless FireWire module, a wireless USB module, a Bluetooth module, or another high-speed wireless networking module. - In the illustrated embodiment, the
bedside controller 300 is powered via both a wired 12 VDC power-over-Ethernet (PoE)connection 328 and abattery 330 disposed within thehousing 302. In one embodiment, thebattery 330 may be sealed within the integrally formedhousing 302 and may be recharged through electrical contacts disposed on the exterior of the housing and electrically coupled to the battery. As shown in the embodiment ofFIG. 3B , thefront wall 350 may include one or moreelectrical contacts 358 through which thebattery 330 may be charged when the controller is mounted to objects with compatible charging structure. In other embodiments, thehousing 302 may include a battery compartment with a removable cover to permit battery replacement. Such a battery compartment cover may be resistant to fluid ingress (e.g., with an IPX4 rating). The besidecontroller 300 may be coupled to a processing system in the catheter lab via thePoE connection 328, over which it receives medical sensing images that have been captured from the patient and rendered on the processing system. In operation, when the bedside controller is coupled to thePoE connection 328, it receives power and communications over the same physical wire. When thebedside controller 300 is disconnected from thePoE connection 328, it runs on battery power and receives data wirelessly via thewireless communications module 326. When used wirelessly in a catheter lab, the beside controller may directly communicate with a processing system (i.e. in an ad-hoc wireless mode), or, alternatively, it may communicate with a wireless network that serves a plurality of wireless devices. In alternative embodiments, thebedside controller 300 may receive power and data through different wired connections, or receive data communications through a wired data connection and power from thebattery 330, or receive data communications through thewireless module 326 and power from a wired electrical connection. In some embodiments, thebedside controller 300 may be used in a semi-wireless configuration, in which thebattery 330 provides backup power to the controller when the controller is temporarily disconnected from a wired power source. For example, if at the beginning of a procedure, thebedside controller 300 is connected to a PoE connection (or other type of wired connection) and during the procedure the controller must be disconnected from the PoE connection to allow for a cabling adjustment, thebattery 330 may keep the controller alive until a PoE connection can be re-established. In this manner, a full power-off and reboot of thecontroller 300 is avoided during a procedure. As shown inFIG. 4 , a DC-DC power converter 332 converts input voltage to a voltage usable by theprocessing platform 320. - It is understood that although the
bedside controller 300 in the illustrated embodiments ofFIGS. 3 and 4 includes specific components described herein, the bedside controller may include any number of additional components, for example a charge regulator interposed between the electrical contacts and the battery, and may be configured in any number of alternative arrangements in alternative embodiments. - With reference now to
FIGS. 5 and 6 , illustrated are examples of locations in which thebedside controller 300 may be mounted.FIG. 5 is a diagrammatic perspective view of a multi-modalitymobile processing system 500. Theprocessing system 500 is disposed on acart 502 that enables the processing system to be easily moved between different locations such as different catheter labs. As shown inFIG. 5 , thebedside controller 300 is mounted to thecart 502 so that it may be transported to catheter labs with the processing system. Thebedside controller 300 is releasably secured to the cart via the self-containedmounting structure 303 that is built into thehousing 302. Further, in some embodiments, thecart 502 may include a dock for thebedside controller 300 such that when the controller is docked on the cart its battery is recharged through theelectrical contacts 358 disposed on thehousing 302. As shown inFIG. 6 , thebedside controller 300 may also releasably attach to anIV pole 600 via the self-containedmounting structure 303. When so attached, thebedside controller 300 may be rolled next to a patient in the sterile field and thus within reach of a clinician who may operate the controller with a single hand. -
FIG. 7 is a high-level flowchart illustrating amethod 700 of conducting a medical sensing workflow with thebedside controller 300 ofFIGS. 3-4 according to various aspects of the present disclosure. Themethod 700 will be described in the context of an IVUS procedure but may equally apply to any number of medical sensing or treatment procedures, such as an OCT procedure, a FLIVUS procedure, an ICE procedure, etc. Themethod 700 begins atblock 702 where a medical sensing workflow is initiated with thebedside controller 300. Using an IVUS procedure as an example, a clinician in the sterile field and adjacent a patient may select the “IVUS” option out of a plurality of modes (e.g., OCT, Chromaflow, FLIVUS, etc) on the bedside controller's GUI to begin the IVUS workflow. Next, inblock 704, after an IVUS imaging catheter has been inserted into the patient, the clinician may select a ‘Live Images’ option on the bedside controller's GUI to receive live images from the catheter. Using the real-time images, the clinician may guide the catheter within the patient to a desired position. In typical embodiments, a processing system may collect raw IVUS data from the catheter and process the data to render IVUS images. The bedside controller retrieves the IVUS images from the processing system and displays them to a user in real-time. Then, inblock 706, after the IVUS catheter has been appropriately positioned in the patient using the live images, the clinician selects a ‘Record’ option on the bedside controller GUI and begins the catheter pull back. The processing system responds to the record command and begins rendering and storing IVUS images. Themethod 700 proceeds to block 708 where, after the IVUS catheter pull back has been completed, the clinician terminates the recording of IVUS images via the bedside controller's GUI. Then, inblock 710, the clinician at the bedside recalls the captured IVUS images on the bedside controller and finds the IVUS images associated with the area of interest. Specifically, the bedside controller may present a condensed view of all captured images and the clinician may navigate through them using gestures on the bedside controller's touch panel to find the target area. Finally, inblock 720, the clinician performs measurements on the IVUS images directly on the bedside controller. The user of the bedside controller creates measurements by interacting with an image through a series of presses, moves and releases using a finger or stylus on the controller's touch-sensitive display. These actions are interpreted by the bedside controller's internal processor and converted to measurements on the display. For precise measurements, the clinician may annotate the images using a stylus or another tool compatible with the bedside controller's touch panel. After the appropriate measurements have been completed, the clinician may save the images to the processing system by selecting the appropriate options in the bedside controller GUI. A method of performing measurements on the bedside controller will be described below. -
FIG. 8 is high-level flowchart of amethod 800 that describes a measurement workflow on thebedside controller 300 ofFIGS. 3A-4 . In one embodiment, themethod 800 may be carried out duringblock 720 of themethod 700 inFIG. 7 as part of a medical sensing workflow on intravascular images. Further, in the illustrated embodiment, themethod 800 of making measurements on thebedside controller 300 is implemented in measurement software stored in thestorage module 324 in the bedside controller. In general, when measuring images, such as intravascular images, a clinician has the option of making different types of measurements such as diameter measurements and area measurements. Typically, when making area measurements, a clinician may either denote the edges of an object by drawings a series of discrete points that are connected in subsequent processing or by drawing a continuous line around the object to the measured. In this regard, themethod 800 of performing measurements on images is “smart” in that it does not require a user to select a particular measurement mode prior to interacting with an image on the bedside controller. For instance, when a user performs a series of measurement inputs on the bedside controller, the GUI software interprets the nature (e.g. shape) of a user's measurement inputs, automatically enters either diameter mode, area-point mode or area-draw mode, and outputs the desired measurement on the controller's display. - In more detail, the
method 800 begins atblock 802 where an image to be measured is displayed on the bedside controller and a user inputs a measurement start point on the image with an input device. For example, the user may use a finger or stylus to indicate a point on a vessel border from which a measurement will commence. Note that prior to selecting the measurement start point, the measurement software did not require the user to select a measurement mode. Next, inblock 804, the user, without removing the input device from the image after indicating the start point, drags the input device across the image a distance to trace a line. Then, inblock 806, the user withdraws the input device from the image at a measurement end point. Themethod 800 proceeds to decision block 808 where the measurement software determines whether the distance between the start point and the end point is less than a threshold value. In one embodiment, the threshold value is equivalent to 10 pixels, but, in alternative embodiments, the threshold value may be smaller or larger or measured in different units. Further, in some embodiments, the threshold value is adjustable either manually by a user or automatically based on detected error rates. If the distance is less than the threshold value, the method proceeds to block 810 where the measurement software enters area-point mode and draws a point on the image corresponding to the end point (i.e. where the user lifted the input device from the touch-enabled display). This sequence is illustrated inFIG. 9 . Specifically, when a user presses (900) an input device on an image and immediately lifts (902) the input device, the input will be interpreted as a point entry and apoint 904 will be drawn on the image. - The
method 800 then proceeds to decision block 812 where it is decided whether additional points are needed to make a measurement on the image. If additional points are needed, the method proceeds to block 814 where a user touches and releases the displayed image at a different location. Note that in this branch ofmethod 800, the measurement software is in area-point mode so that all entries will be interpreted as points and, when an input is detected, a point will be drawn on the image inblock 810 regardless of the distance between a start point and end point of the input. If no additional points are needed to make a measurement indecision block 812, themethod 800 proceeds to block 816, where a user selects a ‘Done’ button in the bedside controller GUI to exit area-point mode. In block 818, the measurement software creates an area measurement using the entered points. For example, in an embodiment directed toward vessel measurement, the measurement software connects the entered points to create a bounding circle at the vessel's outer edge. In one embodiment, the measurement software uses the entered points as seed points to assist edge detection algorithms. - With reference back to decision block 808, if the distance between the start point and the end point is greater than or equal to the threshold, the
method 800 proceeds to decision block 820 where the measurement software determines whether the drawn line is “relatively straight”. That is, it determines whether the user desires to measure a diameter with a line or an area with an enclosed shape. As shown inFIG. 10 , to make such a determination, the measurement software compares intervening points on the traced line between astart point 1000 and anend point 1002 against aboundary threshold 1004. If all intervening points are within theboundary threshold 1004, the measurement software determines that the user desires to make a diameter measurement and transforms the traced line into astraight line 1006 extending from the start point to the end point. The diameter measurement is thus based on the length of thestraight line 1006. In alternative embodiments, however, the measurement software may employ different methods for determining whether the user desires to make a diameter measurement or an area measurement, such as detecting whether intervening points between start and end points increase in distance from the start point before decreasing in distance from the start point or detecting whether the traced line extending through the start point, at least one intervening point, and the end point is arcuate past a threshold degree. Atdecision block 820, if the user's traced line is relatively straight, the method proceeds to block 822 where the measurement software enters diameter mode and outputs a measurement of thestraight line 1006 created between the start and end points. If, however, the traced line is not relatively straight, themethod 800 proceeds to 818 where the measurement software enters area-draw mode. As shown inFIG. 11 , the tracedline 1100 betweenstart point 1102 andend point 1104 extends outside of a boundary threshold (not shown) and is thus not relatively straight, prompting the measurement software to enter area-draw mode. Once this determination is made, the software connects the start and ends points to create anunbroken bounding line 1006 from which an area may be calculated. After an area measurement has been made in block 818 (either in area-point mode or area-draw mode), the method proceeds to decision block 824 where it is determined if another measurement needs to be done. If so, the method proceeds back to block 802 where a user selects another start point on the image without first selecting a measurement mode. If all measurements have been completed, themethod 800 ends. - It is understood that the
methods FIGS. 7 and 8 may, in alternative embodiments, be performed in a different order and may include different and/or additional blocks in some embodiments. For example, workflows for some medical sensing procedure may allow for additional measurement modes, such as volumetric measurements. According to the described aspects of the present disclosure, a user may initiate any such additional measurement modes without first selecting a measurement mode, thus simplifying the workflow. Further, the steps inmethods - Although illustrative embodiments have been shown and described, a wide range of modification, change, and substitution is contemplated in the foregoing disclosure and in some instances, some features of the present disclosure may be employed without a corresponding use of the other features. For example, in some embodiments, the touch-enabled
integrated bedside controllers controllers bedside controller 300 may be implemented in hardware, software, or a combination of both. And the bedside controller may be designed to enable user control in many different network settings such as ad-hoc networks, local area networks, client-server networks, wide area networks, internets, and the controller may have a number of form factors such as a tablet, a smartphone, a laptop, or any other similar device. It is understood that such variations may be made in the foregoing without departing from the scope of the present disclosure. Accordingly, it is appropriate that the appended claims be construed broadly and in a manner consistent with the scope of the present disclosure.
Claims (30)
1. A method of conducting a medical workflow with a touch-sensitive bedside controller, the method comprising:
initiating a medical workflow using a graphical user interface on the bedside controller;
positioning an imaging tool within a patient's body based on images captured by the imaging tools and displayed on the bedside controller;
controlling the initiation and termination of a recordation of images by the imaging tool using the graphical user interface on the bedside controller;
navigating through the recorded images to identify an image of interest using the graphical user interface on the bedside controller; and
performing measurements on the image of interest using the graphical user interface on the bedside controller.
2. The method of claim 1 , wherein the initiating includes selecting the medical workflow out of a plurality of medical workflows presented as selectable options within the graphical user interface.
3. The method of claim 2 , wherein the plurality of medical workflows includes two or more of an intravascular ultrasound (IVUS) imaging workflow, an intravascular photoacoustic (IVPA) imaging workflow, an optical coherence tomography (OCT) workflow, a forward looking IVUS (FL-IVUS) workflow, a fractional flow reserve (FFR) workflow, a coronary flow reserve (CFR) workflow, and an angiography workflow.
4. The method of claim 1 , wherein the performing measurements includes touching and releasing portions of the image of interest as it is displayed on the bedside controller to make one of an area measurement and a diameter measurement.
5. The method of claim 4 , wherein the touching and releasing includes making one of the area measurement and the diameter measurement without first selecting a measurement mode corresponding one of the area measurement and the diameter measurement.
6. The method of claim 1 , wherein the initiating, positioning, controlling, navigating, and performing are performed on the bedside controller in a sterile field.
7. The method of claim 1 , wherein navigating through the recorded images includes navigating through a condensed view of all of the recorded images.
8. The method of claim 1 , wherein navigating through the recorded images includes performing navigation-type gestures on the touch-sensitive display.
9. The method of claim 1 , wherein performing measurements includes annotating the image of interest with the measurements.
10. The method of claim 9 , further including saving to a processing system the image of interest with the measurement annotations thereon.
11. The method of claim 1 , wherein the imaging tool is a sensor disposed on a catheter within a vessel of the patient.
12. The method of claim 11 , wherein the controlling includes coordinating the recordation with a pullback of the catheter.
13. The method of claim 1 , wherein controlling the initiation and termination includes selecting a record option in the graphical user interface to initiate the recordation.
14. The method of claim 1 , further including receiving haptic feedback in response to performing measurements on the image of interest using the graphical user interface.
15. The method of claim 1 , wherein the images captured by the imaging tool are associated with one of intravascular ultrasound (IVUS), intravascular photoacoustic (IVPA), optical coherence tomography (OCT), forward looking IVUS (FL-IVUS), fractional flow reserve (FFR), coronary flow reserve (CFR), and angiography.
16. A bedside controller, comprising:
a housing;
a touch-sensitive display disposed within a surface of the housing and configured to display images and receive user input on the surface;
a processor disposed within the housing;
a communication module disposed within the housing, communicatively coupled to the processor, and configured to transmit and receive medical data; and
a non-transitory computer readable storage module disposed within the housing, communicatively coupled to the processor, and including a plurality of instructions stored therein and executable by the processor, the plurality of instructions including:
instructions for rendering a graphical user interface (GUI) on the touch-sensitive display;
instructions for displaying images of a patient as they are being captured by an imaging tool disposed within the patient's body;
instructions for initiating and terminating a recordation of the images based on user input to the GUI;
instructions for displaying the recorded images within the GUI so that a user may identify an image of interest; and
instructions for making a measurement on the image of interest based on a user measurement input to the GUI.
17. The bedside controller of claim 16 , wherein the images of the patient are associated with one of intravascular ultrasound (IVUS), intravascular photoacoustic (IVPA), optical coherence tomography (OCT), forward looking IVUS (FL-IVUS), fractional flow reserve (FFR), coronary flow reserve (CFR), and angiography.
18. The bedside controller of claim 16 , wherein the plurality of instructions includes instructions for presenting a plurality of workflow modes selectable by a user through the GUI.
19. The bedside controller of claim 18 , wherein the plurality of instructions includes instructions for receiving a user selection of a workflow mode out of the plurality of workflow modes through the GUI.
20. The bedside controller of claim 18 , wherein the plurality of workflow modes includes two or more of an intravascular ultrasound (IVUS) imaging mode, an intravascular photoacoustic (IVPA) imaging mode, an optical coherence tomography (OCT) mode, a forward looking IVUS (FL-IVUS) mode, a fractional flow reserve (FFR) mode, a coronary flow reserve (CFR) mode, and an angiography mode.
21. The bedside controller of claim 16 , wherein the instructions for displaying the recorded images within the GUI includes instructions for displaying a condensed view of all of the recorded images.
22. The bedside controller of claim 16 , wherein the instructions for displaying the recorded images within the GUI includes instructions for receiving gesture inputs through the touch-sensitive display for navigation through the recorded images.
23. The bedside controller of claim 16 , wherein the plurality of instructions includes instructions for annotating the image of interest with the measurements.
24. The bedside controller of claim 23 , wherein the plurality of instructions includes instructions for saving to a processing system the image of interest with the measurement annotations thereon.
25. The bedside controller of claim 16 , wherein the imaging tool is a sensor disposed on a catheter within a vessel of the patient.
26. The bedside controller of claim 25 , wherein the instructions for initiating and terminating include instructions for coordinating the recordation with a pullback of the catheter.
27. The bedside controller of claim 16 , wherein the instructions for making a measurement include instructions for selecting one of an area measurement mode and a diameter measurement mode based on the user measurement input.
28. The bedside controller of claim 16 , wherein the instructions for rendering a GUI on the touch-sensitive display include instructions for providing haptic feedback in response to touch-based input to the GUI.
29. The bedside controller of claim 16 , wherein the instructions for displaying the recorded images include instructions for displaying a three-dimensional rendering of a portion of the patient's body.
30. The bedside controller of claim 29 , wherein the instructions for displaying a three-dimensional rendering include: instructions for receiving multiple concurrent touch inputs via the touch-sensitive display; and instructions for rotating the three-dimensional rendering about an axis based on the multiple concurrent touch inputs.
Priority Applications (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
US13/679,795 US20130123616A1 (en) | 2011-11-16 | 2012-11-16 | Medical Workflow System and Method |
Applications Claiming Priority (2)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
US201161560677P | 2011-11-16 | 2011-11-16 | |
US13/679,795 US20130123616A1 (en) | 2011-11-16 | 2012-11-16 | Medical Workflow System and Method |
Publications (1)
Publication Number | Publication Date |
---|---|
US20130123616A1 true US20130123616A1 (en) | 2013-05-16 |
Family
ID=48280119
Family Applications (3)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
US13/679,795 Abandoned US20130123616A1 (en) | 2011-11-16 | 2012-11-16 | Medical Workflow System and Method |
US13/679,752 Active US8681116B2 (en) | 2011-11-16 | 2012-11-16 | Medical mounting system and method |
US13/679,776 Active US8754865B2 (en) | 2011-11-16 | 2012-11-16 | Medical measuring system and method |
Family Applications After (2)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
US13/679,752 Active US8681116B2 (en) | 2011-11-16 | 2012-11-16 | Medical mounting system and method |
US13/679,776 Active US8754865B2 (en) | 2011-11-16 | 2012-11-16 | Medical measuring system and method |
Country Status (5)
Country | Link |
---|---|
US (3) | US20130123616A1 (en) |
EP (1) | EP2779907B1 (en) |
JP (1) | JP6373758B2 (en) |
CA (1) | CA2855830A1 (en) |
WO (1) | WO2013074800A1 (en) |
Cited By (55)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US8693756B2 (en) | 2007-03-08 | 2014-04-08 | Sync-Rx, Ltd. | Automatic reduction of interfering elements from an image stream of a moving organ |
US8700130B2 (en) | 2007-03-08 | 2014-04-15 | Sync-Rx, Ltd. | Stepwise advancement of a medical tool |
US20140180135A1 (en) * | 2012-12-21 | 2014-06-26 | Volcano Corporation | Chronic total occlusion catheter |
US20140181716A1 (en) * | 2012-12-26 | 2014-06-26 | Volcano Corporation | Gesture-Based Interface for a Multi-Modality Medical Imaging System |
US8855744B2 (en) | 2008-11-18 | 2014-10-07 | Sync-Rx, Ltd. | Displaying a device within an endoluminal image stack |
DE102013210230A1 (en) * | 2013-06-03 | 2014-10-16 | Siemens Aktiengesellschaft | Medical diagnostic and / or therapeutic device |
US20150169001A1 (en) * | 2013-12-18 | 2015-06-18 | Siemens Aktiengesellschaft | Medical facility |
US9095313B2 (en) | 2008-11-18 | 2015-08-04 | Sync-Rx, Ltd. | Accounting for non-uniform longitudinal motion during movement of an endoluminal imaging probe |
US9101286B2 (en) | 2008-11-18 | 2015-08-11 | Sync-Rx, Ltd. | Apparatus and methods for determining a dimension of a portion of a stack of endoluminal data points |
US9144394B2 (en) | 2008-11-18 | 2015-09-29 | Sync-Rx, Ltd. | Apparatus and methods for determining a plurality of local calibration factors for an image |
US20150351704A1 (en) * | 2009-12-21 | 2015-12-10 | Masimo Corporation | Modular patient monitor |
US9244590B1 (en) * | 2013-12-13 | 2016-01-26 | Amazon Technologies, Inc. | Three-dimensional navigation using a two-dimensional surface |
US9305334B2 (en) | 2007-03-08 | 2016-04-05 | Sync-Rx, Ltd. | Luminal background cleaning |
US9351698B2 (en) | 2013-03-12 | 2016-05-31 | Lightlab Imaging, Inc. | Vascular data processing and image registration systems, methods, and apparatuses |
US20160171160A1 (en) * | 2013-07-19 | 2016-06-16 | Ricoh Company, Ltd. | Healthcare system integration |
US9375164B2 (en) | 2007-03-08 | 2016-06-28 | Sync-Rx, Ltd. | Co-use of endoluminal data and extraluminal imaging |
US9610064B2 (en) | 2011-05-31 | 2017-04-04 | Desmond Adler | Multimodal imaging system, apparatus, and methods |
US9629571B2 (en) | 2007-03-08 | 2017-04-25 | Sync-Rx, Ltd. | Co-use of endoluminal data and extraluminal imaging |
US9788735B2 (en) | 2002-03-25 | 2017-10-17 | Masimo Corporation | Body worn mobile medical patient monitor |
US9855384B2 (en) | 2007-03-08 | 2018-01-02 | Sync-Rx, Ltd. | Automatic enhancement of an image stream of a moving organ and displaying as a movie |
US9888969B2 (en) | 2007-03-08 | 2018-02-13 | Sync-Rx Ltd. | Automatic quantitative vessel analysis |
US9913617B2 (en) | 2011-10-13 | 2018-03-13 | Masimo Corporation | Medical monitoring hub |
US9943269B2 (en) | 2011-10-13 | 2018-04-17 | Masimo Corporation | System for displaying medical monitoring data |
US9974509B2 (en) | 2008-11-18 | 2018-05-22 | Sync-Rx Ltd. | Image super enhancement |
US9996921B2 (en) | 2015-05-17 | 2018-06-12 | LIGHTLAB IMAGING, lNC. | Detection of metal stent struts |
US10105107B2 (en) | 2015-01-08 | 2018-10-23 | St. Jude Medical International Holding S.À R.L. | Medical system having combined and synergized data output from multiple independent inputs |
US10109058B2 (en) | 2015-05-17 | 2018-10-23 | Lightlab Imaging, Inc. | Intravascular imaging system interfaces and stent detection methods |
US10149616B2 (en) | 2012-02-09 | 2018-12-11 | Masimo Corporation | Wireless patient monitoring device |
US10172582B2 (en) | 2015-11-18 | 2019-01-08 | Lightlab Imaging, Inc. | X-ray image feature detection and registration systems and methods |
US10222956B2 (en) | 2015-05-17 | 2019-03-05 | Lightlab Imaging, Inc. | Intravascular imaging user interface systems and methods |
US10362962B2 (en) | 2008-11-18 | 2019-07-30 | Synx-Rx, Ltd. | Accounting for skipped imaging locations during movement of an endoluminal imaging probe |
US10409951B2 (en) * | 2012-12-28 | 2019-09-10 | Volcano Corporation | Multi-modality case management system and method |
US10453190B2 (en) | 2015-11-23 | 2019-10-22 | Lightlab Imaging, Inc. | Detection of and validation of shadows in intravascular images |
EP3563757A1 (en) * | 2018-05-03 | 2019-11-06 | Canon U.S.A. Inc. | Devices, systems, and methods to emphasize regions of interest across multiple imaging modalities |
US10499813B2 (en) | 2014-09-12 | 2019-12-10 | Lightlab Imaging, Inc. | Methods, systems and apparatus for temporal calibration of an intravascular imaging system |
US10593037B2 (en) | 2016-04-14 | 2020-03-17 | Lightlab Imaging, Inc. | Method, apparatus, and system to identify branches of a blood vessel |
US10617302B2 (en) | 2016-07-07 | 2020-04-14 | Masimo Corporation | Wearable pulse oximeter and respiration monitor |
US10631754B2 (en) | 2016-05-16 | 2020-04-28 | Lightlab Imaging, Inc. | Intravascular absorbable stent detection and diagnostic methods and systems |
US10646198B2 (en) | 2015-05-17 | 2020-05-12 | Lightlab Imaging, Inc. | Intravascular imaging and guide catheter detection methods and systems |
US10716528B2 (en) | 2007-03-08 | 2020-07-21 | Sync-Rx, Ltd. | Automatic display of previously-acquired endoluminal images |
US10748289B2 (en) | 2012-06-26 | 2020-08-18 | Sync-Rx, Ltd | Coregistration of endoluminal data points with values of a luminal-flow-related index |
US10825568B2 (en) | 2013-10-11 | 2020-11-03 | Masimo Corporation | Alarm notification system |
US10833983B2 (en) | 2012-09-20 | 2020-11-10 | Masimo Corporation | Intelligent medical escalation process |
US10912524B2 (en) | 2006-09-22 | 2021-02-09 | Masimo Corporation | Modular patient monitor |
US11064903B2 (en) | 2008-11-18 | 2021-07-20 | Sync-Rx, Ltd | Apparatus and methods for mapping a sequence of images to a roadmap image |
US11064964B2 (en) | 2007-03-08 | 2021-07-20 | Sync-Rx, Ltd | Determining a characteristic of a lumen by measuring velocity of a contrast agent |
US11109818B2 (en) | 2018-04-19 | 2021-09-07 | Masimo Corporation | Mobile patient alarm display |
US11197651B2 (en) | 2007-03-08 | 2021-12-14 | Sync-Rx, Ltd. | Identification and presentation of device-to-vessel relative motion |
US11287961B2 (en) | 2015-07-25 | 2022-03-29 | Lightlab Imaging, Inc. | Intravascular data visualization and interface systems and methods |
US11631496B2 (en) | 2013-09-12 | 2023-04-18 | Johnson & Johnson Surgical Vision, Inc. | Computer-based operating room support system |
US11715560B2 (en) | 2013-09-12 | 2023-08-01 | Johnson & Johnson Surgical Vision, Inc. | Computer-based operating room support system |
US11883107B2 (en) | 2016-09-28 | 2024-01-30 | Lightlab Imaging, Inc. | Stent planning systems and methods using vessel representation obtained via intravascular probe by determining stent effectiveness score and fractional flow reserve |
US11963736B2 (en) | 2009-07-20 | 2024-04-23 | Masimo Corporation | Wireless patient monitoring system |
WO2024222942A1 (en) * | 2023-04-28 | 2024-10-31 | Shanghai United Imaging Healthcare Co., Ltd. | Methods and systems for controlling medical device |
US12257022B2 (en) | 2018-10-12 | 2025-03-25 | Masimo Corporation | System for transmission of sensor data using dual communication protocol |
Families Citing this family (55)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
JP5452500B2 (en) | 2007-11-26 | 2014-03-26 | シー・アール・バード・インコーポレーテッド | Integrated system for intravascular placement of catheters |
US9521961B2 (en) | 2007-11-26 | 2016-12-20 | C. R. Bard, Inc. | Systems and methods for guiding a medical instrument |
US8781555B2 (en) | 2007-11-26 | 2014-07-15 | C. R. Bard, Inc. | System for placement of a catheter including a signal-generating stylet |
US9532724B2 (en) | 2009-06-12 | 2017-01-03 | Bard Access Systems, Inc. | Apparatus and method for catheter navigation using endovascular energy mapping |
EP2912999B1 (en) | 2010-05-28 | 2022-06-29 | C. R. Bard, Inc. | Apparatus for use with needle insertion guidance system |
US10210956B2 (en) | 2012-10-24 | 2019-02-19 | Cathworks Ltd. | Diagnostically useful results in real time |
WO2014064702A2 (en) | 2012-10-24 | 2014-05-01 | Cathworks Ltd. | Automated measurement system and method for coronary artery disease scoring |
US10799209B2 (en) * | 2012-12-26 | 2020-10-13 | Philips Image Guided Therapy Corporation | Measurement navigation in a multi-modality medical imaging system |
JP6212938B2 (en) * | 2013-05-10 | 2017-10-18 | 富士通株式会社 | Display processing apparatus, system, and display processing program |
CN203829052U (en) * | 2013-06-19 | 2014-09-17 | 迈柯唯医疗设备(苏州)有限公司 | Medical tower crane system |
US10780298B2 (en) | 2013-08-22 | 2020-09-22 | The Regents Of The University Of Michigan | Histotripsy using very short monopolar ultrasound pulses |
EP3954298A3 (en) | 2013-10-24 | 2022-03-16 | Cathworks Ltd. | Vascular characteristic determination with correspondence modeling of a vascular tree |
US11076301B1 (en) * | 2014-02-13 | 2021-07-27 | Marvell Asia Pte, Ltd. | Multi-mode network access device |
JP6526036B2 (en) * | 2014-02-20 | 2019-06-05 | フレゼニウス ヴィアル エスアーエスFresenius Vial SAS | Medical communication system and communication method |
EP3190958B1 (en) | 2014-09-11 | 2025-02-26 | Koninklijke Philips N.V. | Bedside controller for assessment of vessels and associated devices, systems, and methods |
JP2016059739A (en) * | 2014-09-22 | 2016-04-25 | 富士フイルム株式会社 | Portable console, control method for portable console, program for portable console, and radiographic system |
JP6122410B2 (en) * | 2014-09-22 | 2017-04-26 | 富士フイルム株式会社 | Portable console, portable console control method, portable console program, and radiation imaging system |
JP6849592B2 (en) | 2014-12-08 | 2021-03-24 | コーニンクレッカ フィリップス エヌ ヴェKoninklijke Philips N.V. | Patient education for percutaneous coronary intervention treatment |
JP6835719B2 (en) | 2014-12-08 | 2021-02-24 | コーニンクレッカ フィリップス エヌ ヴェKoninklijke Philips N.V. | Bedside interface for percutaneous coronary intervention treatment planning |
CN104866564B (en) * | 2015-05-21 | 2019-03-15 | 百度在线网络技术(北京)有限公司 | Image searching method and device |
US11135454B2 (en) | 2015-06-24 | 2021-10-05 | The Regents Of The University Of Michigan | Histotripsy therapy systems and methods for the treatment of brain tissue |
JP6681740B2 (en) * | 2016-02-25 | 2020-04-15 | ゼネラル・エレクトリック・カンパニイ | Ultrasonic diagnostic device and its control program |
IL263065B2 (en) | 2016-05-16 | 2024-08-01 | Cathworks Ltd | System for vascular assessment |
WO2017199246A1 (en) | 2016-05-16 | 2017-11-23 | Cathworks Ltd. | Vascular selection from images |
US11058388B2 (en) * | 2016-05-20 | 2021-07-13 | Perimeter Medical Imaging, Inc. | Method and system for combining microscopic imaging with X-Ray imaging |
US11020563B2 (en) | 2016-07-14 | 2021-06-01 | C. R. Bard, Inc. | Automated catheter-to-vessel size comparison tool and related methods |
US12112850B2 (en) | 2017-03-30 | 2024-10-08 | Koninklijke Philips N.V. | Functional measurement patient interface module (PIM) for distributed wireless intraluminal sensing systems |
AU2018345850B2 (en) | 2017-10-06 | 2024-04-11 | Emory University | Methods and systems for determining hemodynamic information for one or more arterial segments |
US20190220185A1 (en) * | 2018-01-12 | 2019-07-18 | Mitutoyo Corporation | Image measurement apparatus and computer readable medium |
KR102524586B1 (en) * | 2018-04-30 | 2023-04-21 | 삼성전자주식회사 | Image display device and operating method for the same |
EP3852622B1 (en) | 2018-10-16 | 2025-04-02 | Bard Access Systems, Inc. | Safety-equipped connection systems and methods thereof for establishing electrical connections |
US11246488B2 (en) * | 2018-11-27 | 2022-02-15 | Shimadzu Corporation | Medical imaging device |
AU2019389001A1 (en) | 2018-11-28 | 2021-06-10 | Histosonics, Inc. | Histotripsy systems and methods |
US12079994B2 (en) | 2019-04-01 | 2024-09-03 | Cathworks Ltd. | Methods and apparatus for angiographic image selection |
US11759166B2 (en) | 2019-09-20 | 2023-09-19 | Bard Access Systems, Inc. | Automatic vessel detection tools and methods |
EP4033964B1 (en) | 2019-09-23 | 2025-04-09 | Cathworks Ltd. | Methods, apparatus, and system for synchronization between a three-dimensional vascular model and an imaging device |
US11341569B2 (en) * | 2019-10-25 | 2022-05-24 | 7-Eleven, Inc. | System and method for populating a virtual shopping cart based on video of a customer's shopping session at a physical store |
AU2021213168A1 (en) | 2020-01-28 | 2022-09-01 | The Regents Of The University Of Michigan | Systems and methods for histotripsy immunosensitization |
EP4181791A1 (en) | 2020-07-21 | 2023-05-24 | Bard Access Systems, Inc. | System, method and apparatus for magnetic tracking of ultrasound probe and generation of 3d visualization thereof |
WO2022031762A1 (en) | 2020-08-04 | 2022-02-10 | Bard Access Systems, Inc. | System and method for optimized medical component insertion monitoring and imaging enhancement |
WO2022035760A1 (en) | 2020-08-10 | 2022-02-17 | Bard Access Systems, Inc. | System and method for generating vessel representations in mixed reality/virtual reality |
CN215839160U (en) | 2020-09-03 | 2022-02-18 | 巴德阿克塞斯系统股份有限公司 | Portable Ultrasound Detectors and Systems |
US11992363B2 (en) | 2020-09-08 | 2024-05-28 | Bard Access Systems, Inc. | Dynamically adjusting ultrasound-imaging systems and methods thereof |
CN216257185U (en) | 2020-09-10 | 2022-04-12 | 巴德阿克塞斯系统股份有限公司 | Ultrasound probe and ultrasound system |
WO2022067101A1 (en) | 2020-09-25 | 2022-03-31 | Bard Access Systems, Inc. | Minimum catheter length tool |
EP4216825A2 (en) | 2020-10-02 | 2023-08-02 | Bard Access Systems, Inc. | Ultrasound systems and methods for sustained spatial attention |
CN216167530U (en) | 2020-10-15 | 2022-04-05 | 巴德阿克塞斯系统股份有限公司 | Ultrasound imaging system configured to generate three-dimensional ultrasound images of a target region |
WO2022115479A1 (en) | 2020-11-24 | 2022-06-02 | Bard Access Systems, Inc. | Ultrasound system with target and medical instrument awareness |
US12165315B2 (en) | 2020-12-01 | 2024-12-10 | Bard Access Systems, Inc. | Ultrasound system with pressure and flow determination capability |
CN114569155A (en) | 2020-12-01 | 2022-06-03 | 巴德阿克塞斯系统股份有限公司 | Ultrasound imaging system and method for obtaining ultrasound image by the same |
US12102481B2 (en) | 2022-06-03 | 2024-10-01 | Bard Access Systems, Inc. | Ultrasound probe with smart accessory |
US12137989B2 (en) | 2022-07-08 | 2024-11-12 | Bard Access Systems, Inc. | Systems and methods for intelligent ultrasound probe guidance |
WO2024054935A1 (en) * | 2022-09-09 | 2024-03-14 | yoR Labs, Inc. | Wired / wireless multiple-modality systems for multiple catheters in interventional cardiology |
EP4418993A1 (en) * | 2022-11-11 | 2024-08-28 | Lightlab Imaging, Inc. | Remote medical imaging system processing |
US12156761B1 (en) | 2024-03-05 | 2024-12-03 | yoR Labs, Inc. | Bayesian anatomically-driven, artificial-intelligence based intracardiac echocardiography object detection and prediction |
Citations (3)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US6597808B1 (en) * | 1999-12-06 | 2003-07-22 | Matsushita Electric Industrial Co., Ltd. | User drawn circled region extraction from scanned documents |
US20070038090A1 (en) * | 2005-07-27 | 2007-02-15 | Medison Co., Ltd. | Ultrasound system for displaying an elastic image |
US20070274591A1 (en) * | 2006-05-26 | 2007-11-29 | Elitegroup Computer Systems Co., Ltd. | Input apparatus and input method thereof |
Family Cites Families (30)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US5056865A (en) * | 1989-08-15 | 1991-10-15 | 501 Century Products Co. | Shopping cart locking means |
US5583946A (en) * | 1993-09-30 | 1996-12-10 | Apple Computer, Inc. | Method and apparatus for recognizing gestures on a computer system |
US5583742A (en) * | 1993-12-15 | 1996-12-10 | Alps Electric Co., Ltd. | Computer with protective cover having outwardly projecting cushioning portions |
US5812188A (en) * | 1996-07-12 | 1998-09-22 | Adair; Edwin L. | Sterile encapsulated endoscopic video monitor |
US6429846B2 (en) * | 1998-06-23 | 2002-08-06 | Immersion Corporation | Haptic feedback for touchpads and other touch controls |
GB2345543A (en) * | 1999-01-06 | 2000-07-12 | Intravascular Res Ltd | Ultrasonic visualisation system with remote components |
US6273857B1 (en) | 1999-07-27 | 2001-08-14 | Siemens Medical Systems, Inc | Method and system for correlating exam worksheet values to supporting measurements |
EP1229822B1 (en) * | 1999-10-27 | 2011-02-09 | DiMicine Research IT, LLC | Control System for Medical Data |
EP1292880A1 (en) * | 2000-05-24 | 2003-03-19 | Koninklijke Philips Electronics N.V. | Immediate mouse control of measuring functionalities for medical images |
JP2001353125A (en) * | 2000-06-14 | 2001-12-25 | Olympus Optical Co Ltd | Endoscopic system |
US6829379B1 (en) * | 2000-11-27 | 2004-12-07 | Ge Medical Systems Global Technology Company, Llc | Methods and apparatus to assist and facilitate vessel analysis |
US20020165005A1 (en) * | 2001-05-03 | 2002-11-07 | Interactive Imaging Systems, Inc. | Portable computing device |
DE10239467B4 (en) * | 2002-08-28 | 2006-01-19 | Dräger Medical AG & Co. KGaA | Holding device for utensils on a patient bed |
WO2005020814A1 (en) | 2003-08-28 | 2005-03-10 | Hebah Noshy Mansour | Cervix monitoring system and related devices and methods |
US7166987B2 (en) * | 2003-10-10 | 2007-01-23 | R. F. Tech Co., Ltd | Portable charger for mobile phone |
JP2005312552A (en) * | 2004-04-27 | 2005-11-10 | Olympus Corp | Observation apparatus |
US7713210B2 (en) | 2004-11-23 | 2010-05-11 | St. Jude Medical, Atrial Fibrillation Division, Inc. | Method and apparatus for localizing an ultrasound catheter |
WO2006102305A2 (en) * | 2005-03-21 | 2006-09-28 | Purdue Research Foundation | Sketch beautification |
AU2006254689B2 (en) * | 2005-06-02 | 2012-03-08 | Salient Imaging, Inc. | System and method of computer-aided detection |
KR100791377B1 (en) * | 2005-12-28 | 2008-01-07 | 삼성전자주식회사 | A mobile device providing a user interface and a method of executing a function using the user interface |
JP5161118B2 (en) | 2006-03-01 | 2013-03-13 | ザ ブリガム アンド ウイメンズ ホスピタル, インク. | Arterial imaging system |
US20080033404A1 (en) * | 2006-08-03 | 2008-02-07 | Romoda Laszlo O | Surgical machine with removable display |
WO2009021179A1 (en) * | 2007-08-09 | 2009-02-12 | Volcano Corporation | Controller user interface for a catheter lab intravascular ultrasound system |
US20090051671A1 (en) * | 2007-08-22 | 2009-02-26 | Jason Antony Konstas | Recognizing the motion of two or more touches on a touch-sensing surface |
JP5202916B2 (en) * | 2007-09-28 | 2013-06-05 | 株式会社東芝 | Ultrasound image diagnostic apparatus and control program thereof |
WO2009137659A1 (en) | 2008-05-07 | 2009-11-12 | Infraredx, Inc. | Multimodal catheter system and method for intravascular analysis |
JP2010240198A (en) * | 2009-04-07 | 2010-10-28 | Toshiba Corp | Ultrasonic diagnostic apparatus |
CN102458217A (en) * | 2009-04-20 | 2012-05-16 | 展望者医疗科技股份有限公司 | Imaging system |
US20110009813A1 (en) * | 2009-07-09 | 2011-01-13 | Medtronic Minimed, Inc. | Panning a display of a portable medical device |
JP2011118554A (en) * | 2009-12-01 | 2011-06-16 | Keyence Corp | Image processing apparatus, image processing method and computer program |
-
2012
- 2012-11-15 CA CA2855830A patent/CA2855830A1/en not_active Abandoned
- 2012-11-15 JP JP2014542464A patent/JP6373758B2/en active Active
- 2012-11-15 WO PCT/US2012/065293 patent/WO2013074800A1/en active Application Filing
- 2012-11-15 EP EP12849119.8A patent/EP2779907B1/en active Active
- 2012-11-16 US US13/679,795 patent/US20130123616A1/en not_active Abandoned
- 2012-11-16 US US13/679,752 patent/US8681116B2/en active Active
- 2012-11-16 US US13/679,776 patent/US8754865B2/en active Active
Patent Citations (3)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US6597808B1 (en) * | 1999-12-06 | 2003-07-22 | Matsushita Electric Industrial Co., Ltd. | User drawn circled region extraction from scanned documents |
US20070038090A1 (en) * | 2005-07-27 | 2007-02-15 | Medison Co., Ltd. | Ultrasound system for displaying an elastic image |
US20070274591A1 (en) * | 2006-05-26 | 2007-11-29 | Elitegroup Computer Systems Co., Ltd. | Input apparatus and input method thereof |
Cited By (121)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US11484205B2 (en) | 2002-03-25 | 2022-11-01 | Masimo Corporation | Physiological measurement device |
US10213108B2 (en) | 2002-03-25 | 2019-02-26 | Masimo Corporation | Arm mountable portable patient monitor |
US10335033B2 (en) | 2002-03-25 | 2019-07-02 | Masimo Corporation | Physiological measurement device |
US10869602B2 (en) | 2002-03-25 | 2020-12-22 | Masimo Corporation | Physiological measurement communications adapter |
US10219706B2 (en) | 2002-03-25 | 2019-03-05 | Masimo Corporation | Physiological measurement device |
US9872623B2 (en) | 2002-03-25 | 2018-01-23 | Masimo Corporation | Arm mountable portable patient monitor |
US9795300B2 (en) | 2002-03-25 | 2017-10-24 | Masimo Corporation | Wearable portable patient monitor |
US9788735B2 (en) | 2002-03-25 | 2017-10-17 | Masimo Corporation | Body worn mobile medical patient monitor |
US10912524B2 (en) | 2006-09-22 | 2021-02-09 | Masimo Corporation | Modular patient monitor |
US9968256B2 (en) | 2007-03-08 | 2018-05-15 | Sync-Rx Ltd. | Automatic identification of a tool |
US8693756B2 (en) | 2007-03-08 | 2014-04-08 | Sync-Rx, Ltd. | Automatic reduction of interfering elements from an image stream of a moving organ |
US10226178B2 (en) | 2007-03-08 | 2019-03-12 | Sync-Rx Ltd. | Automatic reduction of visibility of portions of an image |
US8700130B2 (en) | 2007-03-08 | 2014-04-15 | Sync-Rx, Ltd. | Stepwise advancement of a medical tool |
US10307061B2 (en) | 2007-03-08 | 2019-06-04 | Sync-Rx, Ltd. | Automatic tracking of a tool upon a vascular roadmap |
US12053317B2 (en) | 2007-03-08 | 2024-08-06 | Sync-Rx Ltd. | Determining a characteristic of a lumen by measuring velocity of a contrast agent |
US9216065B2 (en) | 2007-03-08 | 2015-12-22 | Sync-Rx, Ltd. | Forming and displaying a composite image |
US9014453B2 (en) | 2007-03-08 | 2015-04-21 | Sync-Rx, Ltd. | Automatic angiogram detection |
US9305334B2 (en) | 2007-03-08 | 2016-04-05 | Sync-Rx, Ltd. | Luminal background cleaning |
US9308052B2 (en) | 2007-03-08 | 2016-04-12 | Sync-Rx, Ltd. | Pre-deployment positioning of an implantable device within a moving organ |
US10499814B2 (en) | 2007-03-08 | 2019-12-10 | Sync-Rx, Ltd. | Automatic generation and utilization of a vascular roadmap |
US9888969B2 (en) | 2007-03-08 | 2018-02-13 | Sync-Rx Ltd. | Automatic quantitative vessel analysis |
US9375164B2 (en) | 2007-03-08 | 2016-06-28 | Sync-Rx, Ltd. | Co-use of endoluminal data and extraluminal imaging |
US11197651B2 (en) | 2007-03-08 | 2021-12-14 | Sync-Rx, Ltd. | Identification and presentation of device-to-vessel relative motion |
US10716528B2 (en) | 2007-03-08 | 2020-07-21 | Sync-Rx, Ltd. | Automatic display of previously-acquired endoluminal images |
US9629571B2 (en) | 2007-03-08 | 2017-04-25 | Sync-Rx, Ltd. | Co-use of endoluminal data and extraluminal imaging |
US9717415B2 (en) | 2007-03-08 | 2017-08-01 | Sync-Rx, Ltd. | Automatic quantitative vessel analysis at the location of an automatically-detected tool |
US11179038B2 (en) | 2007-03-08 | 2021-11-23 | Sync-Rx, Ltd | Automatic stabilization of a frames of image stream of a moving organ having intracardiac or intravascular tool in the organ that is displayed in movie format |
US9008754B2 (en) | 2007-03-08 | 2015-04-14 | Sync-Rx, Ltd. | Automatic correction and utilization of a vascular roadmap comprising a tool |
US9008367B2 (en) | 2007-03-08 | 2015-04-14 | Sync-Rx, Ltd. | Apparatus and methods for reducing visibility of a periphery of an image stream |
US8781193B2 (en) | 2007-03-08 | 2014-07-15 | Sync-Rx, Ltd. | Automatic quantitative vessel analysis |
US9855384B2 (en) | 2007-03-08 | 2018-01-02 | Sync-Rx, Ltd. | Automatic enhancement of an image stream of a moving organ and displaying as a movie |
US11064964B2 (en) | 2007-03-08 | 2021-07-20 | Sync-Rx, Ltd | Determining a characteristic of a lumen by measuring velocity of a contrast agent |
US9974509B2 (en) | 2008-11-18 | 2018-05-22 | Sync-Rx Ltd. | Image super enhancement |
US10362962B2 (en) | 2008-11-18 | 2019-07-30 | Synx-Rx, Ltd. | Accounting for skipped imaging locations during movement of an endoluminal imaging probe |
US11064903B2 (en) | 2008-11-18 | 2021-07-20 | Sync-Rx, Ltd | Apparatus and methods for mapping a sequence of images to a roadmap image |
US9095313B2 (en) | 2008-11-18 | 2015-08-04 | Sync-Rx, Ltd. | Accounting for non-uniform longitudinal motion during movement of an endoluminal imaging probe |
US8855744B2 (en) | 2008-11-18 | 2014-10-07 | Sync-Rx, Ltd. | Displaying a device within an endoluminal image stack |
US11883149B2 (en) | 2008-11-18 | 2024-01-30 | Sync-Rx Ltd. | Apparatus and methods for mapping a sequence of images to a roadmap image |
US9101286B2 (en) | 2008-11-18 | 2015-08-11 | Sync-Rx, Ltd. | Apparatus and methods for determining a dimension of a portion of a stack of endoluminal data points |
US9144394B2 (en) | 2008-11-18 | 2015-09-29 | Sync-Rx, Ltd. | Apparatus and methods for determining a plurality of local calibration factors for an image |
US11963736B2 (en) | 2009-07-20 | 2024-04-23 | Masimo Corporation | Wireless patient monitoring system |
US11900775B2 (en) | 2009-12-21 | 2024-02-13 | Masimo Corporation | Modular patient monitor |
US10354504B2 (en) | 2009-12-21 | 2019-07-16 | Masimo Corporation | Modular patient monitor |
US20150351704A1 (en) * | 2009-12-21 | 2015-12-10 | Masimo Corporation | Modular patient monitor |
US9847002B2 (en) * | 2009-12-21 | 2017-12-19 | Masimo Corporation | Modular patient monitor |
US10943450B2 (en) | 2009-12-21 | 2021-03-09 | Masimo Corporation | Modular patient monitor |
US9610064B2 (en) | 2011-05-31 | 2017-04-04 | Desmond Adler | Multimodal imaging system, apparatus, and methods |
US11241154B2 (en) | 2011-05-31 | 2022-02-08 | Lightlab Imaging, Inc. | Multimodal imaging system, apparatus, and methods |
US11241199B2 (en) | 2011-10-13 | 2022-02-08 | Masimo Corporation | System for displaying medical monitoring data |
US10512436B2 (en) | 2011-10-13 | 2019-12-24 | Masimo Corporation | System for displaying medical monitoring data |
US11179114B2 (en) | 2011-10-13 | 2021-11-23 | Masimo Corporation | Medical monitoring hub |
US9913617B2 (en) | 2011-10-13 | 2018-03-13 | Masimo Corporation | Medical monitoring hub |
US9943269B2 (en) | 2011-10-13 | 2018-04-17 | Masimo Corporation | System for displaying medical monitoring data |
US10925550B2 (en) | 2011-10-13 | 2021-02-23 | Masimo Corporation | Medical monitoring hub |
US9993207B2 (en) | 2011-10-13 | 2018-06-12 | Masimo Corporation | Medical monitoring hub |
US11786183B2 (en) | 2011-10-13 | 2023-10-17 | Masimo Corporation | Medical monitoring hub |
US10188296B2 (en) | 2012-02-09 | 2019-01-29 | Masimo Corporation | Wireless patient monitoring device |
US12109022B2 (en) | 2012-02-09 | 2024-10-08 | Masimo Corporation | Wireless patient monitoring device |
US11083397B2 (en) | 2012-02-09 | 2021-08-10 | Masimo Corporation | Wireless patient monitoring device |
US10149616B2 (en) | 2012-02-09 | 2018-12-11 | Masimo Corporation | Wireless patient monitoring device |
US11918353B2 (en) | 2012-02-09 | 2024-03-05 | Masimo Corporation | Wireless patient monitoring device |
US10984531B2 (en) | 2012-06-26 | 2021-04-20 | Sync-Rx, Ltd. | Determining a luminal-flow-related index using blood velocity determination |
US10748289B2 (en) | 2012-06-26 | 2020-08-18 | Sync-Rx, Ltd | Coregistration of endoluminal data points with values of a luminal-flow-related index |
US11887728B2 (en) | 2012-09-20 | 2024-01-30 | Masimo Corporation | Intelligent medical escalation process |
US10833983B2 (en) | 2012-09-20 | 2020-11-10 | Masimo Corporation | Intelligent medical escalation process |
US9486143B2 (en) * | 2012-12-21 | 2016-11-08 | Volcano Corporation | Intravascular forward imaging device |
US20140180135A1 (en) * | 2012-12-21 | 2014-06-26 | Volcano Corporation | Chronic total occlusion catheter |
US10368836B2 (en) * | 2012-12-26 | 2019-08-06 | Volcano Corporation | Gesture-based interface for a multi-modality medical imaging system |
US20140181716A1 (en) * | 2012-12-26 | 2014-06-26 | Volcano Corporation | Gesture-Based Interface for a Multi-Modality Medical Imaging System |
US10409951B2 (en) * | 2012-12-28 | 2019-09-10 | Volcano Corporation | Multi-modality case management system and method |
US9351698B2 (en) | 2013-03-12 | 2016-05-31 | Lightlab Imaging, Inc. | Vascular data processing and image registration systems, methods, and apparatuses |
US9907527B2 (en) | 2013-03-12 | 2018-03-06 | Lightlab Imaging, Inc. | Vascular data processing and image registration systems, methods, and apparatuses |
US10687777B2 (en) | 2013-03-12 | 2020-06-23 | Lightlab Imaging, Inc. | Vascular data processing and image registration systems, methods, and apparatuses |
DE102013210230A1 (en) * | 2013-06-03 | 2014-10-16 | Siemens Aktiengesellschaft | Medical diagnostic and / or therapeutic device |
US10025901B2 (en) * | 2013-07-19 | 2018-07-17 | Ricoh Company Ltd. | Healthcare system integration |
US20160171160A1 (en) * | 2013-07-19 | 2016-06-16 | Ricoh Company, Ltd. | Healthcare system integration |
US11715560B2 (en) | 2013-09-12 | 2023-08-01 | Johnson & Johnson Surgical Vision, Inc. | Computer-based operating room support system |
US11631496B2 (en) | 2013-09-12 | 2023-04-18 | Johnson & Johnson Surgical Vision, Inc. | Computer-based operating room support system |
US11699526B2 (en) | 2013-10-11 | 2023-07-11 | Masimo Corporation | Alarm notification system |
US10832818B2 (en) | 2013-10-11 | 2020-11-10 | Masimo Corporation | Alarm notification system |
US12009098B2 (en) | 2013-10-11 | 2024-06-11 | Masimo Corporation | Alarm notification system |
US10825568B2 (en) | 2013-10-11 | 2020-11-03 | Masimo Corporation | Alarm notification system |
US11488711B2 (en) | 2013-10-11 | 2022-11-01 | Masimo Corporation | Alarm notification system |
US12230396B2 (en) | 2013-10-11 | 2025-02-18 | Masimo Corporation | Alarm notification system |
US9244590B1 (en) * | 2013-12-13 | 2016-01-26 | Amazon Technologies, Inc. | Three-dimensional navigation using a two-dimensional surface |
US9778691B2 (en) * | 2013-12-18 | 2017-10-03 | Siemens Aktiengesellschaft | Medical facility |
US20150169001A1 (en) * | 2013-12-18 | 2015-06-18 | Siemens Aktiengesellschaft | Medical facility |
US10499813B2 (en) | 2014-09-12 | 2019-12-10 | Lightlab Imaging, Inc. | Methods, systems and apparatus for temporal calibration of an intravascular imaging system |
US10105107B2 (en) | 2015-01-08 | 2018-10-23 | St. Jude Medical International Holding S.À R.L. | Medical system having combined and synergized data output from multiple independent inputs |
US10646198B2 (en) | 2015-05-17 | 2020-05-12 | Lightlab Imaging, Inc. | Intravascular imaging and guide catheter detection methods and systems |
US10109058B2 (en) | 2015-05-17 | 2018-10-23 | Lightlab Imaging, Inc. | Intravascular imaging system interfaces and stent detection methods |
US10222956B2 (en) | 2015-05-17 | 2019-03-05 | Lightlab Imaging, Inc. | Intravascular imaging user interface systems and methods |
US11367186B2 (en) | 2015-05-17 | 2022-06-21 | Lightlab Imaging, Inc. | Detection of metal stent struts |
US10713786B2 (en) | 2015-05-17 | 2020-07-14 | Lightlab Imaging, Inc. | Detection of metal stent struts |
US10902599B2 (en) | 2015-05-17 | 2021-01-26 | Lightlab Imaging, Inc. | Stent detection methods and imaging system interfaces |
US9996921B2 (en) | 2015-05-17 | 2018-06-12 | LIGHTLAB IMAGING, lNC. | Detection of metal stent struts |
US11532087B2 (en) | 2015-05-17 | 2022-12-20 | Lightlab Imaging, Inc. | Stent detection methods and imaging system interfaces |
US11768593B2 (en) | 2015-07-25 | 2023-09-26 | Lightlab Imaging, Inc. | Intravascular data visualization and interface systems and methods |
US11287961B2 (en) | 2015-07-25 | 2022-03-29 | Lightlab Imaging, Inc. | Intravascular data visualization and interface systems and methods |
US12161500B2 (en) | 2015-11-18 | 2024-12-10 | Lightlab Imaging, Inc. | X-ray image feature detection and registration systems and methods |
US10327726B2 (en) | 2015-11-18 | 2019-06-25 | Lightlab Imaging, Inc. | X-ray image feature detection and registration systems and methods |
US11633167B2 (en) | 2015-11-18 | 2023-04-25 | Lightlab Imaging, Inc. | X-ray image feature detection and registration systems and methods |
US11020078B2 (en) | 2015-11-18 | 2021-06-01 | Lightlab Imaging, Inc. | X-ray image feature detection and registration systems and methods |
US10172582B2 (en) | 2015-11-18 | 2019-01-08 | Lightlab Imaging, Inc. | X-ray image feature detection and registration systems and methods |
US10342502B2 (en) | 2015-11-18 | 2019-07-09 | Lightlab Imaging, Inc. | X-ray image feature detection and registration systems and methods |
US11850089B2 (en) | 2015-11-19 | 2023-12-26 | Lightlab Imaging, Inc. | Intravascular imaging and guide catheter detection methods and systems |
US10453190B2 (en) | 2015-11-23 | 2019-10-22 | Lightlab Imaging, Inc. | Detection of and validation of shadows in intravascular images |
US10593037B2 (en) | 2016-04-14 | 2020-03-17 | Lightlab Imaging, Inc. | Method, apparatus, and system to identify branches of a blood vessel |
US11475560B2 (en) | 2016-04-14 | 2022-10-18 | Lightlab Imaging, Inc. | Method, apparatus, and system to identify branches of a blood vessel |
US10631754B2 (en) | 2016-05-16 | 2020-04-28 | Lightlab Imaging, Inc. | Intravascular absorbable stent detection and diagnostic methods and systems |
US12070293B2 (en) | 2016-07-07 | 2024-08-27 | Masimo Corporation | Wearable pulse oximeter and respiration monitor |
US11202571B2 (en) | 2016-07-07 | 2021-12-21 | Masimo Corporation | Wearable pulse oximeter and respiration monitor |
US10617302B2 (en) | 2016-07-07 | 2020-04-14 | Masimo Corporation | Wearable pulse oximeter and respiration monitor |
US11883107B2 (en) | 2016-09-28 | 2024-01-30 | Lightlab Imaging, Inc. | Stent planning systems and methods using vessel representation obtained via intravascular probe by determining stent effectiveness score and fractional flow reserve |
US11844634B2 (en) | 2018-04-19 | 2023-12-19 | Masimo Corporation | Mobile patient alarm display |
US12193849B2 (en) | 2018-04-19 | 2025-01-14 | Masimo Corporation | Mobile patient alarm display |
US11109818B2 (en) | 2018-04-19 | 2021-09-07 | Masimo Corporation | Mobile patient alarm display |
US12067225B2 (en) | 2018-05-03 | 2024-08-20 | Canon U.S.A., Inc. | Devices, systems, and methods to emphasize regions of interest across multiple imaging modalities |
EP3563757A1 (en) * | 2018-05-03 | 2019-11-06 | Canon U.S.A. Inc. | Devices, systems, and methods to emphasize regions of interest across multiple imaging modalities |
US12257022B2 (en) | 2018-10-12 | 2025-03-25 | Masimo Corporation | System for transmission of sensor data using dual communication protocol |
WO2024222942A1 (en) * | 2023-04-28 | 2024-10-31 | Shanghai United Imaging Healthcare Co., Ltd. | Methods and systems for controlling medical device |
Also Published As
Publication number | Publication date |
---|---|
US20130120296A1 (en) | 2013-05-16 |
EP2779907B1 (en) | 2022-06-15 |
CA2855830A1 (en) | 2013-05-23 |
JP6373758B2 (en) | 2018-08-15 |
WO2013074800A1 (en) | 2013-05-23 |
US8681116B2 (en) | 2014-03-25 |
JP2015502206A (en) | 2015-01-22 |
EP2779907A1 (en) | 2014-09-24 |
EP2779907A4 (en) | 2015-11-11 |
US8754865B2 (en) | 2014-06-17 |
US20130120297A1 (en) | 2013-05-16 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
US8681116B2 (en) | Medical mounting system and method | |
US20230012365A1 (en) | Bedside controller for assessment of vessels and associated devices, systems, and methods | |
US20240127930A1 (en) | Bedside interface for percutaneous coronary intervention planning | |
US20210030385A1 (en) | Patient education for percutaneous coronary intervention treatments | |
US9779483B2 (en) | Measurement and enhancement in a multi-modality medical imaging system | |
JP6235610B2 (en) | Measurement and enhancement in multi-modality medical imaging systems | |
US20230301519A1 (en) | Multi-modality medical system and associated devices and methods |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
AS | Assignment |
Owner name: VOLCANO CORPORATION, CALIFORNIA Free format text: ASSIGNMENT OF ASSIGNORS INTEREST;ASSIGNORS:MERRITT, FERGUS;COHEN, ASHER;DE JONG, DUANE;AND OTHERS;SIGNING DATES FROM 20140416 TO 20140429;REEL/FRAME:032785/0693 |
|
STCB | Information on status: application discontinuation |
Free format text: ABANDONED -- FAILURE TO RESPOND TO AN OFFICE ACTION |