US7703919B2 - Application of neuro-ocular wavefront data in vision correction - Google Patents
Application of neuro-ocular wavefront data in vision correction Download PDFInfo
- Publication number
- US7703919B2 US7703919B2 US10/811,368 US81136804A US7703919B2 US 7703919 B2 US7703919 B2 US 7703919B2 US 81136804 A US81136804 A US 81136804A US 7703919 B2 US7703919 B2 US 7703919B2
- Authority
- US
- United States
- Prior art keywords
- subject
- visual
- neuro
- vision
- image
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Expired - Fee Related, expires
Links
- 238000012937 correction Methods 0.000 title claims description 73
- 230000004438 eyesight Effects 0.000 title claims description 45
- 238000000034 method Methods 0.000 claims abstract description 131
- 230000000007 visual effect Effects 0.000 claims abstract description 129
- 210000001747 pupil Anatomy 0.000 claims description 74
- 238000011282 treatment Methods 0.000 claims description 52
- 230000003287 optical effect Effects 0.000 claims description 36
- 238000002679 ablation Methods 0.000 claims description 35
- 230000004075 alteration Effects 0.000 claims description 32
- 230000002980 postoperative effect Effects 0.000 claims description 31
- 230000000875 corresponding effect Effects 0.000 claims description 25
- 239000011159 matrix material Substances 0.000 claims description 25
- 208000001491 myopia Diseases 0.000 claims description 25
- 238000001356 surgical procedure Methods 0.000 claims description 21
- 230000007704 transition Effects 0.000 claims description 18
- 230000007613 environmental effect Effects 0.000 claims description 8
- 206010010071 Coma Diseases 0.000 claims description 4
- 235000004035 Cryptotaenia japonica Nutrition 0.000 claims description 4
- 102000007641 Trefoil Factors Human genes 0.000 claims description 4
- 235000015724 Trifolium pratense Nutrition 0.000 claims description 4
- 201000009310 astigmatism Diseases 0.000 claims description 4
- 230000002596 correlated effect Effects 0.000 claims description 4
- 230000003500 cycloplegic effect Effects 0.000 claims description 4
- 238000011065 in-situ storage Methods 0.000 claims description 4
- 230000004297 night vision Effects 0.000 claims description 4
- 238000002271 resection Methods 0.000 claims description 4
- 238000004088 simulation Methods 0.000 claims description 3
- 238000007620 mathematical function Methods 0.000 claims 4
- 238000005070 sampling Methods 0.000 description 55
- 230000008569 process Effects 0.000 description 49
- 230000006870 function Effects 0.000 description 38
- 238000010586 diagram Methods 0.000 description 33
- 230000000694 effects Effects 0.000 description 18
- 238000012545 processing Methods 0.000 description 16
- 238000012360 testing method Methods 0.000 description 16
- 238000005259 measurement Methods 0.000 description 14
- 238000004364 calculation method Methods 0.000 description 10
- 238000007405 data analysis Methods 0.000 description 10
- 238000004422 calculation algorithm Methods 0.000 description 9
- 230000037361 pathway Effects 0.000 description 9
- 230000000926 neurological effect Effects 0.000 description 8
- 238000013459 approach Methods 0.000 description 7
- 238000005457 optimization Methods 0.000 description 7
- 230000002207 retinal effect Effects 0.000 description 7
- 230000004048 modification Effects 0.000 description 6
- 238000012986 modification Methods 0.000 description 6
- 238000007619 statistical method Methods 0.000 description 6
- 238000004458 analytical method Methods 0.000 description 5
- 210000004087 cornea Anatomy 0.000 description 5
- 238000013479 data entry Methods 0.000 description 5
- 230000001179 pupillary effect Effects 0.000 description 5
- 230000002452 interceptive effect Effects 0.000 description 4
- 230000033001 locomotion Effects 0.000 description 4
- 230000004304 visual acuity Effects 0.000 description 4
- 210000003484 anatomy Anatomy 0.000 description 3
- 230000008901 benefit Effects 0.000 description 3
- 239000013256 coordination polymer Substances 0.000 description 3
- 230000001419 dependent effect Effects 0.000 description 3
- 238000003745 diagnosis Methods 0.000 description 3
- 238000005286 illumination Methods 0.000 description 3
- 239000000463 material Substances 0.000 description 3
- 230000002093 peripheral effect Effects 0.000 description 3
- 208000014733 refractive error Diseases 0.000 description 3
- 210000004556 brain Anatomy 0.000 description 2
- 238000013461 design Methods 0.000 description 2
- 230000004424 eye movement Effects 0.000 description 2
- 238000009499 grossing Methods 0.000 description 2
- 230000006872 improvement Effects 0.000 description 2
- 238000000608 laser ablation Methods 0.000 description 2
- 230000004044 response Effects 0.000 description 2
- 210000001525 retina Anatomy 0.000 description 2
- 238000003860 storage Methods 0.000 description 2
- PEDCQBHIVMGVHV-UHFFFAOYSA-N Glycerine Chemical compound OCC(O)CO PEDCQBHIVMGVHV-UHFFFAOYSA-N 0.000 description 1
- 235000010627 Phaseolus vulgaris Nutrition 0.000 description 1
- 244000046052 Phaseolus vulgaris Species 0.000 description 1
- 230000002776 aggregation Effects 0.000 description 1
- 238000004220 aggregation Methods 0.000 description 1
- 230000005540 biological transmission Effects 0.000 description 1
- 239000003086 colorant Substances 0.000 description 1
- 239000002131 composite material Substances 0.000 description 1
- 230000007812 deficiency Effects 0.000 description 1
- 230000002950 deficient Effects 0.000 description 1
- 238000009795 derivation Methods 0.000 description 1
- 238000011161 development Methods 0.000 description 1
- 230000010339 dilation Effects 0.000 description 1
- 238000005516 engineering process Methods 0.000 description 1
- 238000009472 formulation Methods 0.000 description 1
- 210000003128 head Anatomy 0.000 description 1
- 238000003780 insertion Methods 0.000 description 1
- 230000037431 insertion Effects 0.000 description 1
- 238000013532 laser treatment Methods 0.000 description 1
- 230000007246 mechanism Effects 0.000 description 1
- 239000000203 mixture Substances 0.000 description 1
- 230000001537 neural effect Effects 0.000 description 1
- 238000010606 normalization Methods 0.000 description 1
- 239000013307 optical fiber Substances 0.000 description 1
- 230000008447 perception Effects 0.000 description 1
- 230000000144 pharmacologic effect Effects 0.000 description 1
- 238000007670 refining Methods 0.000 description 1
- 239000004065 semiconductor Substances 0.000 description 1
- 230000001360 synchronised effect Effects 0.000 description 1
- 238000012546 transfer Methods 0.000 description 1
- 230000009466 transformation Effects 0.000 description 1
- 230000004400 visual pathway Effects 0.000 description 1
- 210000000239 visual pathway Anatomy 0.000 description 1
Images
Classifications
-
- A—HUMAN NECESSITIES
- A61—MEDICAL OR VETERINARY SCIENCE; HYGIENE
- A61F—FILTERS IMPLANTABLE INTO BLOOD VESSELS; PROSTHESES; DEVICES PROVIDING PATENCY TO, OR PREVENTING COLLAPSING OF, TUBULAR STRUCTURES OF THE BODY, e.g. STENTS; ORTHOPAEDIC, NURSING OR CONTRACEPTIVE DEVICES; FOMENTATION; TREATMENT OR PROTECTION OF EYES OR EARS; BANDAGES, DRESSINGS OR ABSORBENT PADS; FIRST-AID KITS
- A61F9/00—Methods or devices for treatment of the eyes; Devices for putting in contact-lenses; Devices to correct squinting; Apparatus to guide the blind; Protective devices for the eyes, carried on the body or in the hand
- A61F9/007—Methods or devices for eye surgery
-
- A—HUMAN NECESSITIES
- A61—MEDICAL OR VETERINARY SCIENCE; HYGIENE
- A61B—DIAGNOSIS; SURGERY; IDENTIFICATION
- A61B3/00—Apparatus for testing the eyes; Instruments for examining the eyes
- A61B3/0016—Operational features thereof
- A61B3/0025—Operational features thereof characterised by electronic signal processing, e.g. eye models
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06V—IMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
- G06V40/00—Recognition of biometric, human-related or animal-related patterns in image or video data
- G06V40/10—Human or animal bodies, e.g. vehicle occupants or pedestrians; Body parts, e.g. hands
- G06V40/18—Eye characteristics, e.g. of the iris
- G06V40/193—Preprocessing; Feature extraction
-
- G—PHYSICS
- G16—INFORMATION AND COMMUNICATION TECHNOLOGY [ICT] SPECIALLY ADAPTED FOR SPECIFIC APPLICATION FIELDS
- G16H—HEALTHCARE INFORMATICS, i.e. INFORMATION AND COMMUNICATION TECHNOLOGY [ICT] SPECIALLY ADAPTED FOR THE HANDLING OR PROCESSING OF MEDICAL OR HEALTHCARE DATA
- G16H10/00—ICT specially adapted for the handling or processing of patient-related medical or healthcare data
- G16H10/60—ICT specially adapted for the handling or processing of patient-related medical or healthcare data for patient-specific data, e.g. for electronic patient records
-
- G—PHYSICS
- G16—INFORMATION AND COMMUNICATION TECHNOLOGY [ICT] SPECIALLY ADAPTED FOR SPECIFIC APPLICATION FIELDS
- G16H—HEALTHCARE INFORMATICS, i.e. INFORMATION AND COMMUNICATION TECHNOLOGY [ICT] SPECIALLY ADAPTED FOR THE HANDLING OR PROCESSING OF MEDICAL OR HEALTHCARE DATA
- G16H20/00—ICT specially adapted for therapies or health-improving plans, e.g. for handling prescriptions, for steering therapy or for monitoring patient compliance
- G16H20/40—ICT specially adapted for therapies or health-improving plans, e.g. for handling prescriptions, for steering therapy or for monitoring patient compliance relating to mechanical, radiation or invasive therapies, e.g. surgery, laser therapy, dialysis or acupuncture
-
- G—PHYSICS
- G16—INFORMATION AND COMMUNICATION TECHNOLOGY [ICT] SPECIALLY ADAPTED FOR SPECIFIC APPLICATION FIELDS
- G16H—HEALTHCARE INFORMATICS, i.e. INFORMATION AND COMMUNICATION TECHNOLOGY [ICT] SPECIALLY ADAPTED FOR THE HANDLING OR PROCESSING OF MEDICAL OR HEALTHCARE DATA
- G16H40/00—ICT specially adapted for the management or administration of healthcare resources or facilities; ICT specially adapted for the management or operation of medical equipment or devices
- G16H40/60—ICT specially adapted for the management or administration of healthcare resources or facilities; ICT specially adapted for the management or operation of medical equipment or devices for the operation of medical equipment or devices
- G16H40/63—ICT specially adapted for the management or administration of healthcare resources or facilities; ICT specially adapted for the management or operation of medical equipment or devices for the operation of medical equipment or devices for local operation
-
- A—HUMAN NECESSITIES
- A61—MEDICAL OR VETERINARY SCIENCE; HYGIENE
- A61F—FILTERS IMPLANTABLE INTO BLOOD VESSELS; PROSTHESES; DEVICES PROVIDING PATENCY TO, OR PREVENTING COLLAPSING OF, TUBULAR STRUCTURES OF THE BODY, e.g. STENTS; ORTHOPAEDIC, NURSING OR CONTRACEPTIVE DEVICES; FOMENTATION; TREATMENT OR PROTECTION OF EYES OR EARS; BANDAGES, DRESSINGS OR ABSORBENT PADS; FIRST-AID KITS
- A61F2/00—Filters implantable into blood vessels; Prostheses, i.e. artificial substitutes or replacements for parts of the body; Appliances for connecting them with the body; Devices providing patency to, or preventing collapsing of, tubular structures of the body, e.g. stents
- A61F2/02—Prostheses implantable into the body
- A61F2/14—Eye parts, e.g. lenses or corneal implants; Artificial eyes
- A61F2/147—Implants to be inserted in the stroma for refractive correction, e.g. ring-like implants
-
- A—HUMAN NECESSITIES
- A61—MEDICAL OR VETERINARY SCIENCE; HYGIENE
- A61F—FILTERS IMPLANTABLE INTO BLOOD VESSELS; PROSTHESES; DEVICES PROVIDING PATENCY TO, OR PREVENTING COLLAPSING OF, TUBULAR STRUCTURES OF THE BODY, e.g. STENTS; ORTHOPAEDIC, NURSING OR CONTRACEPTIVE DEVICES; FOMENTATION; TREATMENT OR PROTECTION OF EYES OR EARS; BANDAGES, DRESSINGS OR ABSORBENT PADS; FIRST-AID KITS
- A61F9/00—Methods or devices for treatment of the eyes; Devices for putting in contact-lenses; Devices to correct squinting; Apparatus to guide the blind; Protective devices for the eyes, carried on the body or in the hand
- A61F9/007—Methods or devices for eye surgery
- A61F9/013—Instruments for compensation of ocular refraction ; Instruments for use in cornea removal, for reshaping or performing incisions in the cornea
-
- G—PHYSICS
- G02—OPTICS
- G02C—SPECTACLES; SUNGLASSES OR GOGGLES INSOFAR AS THEY HAVE THE SAME FEATURES AS SPECTACLES; CONTACT LENSES
- G02C2202/00—Generic optical aspects applicable to one or more of the subgroups of G02C7/00
- G02C2202/22—Correction of higher order and chromatic aberrations, wave front measurement and calculation
Definitions
- the present disclosure relates generally to metrics and, more particularly, to visual metrics associated with vision correction.
- ocular aberrometers provide detailed information on the optical characteristics of the eye
- these devices usually provide no information with reference to the neurological pathways between the eye and the brain and they are not capable of localizing the retinal plane preferred by the patient.
- the integrated visual pathway between the eye and the brain can introduce another transfer function, which is undetectable by ocular aberrometry.
- a technique known as visual aberrometry in which the patient's preferred retinal plane is localized and the effects of the neurological pathways are taken into account by combining subject feedback with the physio-optical characteristics of the eye.
- the aggregate effect of the neurological pathway in combination with the characteristics of the eye is referred to herein as “neuro-ocular wavefront error” or “neuro-ocular wavefront data.”
- Examples of visual aberrometers are described in greater detail in U.S. Pat. Nos. 6,000,800 and 6,099,125, which are incorporated herein by reference in their entireties.
- FIG. 2 is a diagram illustrating an embodiment of a pupillometry image, which may be used in a refractometer adapted to acquire neuro-ocular wavefront data.
- FIG. 3 is a diagram illustrating an embodiment of a sampling matrix in a refractometer adapted to acquire neuro-ocular wavefront data.
- FIG. 4 is a diagram showing, more clearly, the treatment envelope of FIG. 3 .
- FIG. 5 is a diagram showing embodiments of various pupil-sampling maps.
- FIG. 8 is a diagram showing a display console with a sample alignment target and a test spot.
- FIG. 9 is a diagram showing a coordinate system of the neuro-ocular wavefront data and a coordinate system for an ablation map that corresponds to the neuro-ocular wavefront data.
- FIG. 10 is a diagram showing an optical zone and a transition zone associated with the ablation map of FIG. 9 .
- FIG. 11 is a diagram showing an embodiment of a sampling map that has a predefined matrix of sampling points.
- FIG. 12 is a diagram showing an embodiment of a sampling map that permits customization of sampling points.
- FIG. 14 is a diagram showing an embodiment of a video display that facilitates on-screen measurements of physical characteristics of the eye.
- FIG. 16 is a block diagram showing an embodiment of a system configured to aggregate data from multiple refractometers.
- FIG. 18 is a flowchart showing, in greater detail, an embodiment of the data-acquisition step from FIG. 17 .
- FIG. 19 is a flowchart showing, in greater detail, an embodiment of the step of identifying visual zones, from FIG. 18 .
- FIGS. 20A and 20B are flowcharts showing, in greater detail, an embodiment of the step of interactively obtaining information, from FIG. 18 .
- FIG. 21 is a flowchart showing, in greater detail, an embodiment of the step of correlating neuro-ocular wavefront data to vision parameters, from FIG. 17 .
- FIGS. 22A and 22B are flowcharts showing, in greater detail, an embodiment of the data-analysis step from FIG. 17 .
- FIG. 23 is a flowchart showing, in greater detail, another embodiment of the data-analysis step from FIG. 17 .
- FIG. 24 is a flowchart showing another embodiment of a method for acquiring and processing neuro-ocular wavefront data.
- FIGS. 27A and 27B are flowcharts showing another embodiment of a method for acquiring and processing neuro-ocular wavefront data, in which multiple sets of data are statistically analyzed.
- FIG. 28 is a table showing definitions of various variables related to the locations of sampling points and their respective corrections.
- FIG. 29 is a table showing definitions of the coordinate systems in FIG. 9 .
- vision correction includes measuring vision with visual acuity determined by eye charts and measuring the refractive error of the eye with test lenses.
- More advanced approaches employ ocular aberrometry, which provides greater details on the refractive properties of the eye. While both of these approaches provide useful results for vision correction, they are nonetheless deficient in many ways.
- visual acuity provides nothing more than a gross estimate of the performance of the patient's visual system.
- Refraction provides only the best sphero-cylindrical correction that is averaged across the pupil.
- Ocular aberrometry while providing more detailed refractive characteristics of the eye, provides no information on the preferred retinal plane and no information related to the neurological pathways associated with the eyes.
- An improvement to both refraction and ocular aberrometry is visual aberrometry, in which the patient preferred retinal plane and the effects of the neurological pathways are taken into account by incorporating a subject feedback in a testing process that assesses multiple points within the pupil.
- the aggregate effect of the location of the patient's preferred retinal plane and the effects of the neurological pathways is known as a “neuro-ocular wavefront error” or “neuro-ocular wavefront data.”
- the present disclosure provides various systems and methods for exploiting the full potential of visual aberrometry. In other words, the present disclosure provides systems and methods in which neuro-ocular wavefront data can be used for diagnosis, treatment, or a combination of both.
- neuro-ocular wavefront data is obtained from a subject (or, more specifically, a patient) and, thereafter, processed.
- processing of the neuro-ocular wavefront data includes correlating the neuro-ocular wavefront data with various subject-dependent or procedure-dependent parameters (hereinafter referred to in the aggregate as “vision parameters”).
- the neuro-ocular wavefront data is often represented as an equation.
- an example equation may be a Zernike polynomial with multiple coefficients. The coefficients in the Zernike polynomial, either alone or in combination with other coefficients, represents a particular characteristic of the patient's visual system.
- the correlated information is used for designing treatments.
- the treatment can be designed by inverting the neuro-ocular wavefront data. For example, if the neuro-ocular wavefront error is defined by a Zernike polynomial, then the inversion of the polynomial provides one or more correction factors that can be applied, thereby effectively correcting for the apparent error that is introduced by the aberrations in the neuro-ocular system.
- the inversion of the equation can be computationally derived using a least-squares fitting algorithm, a minimization algorithm, or other algorithms designed for curve fitting.
- various coefficients of the Zernike polynomial may provide information on the near-vision or far-vision refraction sphere, refraction cylinder, and refraction axis.
- the curve-fitting algorithm provides information on how much correction will be needed to remedy the errors introduced by the refraction sphere, refraction cylinder, and/or refraction axis and the higher order aberrations of the visual system that are not corrected by sphero-cylindrical corrections.
- FIGS. 1 through 27B provide greater details related to various embodiments of systems and methods for acquiring and processing neuro-ocular wavefront data.
- FIG. 1 is a diagram illustrating an eye 110 a, and various axes defined in relation to the eye 110 a.
- FIG. 1 shows the right eye 110 b, the left eye 110 c, and the nose 120 from the viewpoint of the clinician (or other individual performing the test on the subject).
- the coordinate system is right-handed insofar as the horizontal axis x extends from the left of the clinician (right of the subject) to the right of the clinician (left of the subject).
- the vertical axis y extends inferior to superior.
- the negative y-axis extends downward (toward the feet) while the positive y-axis extends upward (toward the head of the subject).
- the x-axis and the y-axis define the pupillary plane. Since the corneal plane is parallel to the pupillary plane, the x-y plane also defines the corneal plane.
- the z-axis is defined as the axis normal to the pupillary plane. In that regard, the direction of light propagation in the eye 110 a is defined by the positive z-axis.
- the coordinate system is defined using polar coordinates, where r represents the radius from the x-y origin, and ⁇ represents the angle measured between the ray projected from (0, 0) to (x, y) with positive angles corresponding to counter-clockwise rotations.
- the normalized radius, ⁇ is computed by taking the radius, r, and dividing by a normalization radius, r norm . In this document, in most cases, r norm is taken to be 3.5 mm.
- FIG. 1 shows a compensation angle dy along the y-axis. This angle is the angular deviation that corresponds to a degree of refractive error, along the y-direction, for a given point on the eye. It should be appreciated that there is a similar compensation angle dx along the x-direction.
- the two compensation angles dx, dy provide an indication of a correction, as described in greater detail below. For simplicity, a summary of these variables is provided in FIG. 28 .
- FIG. 2 is a diagram illustrating an embodiment of a pupillometry image, which may be used in a refractometer adapted to acquire neuro-ocular wavefront data.
- An example of an aberrometer adapted to acquire neuro-ocular wavefront data can be found in U.S. Pat. No. 6,000,800 (hereinafter, “the '800 patent”), which is incorporated herein by reference, as if set forth in its entirety.
- the visual refractometer of the '800 patent is also referred to as the “InterWave” refractometer.
- the pupillometry image includes an image of the eye 110 , which includes a pupil 210 .
- the pupillometry image includes a pupil tracking box 220 , which dynamically tracks the movement of the pupil 210 .
- the motion of the pupil tracking box 220 is substantially synchronous with the motion of the pupil 210 .
- the pupillometry image also includes a location of the first Purkinje image (hereinafter, “the Purkinje location”) 230 , in addition to images from multiple eye-illumination light emitting diodes (LEDs) 240 .
- the Purkinje location a location of the first Purkinje image
- LEDs eye-illumination light emitting diodes
- FIG. 3 is a diagram illustrating an embodiment of a sampling matrix 320 in a refractometer adapted to acquire neuro-ocular wavefront data. Again, for proper orientation, both the x-axis and the y-axis are shown in FIG. 3 .
- the sampling matrix 320 superimposes onto the location of the pupil 210 as identified by the various LED images 230 , 240 .
- the sampling matrix 320 comprises multiple sampling elements 310 a, 310 b (hereinafter “sampling elements 310 ”) that are arranged to form a grid-like pattern.
- a treatment envelope 330 within the sampling matrix 320 defines the area for treatment. This is shown more clearly in FIG. 4 , where the treatment envelope 330 and the sampling matrix 320 are shown without the sampling elements 310 .
- FIG. 6 is a diagram showing an example video display of a refractometer adapted to acquire neuro-ocular wavefront data.
- the video display includes an eye display 610 , which shows the pupil tracking box and the Purkinje location with reference to the x-y plane.
- the video display comprises a sampling map 620 , which shows the sampling points for acquiring the neuro-ocular wavefront data.
- the video display includes an instrument control panel 630 , which permits an operator to adjust various parameters associated with pupillometry.
- the instrument control panel 630 can include a graphical interface for adjusting the Purkinje illumination.
- FIG. 8 is a diagram showing a video display with an alignment target and a test spot.
- the diagram of FIG. 8 is substantially identical to the diagram of FIG. 6 , with the exception of the tracking display 810 .
- the tracking display 810 comprises an alignment target 820 and a test spot 830 .
- the alignment target 820 is configured as cross-hairs while the test spot 830 is configured as a spot.
- the tracking display 810 represents an image that is displayed to the subject. In that regard, for each of the sampling points in the sampling map 620 , the subject is requested to interactively move the test spot 830 to the center of the alignment target 820 .
- the neuro-ocular wavefront can be represented as a Zernike polynomial according to:
- r norm is the normative radius
- F n ( ⁇ , ⁇ ) are the Zernike function terms
- C n are the coefficients for the neuro-ocular wavefront
- N t is the number of coefficients used to define W.
- ⁇ x 2 + y 2 r norm , 0 ⁇ x 2 + y 2 ⁇ r max [ Eq . ⁇ 5 ] where r max is the largest radius within the sampling map.
- the degree of aberration at any sampling point can be defined by the compensation angles:
- the solution of Eq. 1 can be used to determine various neuro-ocular characteristics as well as various treatments for neuro-ocular aberrations.
- the neuro-ocular system can be characterized by solving for the various coefficients in Eq. 1, and these solutions can be used to design various treatments.
- ⁇ [ Eq . ⁇ 21 ]
- Eq. 26 can be depicted as:
- zero-reference measurements are made in order to remove any misalignment between the subject and the measurement instrument.
- the zero-reference is measured by calculating a baseline angular deviation associated with a central point at the entrance pupil of the subject.
- the reference measurements can be acquired multiple times and averaged in order to increase the accuracy of the zero-reference.
- Eq. 1 can be expanded up to five terms as:
- W ⁇ ( r , ⁇ ) ⁇ C 1 ⁇ r norm ⁇ 2 ⁇ ⁇ ⁇ ⁇ cos ⁇ ⁇ ⁇ + C 2 ⁇ r norm ⁇ 2 ⁇ ⁇ ⁇ ⁇ sin ⁇ ⁇ ⁇ + ⁇ C 3 ⁇ r norm ⁇ 6 ⁇ ⁇ 2 ⁇ cos ⁇ ( 2 ⁇ ⁇ ) + ⁇ C 4 ⁇ r norm ⁇ 3 ⁇ ( 2 ⁇ ⁇ 2 - 1 ) + ⁇ C 5 ⁇ r norm ⁇ 6 ⁇ ⁇ 2 ⁇ sin ⁇ ( 2 ⁇ ⁇ ) + ⁇ higher ⁇ ⁇ order ⁇ ⁇ terms . [ Eq . ⁇ 33 ]
- C 3 ′ C 3 + D cyl ⁇ r norm ⁇ cos ⁇ ( 2 ⁇ ⁇ axi ) 4 ⁇ 6 [ Eq . ⁇ 34 ]
- C 4 ′ C 4 - D sph ⁇ r norm 4 ⁇ 3 - D cyl ⁇ r norm 8 ⁇ 3 ⁇ ⁇ and ⁇ : [ Eq . ⁇ 35 ]
- C 5 ′ C 5 + D cyl ⁇ r norm ⁇ sin ⁇ ( 2 ⁇ ⁇ axi ) 4 ⁇ 6 [ Eq . ⁇ 36 ]
- D sph , D cyl , and ⁇ axi represent the ophthalmic sphere value, the ophthalmic cylinder value, and the ophthalmic axis value, respectively, with:
- FIG. 9 is a diagram showing a coordinate system of the neuro-ocular wavefront data and a coordinate system for a laser ablation map, to be used to direct laser treatment in vision correction that corresponds to the neuro-ocular wavefront data.
- FIG. 9 shows a wavefront coordinate system (left) and a coordinate system for laser ablation map (right).
- the wavefront coordinate system and the ablation map coordinate system have a congruent relationship to each other.
- the center of the wavefront coordinate system is the pupil centroid 910 .
- the center of the ablation map coordinate system is the corneal vertex 920 .
- FIG. 29 provides a summary of the variables associated with the calculations in Eqs. 39 through 42.
- FIG. 10 is a diagram showing an optical zone 1010 and a transition zone 1020 associated with the ablation map of FIG. 9 .
- T A-OZ (X,Y) represents the treatment profile of the optical zone
- the treatment profile for the optical zone can be represented according to:
- R OZ represents the radius of the optical zone
- W min represents the minimum value of the wavefront error (or the most negative value of the wavefront error) over the area of the optical zone
- n c represents the index of refraction of the cornea.
- T 0 represents a constant ablation depth, which is added to facilitate the addition of the transition zone
- T A-Oz (X OZ ,Y OZ ) is the optical zone treatment amount at a point where the optical zone boundary intersects the vector from the pupil center to the point (X,Y)
- r is the distance from the mesopic pupil center to the point (X,Y).
- the function ⁇ (x,y) is a smoothing function that is configured to contour the boundary between the optical zone and the transition zone such that there is a smooth transition between these two zones.
- the smoothing function may be a supergaussian having the form:
- f ⁇ ( x , y ) e - 2 ⁇ ( r r 0 ) 2 ⁇ n [ Eq . ⁇ 48 ] where n is a non-zero, positive integer value.
- Eq. 43 can be modified to include any number of variables, such as, for example, optical parameters, subject parameters, and/or environmental parameters, as long as those variables affect the treatment in a relatively linear manner. Defining the treatment to include such variables as T A-L (X,Y), that treatment can be defined according to:
- the coefficients of Eq. 49 can be determined by known statistical methods, so long as neuro-ocular wavefront data from a sufficient population of subjects has been gathered. Such data can be obtained, for some embodiments, in accordance with one or more of the processes set forth in FIGS. 17 through 27B , below.
- r s n 1 ⁇ cos ⁇ ⁇ ⁇ 1 - n ⁇ 2 ⁇ cos ⁇ ⁇ ⁇ 2 n 1 ⁇ cos ⁇ ⁇ ⁇ 1 + n ⁇ 2 ⁇ cos ⁇ ⁇ ⁇ 2 [ Eq . ⁇ 52 ] with ⁇ 2 representing a complex index of refraction for corneal tissue, n 1 representing a complex index of refraction for incident medium, ⁇ 1 representing an incident angle, and ⁇ 2 representing an internal angle refracted according to Snell's law.
- the incident angle is calculated from corneal topographic data, Z, at point (X,Y) in accordance with:
- the ablation depth can be calculated according to:
- T A - T ⁇ ( X , Y ) ⁇ 1 N L ⁇ ( X , Y ) ⁇ ⁇ t i ⁇ ( X , Y ) [ Eq . ⁇ 60 ]
- N L (X,Y) represents the number of layers at point (X,Y)
- These temporarily-skipped zones can be skipped as a result of the operator choosing to conduct the data acquisition out of order, or, alternatively, these zones can be temporarily skipped due to feedback from the subject that indicates difficulty in providing input.
- the dark blue circles 1130 represent those visual zones for which the subject has provided input.
- the red circles 1120 represent those visual zones that have been permanently skipped due to difficulty in obtaining a reliable reading from the subject.
- the light blue circle 1150 represents the visual zone for which data is currently being obtained.
- neuro-ocular wavefront data can be estimated by obtaining input from the subject for each of the visual zones.
- FIG. 13 is a diagram showing an embodiment of a subject data entry screen for a refractometer adapted to acquire neuro-ocular wavefront data.
- subject information can be used to statistically compute various subject-related or environment-related factors that may affect treatment.
- FIG. 13 shows an embodiment of a data entry screen by which the subject data can be provided.
- the data entry screen comprises an input 1305 for the gender of the subject, an input 1310 for the date of birth of the subject, an input 1315 for the type of examination, and inputs 1320 a, 1320 b (collectively referred to herein as 1320 ) for optical characteristics for each eye.
- the optical characteristics can include the sphere 1330 a, 1330 b, the cylinder 1340 a, 1340 b, and the axis 1350 a, 1350 b for each eye. It should be appreciated that the data entry screen can be customized, as a matter of design, to include other fields and, also, to omit certain fields.
- FIG. 14 is a diagram showing an embodiment of a video display that facilitates on-screen measurements of physical characteristics of the eye.
- FIG. 14 shows, with reference to the x-axis and the y-axis, the pupil tracking box 220 and the Purkinje location 230 as well as the eye illumination LED locations 240 .
- the embodiment of FIG. 14 includes an on-screen measurement tool 1410 that can be adjusted with a cursor (not shown) using a graphical user interface (GUI, not shown).
- GUI graphical user interface
- the on-screen measurement tool 1410 can measure the width of the cornea. It should be appreciated that the system can be configured to accommodate other similar measurements.
- the aberrometer that determines the neuro-ocular wavefront error can be configured in a variety of ways, thereby permitting greater accuracy in diagnosing and treating the visual system of a subject.
- FIG. 16 is a block diagram showing an embodiment of a system configured to aggregate data from multiple spatially resolved refractometers. As noted with reference to Eq. 49, statistical analysis can be performed on neuro-ocular wavefront data.
- the embodiment of FIG. 16 provides one example, among others, of a system that is configured to obtain multiple sets of neuro-ocular wavefront data.
- multiple clinical sites 1605 , 1620 are communicatively coupled to a server 1640 over a network 1635 , such as, for example, the Internet.
- FIG. 16 also shows a client 1655 that is communicatively coupled to the Internet 1635 .
- Each clinical site 1605 , 1620 can include multiple refractometers 1610 a . . . 1610 n (collectively 1610 ), 1625 a . . . 1625 n (collectively 1625 ) that are communicatively coupled to the Internet 1635 through routers 1615 , 1630 that are located at their respective clinical sites 1605 , 1620 . Since networked environments are known in the art, further discussion of the routers 1615 , 1630 and the Internet 1635 is omitted here.
- the client processor 1660 is configured to generate requests to the server 1640 and receive data in response to those queries.
- the client processor 1660 is further configured to perform statistical analysis on the requested data. Since those statistical processes, in addition to the hardware and software for performing the statistical analysis, are known in the art, further discussion of those processor functions is omitted here.
- multiple sets of neuro-ocular wavefront data can be aggregated to better estimate various effects, such as, for example, subject effects, environmental effects, and optical effects.
- FIGS. 17 through 27B show various processes associated with neuro-ocular wavefront data acquisition and analysis. It should be appreciated that the following processes, in some embodiments, can be executed by the above-described systems. Alternatively, the processes of FIGS. 17 through 27B can also be executed with other devices that are configured appropriately.
- FIGS. 17 through 27B can be seen as an interactive process in which a subject (or patient) provides neuro-ocular wavefront data by interacting with a spatially-resolved refractometer.
- a subject or patient
- a spatially-resolved refractometer For example, in some embodiments, the subject is placed in front of a refractometer, and an eye-tracking algorithm tracks the location of the subject's pupil in order to eliminate much of the movement artifacts that may be present due to natural eye movements.
- the eye, and more specifically the pupil is divided into multiple zones. For some embodiments, those zones are defined by a matrix of sampling points, which can be predefined sampling points or customized sampling points.
- the refractive characteristic of each zone is obtained interactively from the subject.
- the obtained data includes neuro-ocular information, rather than merely ocular information.
- a neuro-ocular wavefront can be estimated from those discreet sampling points.
- the neuro-ocular wavefront data provides some information on the neuro-ocular pathways, rather than merely providing information from the eye alone.
- FIG. 17 is a flowchart showing an embodiment of a method for acquiring and processing neuro-ocular wavefront data.
- some embodiments of the process can be broadly seen as comprising two steps.
- the data acquisition step ( 1705 ) comprises the step of obtaining ( 1720 ) neuro-ocular wavefront data from a subject
- the data analysis step ( 1725 ) comprises the step of ascertaining characteristics of a visual system from the obtained neuro-ocular wavefront data.
- the neuro-ocular wavefront data can be obtained ( 1720 ), in some embodiments, using one or more of the systems described with reference to FIGS. 1 through 16 .
- the data analysis step ( 1725 ) comprises the step of correlating ( 1730 ) neuro-ocular wavefront data to parameters associated with the visual system of the subject.
- the parameter associated with the visual system can be, but is not limited to, an optical parameter, a subject parameter, or an environment parameter.
- the optical parameters can include, but are not limited to, photopic pupil diameter, mesopic pupil diameter, cycloplegic pupil diameter, near-vision preoperative refraction sphere, near-vision preoperative refraction cylinder, near-vision preoperative refraction axis, far-vision preoperative refraction sphere, far-vision preoperative refraction cylinder, far-vision preoperative refraction axis, near-vision postoperative refraction sphere, near-vision postoperative refraction cylinder, near-vision postoperative refraction axis, far-vision postoperative refraction sphere, far-vision postoperative refraction cylinder, far-vision postoperative refraction axis, left eye, right eye, asphericity, axis angle, optical zone diameter, transition zone diameter, central pachymetry, spherical aberration as a percent of total root-mean-square (RMS)
- RMS root-mean-square
- the subject parameters can include, but are not limited to, age, side of dominant eye, preference between day vision and night vision, treatment purpose, ethnicity, iris color, gender, or any combination of these subject parameters.
- the environmental parameters can include, but are not limited to, temperature, humidity, microkeratome used for corneal resection, flap size, time elapsed from opening of flap to ablation, surgeon, estimated total time during opening of flap, expected flap thickness, procedure type, scanner used, laser used, day of surgery, location of flap hinge, or any combination of these environmental parameters.
- the data analysis step ( 1725 ) proceeds to calculate ( 1740 ) a correction corresponding to one or more treatments for reducing the anomalies in a visual system of the subject.
- the calculation of the correction can be seen as a calculation of one or more correction factors.
- the correction can correspond to a prescription for spectacles. In other embodiments, the correction can correspond to a prescription for contact lenses.
- the correction can correspond to a treatment for a refractive surgical technique, such as, but not limited to, radial keratotomy (RK), astigmatic keratotomy (AK), automated lamellar keratoplasty (ALK), photorefractive keratectomy (PRK), laser in situ keratomileusis (LASIK), intracorneal ring segments (Intacs), intracorneal lens insertion, laser thermal keratoplasty (LTK), phakic intraocular lenses, or any combination of these refractive surgical techniques.
- the correction factors can correspond to one or more factors that affect an ablation profile for a corresponding refractive surgical method.
- FIG. 18 is a flowchart showing, in greater detail, an embodiment of the data-acquisition step ( 1720 ) from FIG. 17 .
- the step of obtaining ( 1720 ) the neuro-ocular wavefront data can be seen, for some embodiments, as a two-step process.
- the process identifies ( 1820 ) visual zones corresponding to different regions of the eye.
- the process interactively obtains ( 1830 ) information related to the visual zones.
- the information can be obtained using an interactive refractometer, such as, for example, one or more refractometers that have been described with reference to FIGS. 1 through 16 .
- FIG. 19 is a flowchart showing, in greater detail, an embodiment of the step of identifying ( 1820 ) visual zones, from FIG. 18 .
- the process identifies ( 1920 ) an area associated with the entrance pupil of the eye.
- the process overlays ( 1930 ) a virtual matrix onto the identified area.
- the area associated with the entrance pupil of the eye is identified and tracked using the Purkinje method, as described above.
- the virtual matrix is overlayed ( 1930 ) onto the pupil area, that matrix will be in substantially the same place, regardless of spurious eye movements.
- FIGS. 20A and 20B are flowcharts showing, in greater detail, an embodiment of the step of interactively obtaining ( 1830 ) information, from FIG. 18 .
- an embodiment of the process begins ( 2005 ) when a reticule image is projected ( 2010 ) at the center of the eye and data is gathered from the various sampling points or visual zones.
- the process determines ( 2015 ) whether or not all visual zones have been analyzed and, so long as all of the visual zones have not been analyzed, further determines ( 2025 ) whether or not a customized sampling map should be used for the analysis. If the process determines ( 2025 ) that the analysis is customized, then the next customized visual zone is selected ( 2030 ).
- this can be done interactively with the operator of the refractometer through a graphical user interface (GUI) similar to that shown in FIG. 12 .
- GUI graphical user interface
- the process determines ( 2025 ) that the analysis is not customized, then the next visual zone is selected ( 2035 ) from a predefined matrix.
- the predefined matrix can be configured similar to the matrix shown in FIG. 11 . Regardless of whether a predefined visual zone is selected ( 2035 ) or a customized visual zone is selected ( 2030 ), the process continues to FIG. 20B .
- a target image is projected ( 2050 ) at the selected visual zone.
- the subject is queried ( 2055 ) for input, and that input is received ( 2060 ) from the subject.
- the query can be configured as a display having an alignment target and a test spot, similar to the display shown in FIG. 8 .
- the input from the subject can be received, for some embodiments, using a joystick or other input mechanism that permits the subject to align the test spot with the alignment target.
- the received input is then stored ( 2065 ), and the process returns to FIG. 20A to determine ( 2015 ) whether or not all visual zones have been analyzed. In the event that all visual zones have been analyzed, the process exits ( 2020 ).
- the blur image depicts a composite image perceived by the patient viewing a point source at an infinite distance, such as a bright star or a street light at a great distance.
- the pupil is divided into three circular zones having outer diameters of approximately 3.65 mm, 5.16 mm, and 6.32 mm. Rays from an object are traced through the pupil and then to the image plane where they are color coded according to the pupillary zone through which they pass.
- the image plane is determined with spatially resolved refractometry by localizing the patient preferred retinal plane and taking into account the effects of neural processing. This depiction provides a simultaneous estimate of vision quality for three different pupil diameters.
- the blur anatomy diagnostic is very similar to a conventional “spot diagram” used in optical system engineering, with the exception that the color coding provides additional information for different pupil diameters.
- the location where a ray is incident on the retina can be calculated from the wavefront error using:
- the root-mean-square (RMS) blur radius of the blur pattern can be calculated using a standard statistical method.
- the first step in performing this calculation is to determine the coordinates of the blur centroid as:
- the method used to select the zone diameters there is variability associated with the method used to select the zone diameters.
- One embodiment of a method used to determine the zone diameters is denoted herein as the “equal area” method. This method endeavors to select the zone sizes by specifying that the contribution of each zone with relation to the image brightness should be equal for each zone. If the pupil diameter under pharmacological dilation is presumed to be the maximum pupil diameter possible, then the diameter of the i th zone, D i , can be determined from the maximum pupil diameter, D max , and the number of zones, N, using:
- zone size selection is that the peripheral zones get progressively thinner. Besides having the effect of equalizing the intensity contribution of each zone, this method is also amenable for optimizing zone diameters of eyes with third-order spherical aberration. The reason being that, in this approach, zones designed according to Eq. 70 will become narrow at the same rate that the variation in sphere error increases. Therefore, each zone defined using this method will be treated with substantially the amount of spherical correction.
- Another and more comprehensive method of selecting the zone diameters would be to utilize the same optimization technique used to optimize sphere, cylinder, and axis, except that zone diameter now becomes one of the variables as an input to the blur diameter calculation.
- a more complicated merit function can be designed, such that the average blur for each zone is quantified.
- the total vision blur represents the sum of the RMS blur for all three zones.
- FIG. 21 is a flowchart showing, in greater detail, an embodiment of the step of correlating ( 1730 ) the neuro-ocular wavefront data to vision parameters, from FIG. 17 .
- FIG. 21 shows an embodiment in which the blur image is obtained.
- the blur image provides an emulation of the degree of blur that a subject sees without corrected vision. Additionally, depending on the parameters of interest, the blur image can also provide information of how a particular subject sees, should that subject's vision be corrected using various treatments.
- the process begins by computing ( 2120 ) a pupil function from the neuro-ocular wavefront data.
- the pupil function can be calculated in accordance with:
- ⁇ is the wavelength of incident light
- D pupil is the diameter of the entrance pupil
- T(x,y) represents an apodizing function that can be used to model variations in effective pupillary transmission.
- FIGS. 22A and 22B are flowcharts showing, in greater detail, an embodiment of the data-analysis step ( 1725 ) from FIG. 17 .
- some embodiments of the data-analysis step ( 1725 ) begin when acquired data is retrieved ( 2210 ) and a model equation is obtained ( 2215 ).
- the acquired data represents the neuro-ocular waveform, while the model equation represents, for some embodiments, a Zernike polynomial.
- the model equation is an expansion of the Zernike polynomial, similar to that shown in Eq. 33.
- the tilt terms of the Zernike polynomial are set ( 2220 ) to zero, as shown with reference to Eqs. 44 and 45. Thereafter, the sphero-cylinder terms are adjusted ( 2225 ) in accordance with Eqs. 34 through 36. Once the sphero-cylinder terms have been adjusted ( 2225 ), an optical zone ablation profile is computed ( 2230 ). For some embodiments, this is done in accordance with Eq. 46. The process continues to FIG. 22B .
- the process further computes ( 2245 ) a transition zone ablation profile.
- the transition zone ablation profile is computed ( 2245 ) in accordance with Eq. 47.
- the optical zone ablation profile and the transition zone ablation profile are then combined ( 2250 ) to produce a total ablation profile.
- the process determines ( 2255 ) whether or not the boundary between the two zones is smooth. If the boundary is not smooth, then the transition zone ablation profile is recomputed ( 2260 ), and that transition zone ablation profile is combined ( 2250 ) with the optical zone ablation profile.
- FIG. 23 is a flowchart showing, in greater detail, another embodiment of the data-analysis step ( 1725 ) from FIG. 17 .
- the data-analysis step ( 1725 ) can be seen as a two-step process in which an equation is generated ( 2310 ) from the stored inputs, and correction factors are calculated ( 2315 ) by inverting the equation.
- the process of FIG. 23 can be seen as the computation of the coefficients in accordance with one or more of Eqs. 1 through 38.
- various predicted vision parameters are derived ( 2430 ).
- the process determines ( 2435 ) whether or not the vision parameters are similar to the psychometric parameters. If there is a substantial discrepancy between the vision parameters and the psychometric parameters, then the fitting parameters for the equation are adjusted ( 2440 ) and the neuro-ocular wavefront data is fit ( 2425 ) using the adjusted parameters. If the vision parameters and the psychometric parameters are in substantial agreement, then the coefficients of the equation are stored ( 2445 ).
- the deviation between the preoperative neuro-ocular wavefront data and the postoperative neuro-ocular wavefront data is determined ( 2530 ). From the determined deviation, future corrections are calculated ( 2535 ). As seen from the embodiment of FIG. 25 , by having both postoperative and preoperative neuro-ocular wavefront data, future treatments can be refined.
- FIG. 26 is a flowchart showing an embodiment of a method for acquiring and processing neuro-ocular wavefront data, in which multiple sets of data are statistically analyzed.
- multiple sets of neuro-ocular wavefront data are interactively obtained ( 2620 ).
- the multiple sets of data can be obtained from multiple clinical sites, such as those shown in FIG. 16 .
- the interactively-obtained neuro-ocular wavefront data is then stored ( 2630 ).
- the data is stored locally, while, for other embodiments, the data is stored remotely. Thereafter, the stored neuro-ocular wavefront data is statistically analyzed ( 2640 ).
- FIGS. 27A and 27B are flowcharts showing another embodiment of a method for acquiring and processing neuro-ocular wavefront data, in which multiple sets of data are statistically analyzed.
- the process begins with the interactive obtaining ( 2710 ) of preoperative neuro-ocular wavefront data from an initial subject. Thereafter, an initial correction is calculated ( 2715 ) from the preoperative neuro-ocular wavefront data of the initial subject. Also, an initial result of applying that correction is predicted ( 2720 ). This can be done in accordance with one or more of the above-identified Eqs. 1 through 75.
- postoperative neuro-ocular wavefront data is interactively obtained ( 2725 ) from the initial subject.
- a deviation between the postoperative neuro-ocular wavefront data and the preoperative neuro-ocular wavefront data is determined ( 2730 ) for the initial subject.
- the process continues to FIG. 27B , where the various coefficients of the corresponding equations are adjusted ( 2745 ) as a function of the determined deviation. It should be appreciated that the coefficients are adjusted as needed. Thus, for some embodiments, only a subset of the coefficients are adjusted to account for substantial deviations from the predicted values.
- the process of FIG. 27B continues by interactively obtaining ( 2750 ) preoperative neuro-ocular wavefront data for a subsequent subject. Thereafter, a correction is calculated ( 2755 ) for that subject using the adjusted equation. The process then predicts ( 2760 ) a result of applying that calculated ( 2755 ) correction.
- postoperative neuro-ocular wavefront data is interactively obtained ( 2765 ) from that subject, and the deviation between the predicted result and the postoperative neuro-ocular wavefront data for that subject is determined ( 2770 ). From that determined ( 2770 ) deviation, the process further determines ( 2775 ) whether or not the adjustment to the equation was appropriate.
- the process ends, and the derivation of the coefficients for the equation has been sufficiently refined.
- the process recursively readjusts the coefficients of the equation until an appropriate adjustment has been made.
- visual aberrometry is an improvement to both visual acuity and ocular aberrometry, insofar as visual aberrometry uses neuro-ocular wavefront data.
- visual aberrometry determines the patient preferred retinal plane and takes into account the neurological pathways by combining subject feedback with the physio-optical characteristics of the eye.
- the aggregate effect of the neurological pathway, in combination with the characteristics of the eye, provides more accurate information for vision diagnosis, vision treatment, or a combination of both.
- the processor, and the components that perform the various functions of the processor may be implemented in hardware, software, firmware, or a combination thereof.
- the processor and its functional components are implemented in software or firmware that is stored in a memory and that is executed by a suitable instruction execution system.
- the processor and its functional components can be implemented with any or a combination of the following technologies, which are all well known in the art: a discrete logic circuit(s) having logic gates for implementing logic functions upon data signals, an application specific integrated circuit (ASIC) having appropriate combinational logic gates, a programmable gate array(s) (PGA), a field programmable gate array (FPGA), etc.
- the neuro-ocular wavefront program and the statistical regression program which comprise an ordered listing of executable instructions for implementing logical functions, can be embodied in any computer-readable medium for use by or in connection with an instruction execution system, apparatus, or device, such as a computer-based system, processor-containing system, or other system that can fetch the instructions from the instruction execution system, apparatus, or device and execute the instructions.
- a “computer-readable medium” can be any means that can contain, store, communicate, propagate, or transport the program for use by or in connection with the instruction execution system, apparatus, or device.
- the computer-readable medium can be, for example but not limited to, an electronic, magnetic, optical, electromagnetic, infrared, or semiconductor system, apparatus, device, or propagation medium.
- the computer-readable medium would include the following: an electrical connection (electronic) having one or more wires, a portable computer diskette (magnetic), a random access memory (RAM) (electronic), a read-only memory (ROM) (electronic), an erasable programmable read-only memory (EPROM or Flash memory) (electronic), an optical fiber (optical), and a portable compact disc read-only memory (CDROM) (optical).
- an electrical connection having one or more wires
- a portable computer diskette magnetic
- RAM random access memory
- ROM read-only memory
- EPROM or Flash memory erasable programmable read-only memory
- CDROM portable compact disc read-only memory
- the computer-readable medium could even be paper or another suitable medium upon which the program is printed, as the program can be electronically captured via, for instance, optical scanning of the paper or other medium, then compiled, interpreted or otherwise processed in a suitable manner if necessary, and then stored in a computer memory.
Landscapes
- Health & Medical Sciences (AREA)
- Engineering & Computer Science (AREA)
- General Health & Medical Sciences (AREA)
- Public Health (AREA)
- Biomedical Technology (AREA)
- Ophthalmology & Optometry (AREA)
- Life Sciences & Earth Sciences (AREA)
- Medical Informatics (AREA)
- Epidemiology (AREA)
- Surgery (AREA)
- Primary Health Care (AREA)
- Heart & Thoracic Surgery (AREA)
- Nuclear Medicine, Radiotherapy & Molecular Imaging (AREA)
- Animal Behavior & Ethology (AREA)
- Physics & Mathematics (AREA)
- Veterinary Medicine (AREA)
- General Business, Economics & Management (AREA)
- Computer Vision & Pattern Recognition (AREA)
- Business, Economics & Management (AREA)
- Signal Processing (AREA)
- Molecular Biology (AREA)
- Urology & Nephrology (AREA)
- Vascular Medicine (AREA)
- Biophysics (AREA)
- Human Computer Interaction (AREA)
- General Physics & Mathematics (AREA)
- Multimedia (AREA)
- Theoretical Computer Science (AREA)
- Eye Examination Apparatus (AREA)
- Prostheses (AREA)
Abstract
Description
where rnorm is the normative radius, Fn(ρ,θ) are the Zernike function terms, and Cn are the coefficients for the neuro-ocular wavefront and Nt is the number of coefficients used to define W. Eq. 1 can be re-written as:
where:
B n =r norm ·C n [Eq. 3]
x=ρ cos(θ), y=ρ sin(θ) [Eq. 4]
with:
where rmax is the largest radius within the sampling map.
where Xk(x) is the basis function for ƒ, and ak is the coefficient associated with its respective basis function. Given Eq. 8, the solutions for coefficients can be obtained using a least-squares algorithm (or any other fitting routine) such that the merit function:
is minimized. Defining
α=dx [Eq. 10]
and
β=dy [Eq. 11],
the compensation angles for each sampling point can be represented as:
where Np is the number of sample points and
nε(1, 2, . . . , Nt) [Eq. 18].
Simplifying Eq. 20 into a single matrix operation results in:
Using matrix notation, Eq. 22 becomes:
S=G T G+H T H [Eq. 24]
and Eq. 23 becomes:
T=G T α+H Tβ [Eq. 25]
T=SC [Eq. 26]
where C is a vector that contains the Zernike function terms. Alternatively, Eq. 26 can be depicted as:
to which the solution is:
C=inv(S)T [Eq. 28]
or, written differently:
where:
Ω=inv(S) [Eq. 30].
dx i,saved =dx i,acquired −dx ref [Eq. 31]
while the zero-reference for the y-axis is given as:
dy i,saved =dy i,acquired −dy ref [Eq. 32]
where Dsph, Dcyl, and θaxi represent the ophthalmic sphere value, the ophthalmic cylinder value, and the ophthalmic axis value, respectively, with:
where the subscript CP denotes the corneal plane and the subscript MR denotes the eyeglass plane.
x0=PurkinjeOffsetMMx [Eq. 39]
and:
y0=PurkinjeOffsetMMy [Eq. 40],
then the coordinate systems of
x=X+x 0 [Eq. 41]
and:
y=Y+y 0 [Eq. 42].
T A-T(X,Y)=T A-OZ(X,Y)+T A-TZ(X,Y) [Eq. 43].
where TA-OZ(X,Y) represents the treatment profile of the optical zone, and TA-TZ(X,Y) represents the treatment profile of the transition zone. Since the first two coefficients, C1 and C2, of Eq. 1 represent tilt of the wavefront error, which affects the fixation direction of each eye, these terms are effectively removed from ablation calculations by setting:
C1=0 [Eq. 44]
and;
C2=0 [Eq. 45].
where ROZ represents the radius of the optical zone, Wmin represents the minimum value of the wavefront error (or the most negative value of the wavefront error) over the area of the optical zone, and nc represents the index of refraction of the cornea. Additionally, from the definitions of
where T0 represents a constant ablation depth, which is added to facilitate the addition of the transition zone, TA-Oz(XOZ,YOZ) is the optical zone treatment amount at a point where the optical zone boundary intersects the vector from the pupil center to the point (X,Y) , and r is the distance from the mesopic pupil center to the point (X,Y). In Eq. 47, the function ƒ(x,y) is a smoothing function that is configured to contour the boundary between the optical zone and the transition zone such that there is a smooth transition between these two zones. As such, in some embodiments, the smoothing function may be a supergaussian having the form:
where n is a non-zero, positive integer value.
where Wn represents the magnitude of the weighting term associated with the variable ξn, the symbol η represents material properties, κ(x,y) represents an ablation efficiency compensation function, and
represents other spatially-dependent, predictable processes. Here, ξn can represent various factors, such as, for example, photopic pupil diameter, mesopic pupil diameter, cycloplegic pupil diameter, near-vision preoperative refraction sphere, near-vision preoperative refraction cylinder, near-vision preoperative refraction axis, far-vision preoperative refraction sphere, far-vision preoperative refraction cylinder, far-vision preoperative refraction axis, near-vision postoperative refraction sphere, near-vision postoperative refraction cylinder, near-vision postoperative refraction axis, far-vision postoperative refraction sphere, far-vision postoperative refraction cylinder, far-vision postoperative refraction axis, left eye, right eye, asphericity, axis angle, optical zone diameter, transition zone diameter, central pachymetry, corneal topographic measurements, spherical aberration as a percent of total root-mean-square (RMS) aberration, coma as a percent of total RMS aberration, trefoil as a percent of total RMS aberration, high-order aberrations as a percent of total RMS aberration, astigmatism index, corneal width, front surface corneal curvature, back surface corneal curvature, front-to-back alignment, age, side of dominant eye, preference between day vision and night vision, treatment purpose, ethnicity, iris color, gender, temperature, humidity, microkeratome used for corneal resection, flap size, time elapsed from opening of flap to ablation, surgeon, estimated total time during opening of flap, expected flap thickness, procedure type, scanner used, laser used, day of surgery, location of flap hinge, or any combination of these variables, which are also referred to herein as confounding parameters. The coefficients of Eq. 49 can be determined by known statistical methods, so long as neuro-ocular wavefront data from a sufficient population of subjects has been gathered. Such data can be obtained, for some embodiments, in accordance with one or more of the processes set forth in
with ñ2 representing a complex index of refraction for corneal tissue, n1 representing a complex index of refraction for incident medium, θ1 representing an incident angle, and θ2 representing an internal angle refracted according to Snell's law. The incident angle is calculated from corneal topographic data, Z, at point (X,Y) in accordance with:
thereby resulting in:
where R is the distance from the corneal vertex to the point of interest (calculated as R=√{square root over (X2+Y2)}), and φ is the azimuthal angle such that
X 2 +Y 2 +Z 2(1+K)−2ZR 0=0 [Eq. 55]
where K is a conic constant, and R0 is a paraxial radius of curvature, then:
where d0 is the ablation depth when φ=e·φ0 or φ=2.718·φ0, φ is the applied fluence, and φ0 is the ablation threshold. Thus, for a small beam ablation where the laser bean width is much smaller than the pupil diameter, Eq. 57 can be rewritten as:
Given this, a rudimentary approximation of the ablation efficiency can be represented according to:
If treatment is applied in N layers, then the transition zone treatment can be represented as:
where NL(X,Y) represents the number of layers at point (X,Y), and:
with tlayer being a nominal value for the removal rate and ti(X,Y) being the actual material removed per layer at that point. If treatment is presumed to be constant across the pupil plane, then:
Thus, according to Eqs. 60 and 62:
where x, and y are the horizontal and vertical coordinates in the entrance pupil plane. In this formulation, the coordinate position actually represents the perceived location of the ray in object space. Thus, Eqs. 64 and 64 are measured in radians or milliradians.
Once the coordinate of the centroid is determined, the calculation of RMS blur radius is completed according to:
RMS blur radius=ƒ(φ, γ, θ) [Eq. 69]
Thus, for example, when N=3, the outer diameters of the three zones is:
Total Vision Blur=ƒ(φ1, γ1, θ1, 0, D 1)+ƒ(φ2, γ2, θ2 , D 1 , D 2)+ƒ(φ3, γ3, θ3 , D 2 , D 3) [Eq. 72].
The total vision blur represents the sum of the RMS blur for all three zones.
where □ is the wavelength of incident light, Dpupil is the diameter of the entrance pupil, and T(x,y) represents an apodizing function that can be used to model variations in effective pupillary transmission. For example, the Stiles-Crawford effect can be simulated according to:
T(x,y)=e −a(x 2+y2) [Eq. 77]
with a being 0.16 (according to Stiles) or 0.105 (according to Moon and Spencer).
I PSF(α, β)=C 0 ·|F(u,v)|u=α/λ,v=β/λ 2 [Eq. 78]
where (□, □ represents the angular position of a point in the far-field, F(u,v) is defined as the Fourier Transform of P(x,y), and C0 is a constant.
Claims (77)
Priority Applications (2)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
US10/811,368 US7703919B2 (en) | 2003-03-28 | 2004-03-26 | Application of neuro-ocular wavefront data in vision correction |
US12/799,503 US7926944B2 (en) | 2003-03-28 | 2010-04-26 | Application of neuro-ocular wavefront data in vision correction |
Applications Claiming Priority (2)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
US45848003P | 2003-03-28 | 2003-03-28 | |
US10/811,368 US7703919B2 (en) | 2003-03-28 | 2004-03-26 | Application of neuro-ocular wavefront data in vision correction |
Related Child Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
US12/799,503 Division US7926944B2 (en) | 2003-03-28 | 2010-04-26 | Application of neuro-ocular wavefront data in vision correction |
Publications (2)
Publication Number | Publication Date |
---|---|
US20050134799A1 US20050134799A1 (en) | 2005-06-23 |
US7703919B2 true US7703919B2 (en) | 2010-04-27 |
Family
ID=33131800
Family Applications (2)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
US10/811,368 Expired - Fee Related US7703919B2 (en) | 2003-03-28 | 2004-03-26 | Application of neuro-ocular wavefront data in vision correction |
US12/799,503 Expired - Fee Related US7926944B2 (en) | 2003-03-28 | 2010-04-26 | Application of neuro-ocular wavefront data in vision correction |
Family Applications After (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
US12/799,503 Expired - Fee Related US7926944B2 (en) | 2003-03-28 | 2010-04-26 | Application of neuro-ocular wavefront data in vision correction |
Country Status (2)
Country | Link |
---|---|
US (2) | US7703919B2 (en) |
WO (1) | WO2004086952A2 (en) |
Cited By (5)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US8789951B2 (en) | 2012-01-10 | 2014-07-29 | Digitalvision, Llc | Intra-ocular lens optimizer |
US8888289B2 (en) | 2012-01-10 | 2014-11-18 | Digitalvision, Llc | Refractometer with a remote wavefront generator |
US9265458B2 (en) | 2012-12-04 | 2016-02-23 | Sync-Think, Inc. | Application of smooth pursuit cognitive testing paradigms to clinical drug development |
US9380976B2 (en) | 2013-03-11 | 2016-07-05 | Sync-Think, Inc. | Optical neuroinformatics |
US20210333573A1 (en) * | 2020-11-23 | 2021-10-28 | Bingsong Wei | Method for preparing double-sided composite thinning zoom concave lens |
Families Citing this family (17)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US7357509B2 (en) * | 2003-04-28 | 2008-04-15 | University Of Rochester | Metrics to predict subjective impact of eye's wave aberration |
US20060197911A1 (en) * | 2003-05-02 | 2006-09-07 | University Of Rochester | Sharpness metric for vision quality |
US8454160B2 (en) | 2006-02-24 | 2013-06-04 | Amo Development, Llc | Zone extension systems and methods |
US8474974B2 (en) | 2006-02-24 | 2013-07-02 | Amo Development Llc. | Induced high order aberrations corresponding to geometrical transformations |
US7695136B2 (en) * | 2007-08-01 | 2010-04-13 | Amo Development, Llc. | Wavefront refractions and high order aberration correction when wavefront maps involve geometrical transformations |
DE102006036958B4 (en) * | 2006-08-08 | 2009-11-26 | Carl Zeiss Vision Gmbh | Method for determining the corrective properties of a visual aid |
DE102009010467A1 (en) * | 2009-02-26 | 2010-09-09 | Carl Zeiss Vision Gmbh | Method and device for determining the position of the ocular pivot point |
US8486054B2 (en) * | 2009-09-14 | 2013-07-16 | Carl Zeiss Meditec Ag | Optometrist client |
DE102011009473B4 (en) * | 2010-04-28 | 2022-03-17 | Rodenstock Gmbh | Computer-implemented method for calculating a spectacle lens with viewing-angle-dependent prescription data, device for calculating or optimizing a spectacle lens, computer program product, storage medium, method for manufacturing a spectacle lens, and use of a spectacle lens |
DK2592995T3 (en) | 2010-07-16 | 2017-02-13 | Zeiss Carl Vision Inc | PUPIL DEPENDENT WAVE FRONT FRACTION |
US9931199B2 (en) * | 2014-05-05 | 2018-04-03 | Roberto Gustavo ALBERTAZZI | Methods and apparatus for treating keratoconus |
EP3861923A1 (en) * | 2020-02-07 | 2021-08-11 | Essilor International | Method for determining at least one parameter of an eye of a person |
CN111414972B (en) * | 2020-03-30 | 2023-09-05 | 王雁 | XGBoost-based eye refraction correction multisource data automatic classification method |
CN111956177B (en) * | 2020-08-17 | 2022-01-11 | 中山大学中山眼科中心 | System for quantifying deformation of visual object |
CN112415774A (en) * | 2020-12-14 | 2021-02-26 | 上海美沃精密仪器股份有限公司 | Design method of corneal contact lens |
CN113963416B (en) * | 2021-11-05 | 2024-05-31 | 北京航空航天大学 | Eye movement interaction method and system based on laser visual feedback |
DE102023103190A1 (en) | 2023-02-09 | 2024-08-14 | Schwind Eye-Tech-Solutions Gmbh | Method for providing control data for an ophthalmic laser |
Citations (10)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US5258791A (en) | 1990-07-24 | 1993-11-02 | General Electric Company | Spatially resolved objective autorefractometer |
US5480396A (en) | 1994-12-09 | 1996-01-02 | Simon; Gabriel | Laser beam ophthalmological surgery method and apparatus |
US5599340A (en) | 1994-12-09 | 1997-02-04 | Simon; Gabriel | Laser beam ophthalmological surgery method and apparatus |
US5777719A (en) | 1996-12-23 | 1998-07-07 | University Of Rochester | Method and apparatus for improving vision and the resolution of retinal images |
US6000800A (en) | 1998-06-22 | 1999-12-14 | Schepens Eye Research Institute | Coaxial spatially resolved refractometer |
US6099125A (en) | 1998-12-07 | 2000-08-08 | Schepens Eye Research Foundation | Coaxial spatially resolved refractometer |
US6499843B1 (en) | 2000-09-13 | 2002-12-31 | Bausch & Lomb Incorporated | Customized vision correction method and business |
US6511180B2 (en) | 2000-10-10 | 2003-01-28 | University Of Rochester | Determination of ocular refraction from wavefront aberration data and design of optimum customized correction |
US20030107814A1 (en) * | 2001-12-11 | 2003-06-12 | Altmann Griffith E. | Method and apparatus for improving the dynamic range and accuracy of a Shack-Hartmann wavefront sensor |
US20080106698A1 (en) * | 2002-12-06 | 2008-05-08 | Visx, Incorporated | Presbyopia correction using patient data |
Family Cites Families (5)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US577719A (en) * | 1897-02-23 | chitsoff | ||
WO2000066139A2 (en) * | 1999-04-29 | 2000-11-09 | Ista Pharmaceuticals, Inc. | Biochemical methods that eliminate corneal scars, opacification and haze |
US6609793B2 (en) * | 2000-05-23 | 2003-08-26 | Pharmacia Groningen Bv | Methods of obtaining ophthalmic lenses providing the eye with reduced aberrations |
JP2004504105A (en) * | 2000-07-21 | 2004-02-12 | オハイオ ステイト ユニバーシティ | Method and apparatus for refractive eye surgery |
IL137635A0 (en) * | 2000-08-01 | 2001-10-31 | Visionix Ltd | Apparatus for interactive optometry |
-
2004
- 2004-03-26 WO PCT/US2004/009569 patent/WO2004086952A2/en active Application Filing
- 2004-03-26 US US10/811,368 patent/US7703919B2/en not_active Expired - Fee Related
-
2010
- 2010-04-26 US US12/799,503 patent/US7926944B2/en not_active Expired - Fee Related
Patent Citations (11)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US5258791A (en) | 1990-07-24 | 1993-11-02 | General Electric Company | Spatially resolved objective autorefractometer |
US5480396A (en) | 1994-12-09 | 1996-01-02 | Simon; Gabriel | Laser beam ophthalmological surgery method and apparatus |
US5599340A (en) | 1994-12-09 | 1997-02-04 | Simon; Gabriel | Laser beam ophthalmological surgery method and apparatus |
US5777719A (en) | 1996-12-23 | 1998-07-07 | University Of Rochester | Method and apparatus for improving vision and the resolution of retinal images |
US6095651A (en) | 1996-12-23 | 2000-08-01 | University Of Rochester | Method and apparatus for improving vision and the resolution of retinal images |
US6000800A (en) | 1998-06-22 | 1999-12-14 | Schepens Eye Research Institute | Coaxial spatially resolved refractometer |
US6099125A (en) | 1998-12-07 | 2000-08-08 | Schepens Eye Research Foundation | Coaxial spatially resolved refractometer |
US6499843B1 (en) | 2000-09-13 | 2002-12-31 | Bausch & Lomb Incorporated | Customized vision correction method and business |
US6511180B2 (en) | 2000-10-10 | 2003-01-28 | University Of Rochester | Determination of ocular refraction from wavefront aberration data and design of optimum customized correction |
US20030107814A1 (en) * | 2001-12-11 | 2003-06-12 | Altmann Griffith E. | Method and apparatus for improving the dynamic range and accuracy of a Shack-Hartmann wavefront sensor |
US20080106698A1 (en) * | 2002-12-06 | 2008-05-08 | Visx, Incorporated | Presbyopia correction using patient data |
Cited By (6)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US8789951B2 (en) | 2012-01-10 | 2014-07-29 | Digitalvision, Llc | Intra-ocular lens optimizer |
US8888289B2 (en) | 2012-01-10 | 2014-11-18 | Digitalvision, Llc | Refractometer with a remote wavefront generator |
US9265458B2 (en) | 2012-12-04 | 2016-02-23 | Sync-Think, Inc. | Application of smooth pursuit cognitive testing paradigms to clinical drug development |
US9380976B2 (en) | 2013-03-11 | 2016-07-05 | Sync-Think, Inc. | Optical neuroinformatics |
US20210333573A1 (en) * | 2020-11-23 | 2021-10-28 | Bingsong Wei | Method for preparing double-sided composite thinning zoom concave lens |
US11789290B2 (en) * | 2020-11-23 | 2023-10-17 | Bingsong Wei | Method for preparing double-sided composite thinning zoom concave lens |
Also Published As
Publication number | Publication date |
---|---|
WO2004086952A2 (en) | 2004-10-14 |
US20050134799A1 (en) | 2005-06-23 |
US20100231855A1 (en) | 2010-09-16 |
WO2004086952A3 (en) | 2005-06-23 |
US7926944B2 (en) | 2011-04-19 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
US7926944B2 (en) | Application of neuro-ocular wavefront data in vision correction | |
US7320517B2 (en) | Compound modulation transfer function for laser surgery and other optical applications | |
US7762668B2 (en) | Residual accommodation threshold for correction of presbyopia and other presbyopia correction using patient data | |
EP1786312B1 (en) | Ocular wavefront-correction profiling | |
US8911086B2 (en) | Compound modulation transfer function for laser surgery and other optical applications | |
US7377648B2 (en) | Volumetric point spread function for eye diagnosis and treatment | |
EP2872030A1 (en) | Process and apparatus for determining optical aberrations of an eye | |
US8342686B2 (en) | Compound modulation transfer function for laser surgery and other optical applications | |
US7273277B2 (en) | Advanced vision intervention algorithm | |
US11998271B2 (en) | Methods and systems of optical diagnosis for refractive eye exams |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
AS | Assignment |
Owner name: DIGITAL VISION, LLC, GEORGIA Free format text: ASSIGNMENT OF ASSIGNORS INTEREST;ASSIGNORS:THOMPSON, KEITH P.;GARCIA, JOSE R.;STAVER, PHILLIP RANDALL;REEL/FRAME:016202/0712;SIGNING DATES FROM 20050112 TO 20050120 Owner name: DIGITAL VISION, LLC,GEORGIA Free format text: ASSIGNMENT OF ASSIGNORS INTEREST;ASSIGNORS:THOMPSON, KEITH P.;GARCIA, JOSE R.;STAVER, PHILLIP RANDALL;SIGNING DATES FROM 20050112 TO 20050120;REEL/FRAME:016202/0712 |
|
AS | Assignment |
Owner name: VISION SERVICE PLAN, CALIFORNIA Free format text: SECURITY AGREEMENT;ASSIGNOR:DIGITALVISION, LLC;REEL/FRAME:030555/0657 Effective date: 20121106 |
|
AS | Assignment |
Owner name: DIGITALVISION, LLC, GEORGIA Free format text: RELEASE BY SECURED PARTY;ASSIGNOR:VISION SERVICE PLAN;REEL/FRAME:030779/0142 Effective date: 20130627 |
|
AS | Assignment |
Owner name: DIGITALVISION, LLC, GEORGIA Free format text: RELEASE BY SECURED PARTY;ASSIGNOR:VISION SERVICE PLAN;REEL/FRAME:030784/0505 Effective date: 20130627 |
|
AS | Assignment |
Owner name: NATIONAL VISION, INC., GEORGIA Free format text: SECURITY AGREEMENT;ASSIGNOR:DIGITALVISION, LLC;REEL/FRAME:030899/0877 Effective date: 20130701 |
|
FPAY | Fee payment |
Year of fee payment: 4 |
|
FEPP | Fee payment procedure |
Free format text: MAINTENANCE FEE REMINDER MAILED (ORIGINAL EVENT CODE: REM.) |
|
LAPS | Lapse for failure to pay maintenance fees |
Free format text: PATENT EXPIRED FOR FAILURE TO PAY MAINTENANCE FEES (ORIGINAL EVENT CODE: EXP.) |
|
STCH | Information on status: patent discontinuation |
Free format text: PATENT EXPIRED DUE TO NONPAYMENT OF MAINTENANCE FEES UNDER 37 CFR 1.362 |
|
FP | Lapsed due to failure to pay maintenance fee |
Effective date: 20180427 |