US20080031494A1 - Fluorescent light tomography - Google Patents

Fluorescent light tomography Download PDF

Info

Publication number
US20080031494A1
US20080031494A1 US11/829,927 US82992707A US2008031494A1 US 20080031494 A1 US20080031494 A1 US 20080031494A1 US 82992707 A US82992707 A US 82992707A US 2008031494 A1 US2008031494 A1 US 2008031494A1
Authority
US
United States
Prior art keywords
mouse
light
fluorescent
imaging
camera
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Granted
Application number
US11/829,927
Other versions
US7555332B2 (en
Inventor
Bradley Rice
Chaincy Kuo
Daniel Stearns
Heng Xu
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Xenogen Corp
Original Assignee
Xenogen Corp
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Priority claimed from US10/606,976 external-priority patent/US7616985B2/en
Application filed by Xenogen Corp filed Critical Xenogen Corp
Priority to US11/829,927 priority Critical patent/US7555332B2/en
Assigned to XENOGEN CORPORATION reassignment XENOGEN CORPORATION ASSIGNMENT OF ASSIGNORS INTEREST (SEE DOCUMENT FOR DETAILS). Assignors: KUO, CHAINCY, XU, HENG, RICE, BRADLEY W., STEARNS, DANIEL G.
Publication of US20080031494A1 publication Critical patent/US20080031494A1/en
Application granted granted Critical
Publication of US7555332B2 publication Critical patent/US7555332B2/en
Anticipated expiration legal-status Critical
Expired - Lifetime legal-status Critical Current

Links

Images

Classifications

    • AHUMAN NECESSITIES
    • A61MEDICAL OR VETERINARY SCIENCE; HYGIENE
    • A61BDIAGNOSIS; SURGERY; IDENTIFICATION
    • A61B5/00Measuring for diagnostic purposes; Identification of persons
    • A61B5/0059Measuring for diagnostic purposes; Identification of persons using light, e.g. diagnosis by transillumination, diascopy, fluorescence
    • A61B5/0073Measuring for diagnostic purposes; Identification of persons using light, e.g. diagnosis by transillumination, diascopy, fluorescence by tomography, i.e. reconstruction of 3D images from 2D projections
    • GPHYSICS
    • G01MEASURING; TESTING
    • G01NINVESTIGATING OR ANALYSING MATERIALS BY DETERMINING THEIR CHEMICAL OR PHYSICAL PROPERTIES
    • G01N21/00Investigating or analysing materials by the use of optical means, i.e. using sub-millimetre waves, infrared, visible or ultraviolet light
    • G01N21/17Systems in which incident light is modified in accordance with the properties of the material investigated
    • G01N2021/178Methods for obtaining spatial resolution of the property being measured
    • G01N2021/1785Three dimensional
    • G01N2021/1787Tomographic, i.e. computerised reconstruction from projective measurements
    • GPHYSICS
    • G01MEASURING; TESTING
    • G01NINVESTIGATING OR ANALYSING MATERIALS BY DETERMINING THEIR CHEMICAL OR PHYSICAL PROPERTIES
    • G01N21/00Investigating or analysing materials by the use of optical means, i.e. using sub-millimetre waves, infrared, visible or ultraviolet light
    • G01N21/62Systems in which the material investigated is excited whereby it emits light or causes a change in wavelength of the incident light
    • G01N21/63Systems in which the material investigated is excited whereby it emits light or causes a change in wavelength of the incident light optically excited
    • G01N21/64Fluorescence; Phosphorescence
    • G01N2021/6417Spectrofluorimetric devices
    • G01N2021/6419Excitation at two or more wavelengths
    • GPHYSICS
    • G01MEASURING; TESTING
    • G01NINVESTIGATING OR ANALYSING MATERIALS BY DETERMINING THEIR CHEMICAL OR PHYSICAL PROPERTIES
    • G01N21/00Investigating or analysing materials by the use of optical means, i.e. using sub-millimetre waves, infrared, visible or ultraviolet light
    • G01N21/62Systems in which the material investigated is excited whereby it emits light or causes a change in wavelength of the incident light
    • G01N21/63Systems in which the material investigated is excited whereby it emits light or causes a change in wavelength of the incident light optically excited
    • G01N21/64Fluorescence; Phosphorescence
    • G01N2021/6417Spectrofluorimetric devices
    • G01N2021/6421Measuring at two or more wavelengths
    • GPHYSICS
    • G01MEASURING; TESTING
    • G01NINVESTIGATING OR ANALYSING MATERIALS BY DETERMINING THEIR CHEMICAL OR PHYSICAL PROPERTIES
    • G01N21/00Investigating or analysing materials by the use of optical means, i.e. using sub-millimetre waves, infrared, visible or ultraviolet light
    • G01N21/62Systems in which the material investigated is excited whereby it emits light or causes a change in wavelength of the incident light
    • G01N21/63Systems in which the material investigated is excited whereby it emits light or causes a change in wavelength of the incident light optically excited
    • G01N21/64Fluorescence; Phosphorescence
    • G01N21/6428Measuring fluorescence of fluorescent products of reactions or of fluorochrome labelled reactive substances, e.g. measuring quenching effects, using measuring "optrodes"
    • G01N2021/6439Measuring fluorescence of fluorescent products of reactions or of fluorochrome labelled reactive substances, e.g. measuring quenching effects, using measuring "optrodes" with indicators, stains, dyes, tags, labels, marks
    • GPHYSICS
    • G01MEASURING; TESTING
    • G01NINVESTIGATING OR ANALYSING MATERIALS BY DETERMINING THEIR CHEMICAL OR PHYSICAL PROPERTIES
    • G01N21/00Investigating or analysing materials by the use of optical means, i.e. using sub-millimetre waves, infrared, visible or ultraviolet light
    • G01N21/17Systems in which incident light is modified in accordance with the properties of the material investigated
    • G01N21/47Scattering, i.e. diffuse reflection
    • G01N21/49Scattering, i.e. diffuse reflection within a body or fluid

Definitions

  • the present invention relates to imaging with light.
  • the present invention relates to systems and methods for obtaining a three-dimensional representation of a fluorescent probe distribution within a scattering medium.
  • Imaging with light is gaining popularity in biomedical applications.
  • One currently popular light imaging application involves the capture of low intensity light from a biological sample such as a mouse or other small animal. This technology is known as in vivo optical imaging.
  • a light emitting probe inside the sample indicates where an activity of interest might be taking place.
  • cancerous tumor cells are labeled with light emitting reporters or probes, such as fluorescent proteins or dyes.
  • the desirable imaging information often pertains to the location and concentration of the fluorescent source inside the subject, particularly a three-dimensional (3D) characterization of the fluorescent source.
  • 3D three-dimensional
  • the present invention provides systems and methods for obtaining a representation of a fluorescent light distribution inside an animal.
  • the fluorescent light distribution can be used to indicate the presence and location of a fluorescent probe in the animal.
  • the systems and methods determine a representation of the fluorescent light distribution inside the animal.
  • the distribution may indicate the location, size, concentration and/or brightness of one or more fluorescent probes in the animal.
  • the present invention relates to a method for obtaining a three-dimensional representation of a fluorescent probe distribution located inside an animal.
  • the method includes obtaining one or more fluorescent images of at least a portion of the animal.
  • the method also includes obtaining a three dimensional representation of a surface portion of the animal.
  • the method further includes mapping fluorescent image data from the one or more fluorescent images to the three dimensional representation of the surface portion of the animal to create fluorescent light emission data from the surface portion of the animal.
  • the method additionally includes determining a three-dimensional representation of the fluorescent probe distribution internal to the animal using the fluorescent light emission data from the surface portion of the animal.
  • a first fluorescent image includes a first trans-illumination position for an excitation light source relative to a camera; a second fluorescent image includes a second trans-illumination position for the excitation light source relative to the camera
  • the three-dimensional imaging accommodates for autofluoresence in the animal being imaged.
  • light imaging methods determine autofluoresence data in the animal and alter fluorescent light emission data from a surface of the animal with the autofluorescence data, before determining the three-dimensional representation of the fluorescent light distribution internal to the animal.
  • the present invention relates to an imaging system for obtaining a representation of a fluorescent probe distribution located inside an animal.
  • the imaging system comprises an imaging chamber and a processing system.
  • the imaging chamber includes a set of walls enclosing an interior cavity, a stage configured to support the animal within the interior cavity, a fluorescent excitation source, and a camera.
  • the processing system includes a processor and memory.
  • the memory includes instructions for obtaining one or more fluorescent images of at least a portion of the animal and instructions for obtaining a three dimensional representation of a surface portion of the animal.
  • the memory also includes instructions for mapping fluorescent image data from the one or more fluorescent images to the three dimensional representation of the surface portion of the animal to create fluorescent light emission data from the surface portion of the animal.
  • the memory further includes instructions for determining a three-dimensional representation of the fluorescent probe distribution internal to the animal using the fluorescent light emission data from the surface portion of the animal.
  • the present invention relates to logic encoded in one or more tangible media for execution and, when executed, operable to obtain a three-dimensional representation of a fluorescent probe distribution located inside a mammal.
  • FIG. 1 shows a simplified pictorial of diffusive light propagation into, through, and out from, a mouse.
  • FIG. 2 illustrates a method for obtaining a 3D representation of a fluorescent probe distribution located inside a mammal in accordance with one embodiment of the present invention.
  • FIG. 3 shows a process flow for fluorescent image capture suitable for use in the method of FIG. 2 .
  • FIG. 4 shows sample images of: autofluorescence of a mouse, fluorescence of an internal probe in the mouse; and a combination of the fluorescence and autofluorescence.
  • FIG. 5 illustrates a sample relationship for converting 2D camera data to surface data for a sample surface element.
  • FIG. 6A schematically shows trans-illumination in accordance with one embodiment.
  • FIG. 6B schematically shows epi-illumination in accordance with one embodiment.
  • FIG. 7 illustrates sample images, each taken with a different trans-illumination position of an excitation light source.
  • FIG. 8 shows sample side/lateral and top/dorsal view of a mouse representation that includes volume elements suitable for use in the method of FIG. 7 .
  • FIG. 9 shows a process flow for obtaining a three-dimensional representation of a fluorescent probe distribution located inside mouse in accordance with a specific embodiment of the present invention.
  • FIG. 10 expands upon a model used in the method of FIG. 9 .
  • FIG. 11 illustrates a schematic diagram of a planar approximation for light emitting from a mouse.
  • FIG. 12 illustrates a schematic diagram that models excitation light as it enters the mouse.
  • FIGS. 13A and 13B show sample reconstructed results for a fluorescent probe distribution within a phantom mouse and real mouse, respectively.
  • FIGS. 14A and 14B illustrate an imaging system configured to capture photographic, fluorescent and structured light images of a mouse in accordance with one embodiment of the present invention.
  • FIG. 1 shows a simplified pictorial of fluorescent light imaging in a mouse 2 .
  • the fluorescent probe distribution may include light data that describes one or more of: an estimated light intensity of one or more fluorescent probes 5 in mouse 2 , an estimated location of the one or more fluorescent probes 5 , an estimated size or shape of a distribution of fluorescent probes 5 , and/or spectral characteristics of the one or more fluorescent probes 5 .
  • a single fluorescent probe 5 is reconstructed as a point.
  • the probe 5 is reconstructed as a complex structure with dimensions characterized spatially in 3D.
  • the fluorescent light distribution in FIG. 1 shows a single probe 5
  • the mouse 2 may include multiple sites and fluorescent probes 5 .
  • the remaining discussion will mainly refer to a single internal fluorescent probe 5 ; it is understood that tomographic reconstruction and processing described herein is well suited for finding and reconstructing multiple fluorescent probes 5 in a mouse 2 or other object being imaged.
  • Fluorescent probe 5 generally refers to any object or molecule that produces fluorescent light.
  • the fluorescent probe 5 absorbs incident energy of a certain wavelength or wavelength range and, in response, emits light energy at a different wavelength or wavelength range.
  • the absorption of light is often referred to as the “excitation”, while the emission of longer wave lights as the “emission”.
  • the output wavelength range is referred to herein as ‘output spectrum’.
  • Fluorescent probe 5 may include one or more fluorescent light emitting molecules, called ‘flourophores’.
  • a flourophore refers to a molecule or a functional group in a molecule that absorbs energy of a specific wavelength and re-emits energy at a different wavelength. Many commercially available fluorophores are suitable for use with mouse 2 .
  • Suitable fluorophores include Qdot® 605, Qdot® 800, AlexaFluor® 680 and AlexaFluor® 750 as provided by Invitrogen of San Diego, Calif. Both organic and inorganic substances can exhibit fluorescent properties, and are suitable for use with fluorescent probe 5 . In one embodiment, fluorescent probe 5 emits light in the range of about 400 nanometers to about 1300 nanometers.
  • the fluorescent probe distribution may be internal to any of a variety of light-emitting objects, animals or samples that contain light-emitting molecules.
  • Objects may include, for example, tissue culture plates and multi-well plates (including 96, 384 and 864 well plates).
  • Animals including a fluorescent probe distribution may include mammals such as a human, a small mammal such as a mouse, cat, primate, dog, rat or other rodent.
  • Other animals may include birds, zebra-fish, mosquitoes and fruit flies, for example.
  • Other objects and samples are also suitable for use herein, such as eggs and plants.
  • the remaining disclosure will show and describe a mouse 2 as an imaging object that contains a fluorescent probe.
  • FIG. 1 shows a simplified pictorial of diffusive light propagation into, through, and out from, mouse 2 .
  • An excitation light source 4 produces incident light 6 that enters a portion of mouse 2 .
  • the incident light 6 scatters in the mouse tissues and some of it eventually reaches an internal fluorescent probe 5 .
  • fluorescent probe 5 When excited by incident light 6 , fluorescent probe 5 emits fluorescent light 7 from within mouse 2 .
  • the fluorescent photons 7 scatter and travel through tissue in the mouse to one or more surfaces 9 ; the light emitted from the surface may then be detected by a camera 20 .
  • a difficulty in tomographic imaging mice is that the complex surface of the mouse will change with each mouse, and potentially each time the mouse is imaged (as its position and body shifts).
  • the probe may also change each time the mouse is imaged—in position, size, strength, and spectral distribution.
  • the difficulty is determining the 3D parameters of an internal fluorescent probe distribution, such as the 3D location, size and brightness distribution of fluorescent probe 5 , given that many parameters needed for tomographic imaging may change with each trial.
  • Topographic imaging determines a surface representation of an object, or a portion thereof.
  • the present invention uses structured light to determine a surface topography for at least a portion of the mouse.
  • Tomographic imaging refers to information inside the mouse surface.
  • An exemplary illustration of topographic vs. tomographic imaging uses a 2D planar slice through the mouse: topography gives the surface (the outer bounding line), while tomography provides information inside the bounding surface.
  • tissue in mouse 2 also autofluoresces.
  • Tissue autofluorescence may act as a source of background or noise to tomographic imaging of a fluorescent probe distribution, and techniques described below also a) model autofluorescence and b) separate the contributions of tissue autofluorescence from light emitted from the mouse surface. This isolates light emitted from the mouse surface that corresponds to fluorescent probe 5 .
  • the present invention overcomes these difficulties and permits real-time fluorescent tomographic imaging, despite variability and complexity of the mouse surface, the effects of autofluorescence, or internal fluorescent probe distribution.
  • simplifying approximations to a photon diffusion model are implemented in order to expedite the computation time required to perform a reconstruction of the light corresponding to fluorescent probe 5 .
  • reconstruction times of less than 5 minutes may be achieved—compared with hours or days for methods that use FEM or Monte Carlo modeling.
  • FIG. 2 illustrates a method 200 for obtaining a 3D representation of a fluorescent light distribution located inside a mouse in accordance with one embodiment of the present invention. Processes in accordance with the present invention may include up to several additional steps not described or illustrated herein in order not to obscure the present invention.
  • Method 200 obtains one or more fluorescent images of at least a portion of the mouse ( 202 ).
  • the images include fluorescent image data that describes fluorescent light emitted from the mouse.
  • the images may be recalled from memory (previously captured) and/or captured in real time by a camera and imaging system, such as that described below with respect to FIGS. 14A and 14B .
  • FIG. 3 describes one suitable method for image capture.
  • the fluorescent image data describes light that falls upon a camera or other photon detector that is distant from the mouse.
  • the fluorescent image data is stored in the images in two-dimensions (2D).
  • Method 200 maps the 2D fluorescent image data onto a surface of the mouse ( 206 ). Before the mapping can occur, method 200 obtains a surface representation of at least a portion of the mouse ( 204 ).
  • the surface portion may include all of the mouse, or a smaller portion. Typically, this portion includes parts of the mouse that the fluorescent image data will be mapped onto.
  • the surface representation refers to a mathematical description or approximation of the actual surface of the mouse, or a portion thereof.
  • the surface representation need not include the entire mouse, and may include a portion of the mouse relevant to a particular imaging scenario. With a mouse for example, the surface representation might not necessarily include distal portions of the tail and distal portions of every foot. Thus, the surface representation is meant to broadly refer to any surface portion of the mouse and not necessarily the entire mouse.
  • the surface representation includes one or more surface elements or regions of interest on the sample that produce surface light emission data related to the internal probe.
  • the surface representation is often displayed in a pictorial depiction such as a 3D depiction (see FIGS. 13A and 13B ).
  • Suitable techniques to obtain a surface representation include structured light, or another imaging modality such as computer tomography (CT) or magnetic resonance imaging (MRI), for example.
  • CT computer tomography
  • MRI magnetic resonance imaging
  • the surface representation may be divided into a surface mesh comprising a set of surface elements, as will be described below.
  • structured light is used to obtain a surface representation of the mouse.
  • Structured light uses a set of lines of light that are projected down on the mouse at an angle (at about 30 degrees, for example) to the surface normal.
  • the mouse generates structured light surface information as each light line reacts to the shape of the animal. Cumulatively, the lines of light each bend or alter in spacing as they pass over the mouse.
  • the structured light surface information can be measured by a camera and used to determine the height of the surface at surface portions of the mouse that are illuminated by the structured light source. These surface portions are the portions of the mouse that face the camera (for a current position of the mouse relative to the camera). The position of the mouse relative to the camera may be changed to gain multiple structured light images and structured light information from multiple views.
  • a camera captures the structured light surface information, digitizes the information and produces one or more structured light images.
  • a processor operating from stored instructions, produces a 3D surface representation of the mouse—or a portion of the object facing the camera—using the structured light information. More specifically, a processing system, running on stored instructions for generating a topographic representation (a surface map) from the structured light surface information, builds a 3D topographic representation of the mouse using the structured light surface information. If multiple views are used, structured light topographies from these multiple views may be “stitched together” to provide a fuller surface representation from different angles. Structured light image capture, hardware and processing suitable for use with a mouse is described further in commonly owned and pending patent application Ser. No. 11/127,842 and entitled “Structured Light Imaging Apparatus”, which is incorporated herein by reference in its entirety.
  • process flow 200 maps the fluorescent image data in the 2D fluorescent images to fluorescent image data at a surface of the mouse ( 206 ). This converts 2D light data collected at a camera to 3D light data at a 3D surface of the mouse. In one embodiment, the mapping converts radiance data from the fluorescent images to photon density just inside the surface.
  • the mapping manipulates 2D camera data according to the geometry between the mouse surface and the camera lens to derive values of the light emission intensity (or radiance) at the surface.
  • a variety of techniques can be used to map camera light data to the mouse surface.
  • the mapping uses a simple 3D translation based on the relative position between the camera and mouse surface.
  • four spatial relationships between a camera and a stage that supports the mouse include: the camera and stage both do not move, the camera moves relative to the stage, the stage moves relative to the camera, and the camera and stage both move relative to each other. In any of these cases, the position of the camera relative to the stage/mouse is known. This permits a 3D translation using the known coordinates from the camera to the mouse surface.
  • FIG. 5 illustrates a relationship for converting 2D camera data to surface data for a sample surface element 402 .
  • FIG. 5 shows a relationship between surface element 402 (on the mouse), image 404 , and an entrance pupil or camera 406 of an imaging system. Light emitted from surface element 402 passes through entrance pupil 406 and is recorded in image 404 .
  • the angle of emission with respect to the surface normal is ⁇ , which is known from the surface topography determined in 204 .
  • the entrance pupil 406 subtends a small solid angle d ⁇ .
  • the imaging system may collect light emitted from surface element 402 on the sample at an angle ⁇ (measured with respect to the normal to surface element 402 ) into the solid angle d ⁇ subtended by the entrance pupil. This information may then be used to convert image data obtained by the camera into the surface emission intensity corresponding to the surface geometry.
  • Emission of light from a mouse surface may be specified in units of radiance, such as photons/sec/cm 2 /steradian.
  • an imaging system captures images of the mouse and reports surface intensity in units of radiance.
  • Surface radiance can be converted to photon density just inside the mouse surface, using a model for photon propagation at the tissue-air interface, as described below.
  • the mapping may produce a surface emission data vector that includes photon density at each surface element for the mouse topography. The photon density just inside the surface are then related to a light emitting probe distribution inside the mouse tissue using a diffusion model.
  • method 200 then models the contributions of tissue autofluorescence to the light emitted from the mouse ( 208 ).
  • Autofluorescence refers to the natural fluorescence of substances within a material or organism. Mammalian tissue has autofluorescence properties that will affect fluorescent imaging.
  • a camera receives image data that includes both: a) light escaping from the mouse surface due to autofluorescence of tissue in the mouse, and b) light escaping from the mouse surface due to fluorescent probe 5 . From a camera's perspective, these two contributions are often mixed.
  • autofluorescence is determined by measurements made in control animals (animals without a fluorescent probe). In this case, an average autofluorescence yield per unit volume of tissue can be derived from images of autofluorescence. The autofluorescence yield can then be used in a forward model of light propagation, e.g., see Eq. 7 below and its associated description.
  • autofluorescence determination techniques may be used. For example, when performing transillumination fluorescence measurements, the autofluorescence pattern on a surface of the animal typically matches a transillumination pattern observed when imaging with an emission filter and no excitation filter. This provides another way to determine and correct for autofluorescence.
  • the process may include: 1) acquiring a normal fluorescence image with specified excitation filter and emission filter (image 250 ); 2) acquiring another image with excitation filter removed (image 252 ); multiplying image 252 by an appropriate scale factor and subtracting it from image A, thus reducing the contribution of autofluorescence from image A.
  • spectral unmixing Another method to reduce autofluorescence is to use spectral unmixing.
  • images are acquired with an array of different emission or excitation filters.
  • Spectral unmixing software separates autofluorescence from the fluorescent probe emission, using known spectral characteristics of the images.
  • Method 200 subtracts the modeled tissue autofluorescence from the light emitted from the surface (as calculated in 206 ), which isolates the light/signal due to the fluorescent probe 5 . This is shown pictorially in FIG. 4 with two epi-illumination images: image 250 shows the autofluorescence of a mouse; image 252 shows the fluorescence of the internal probe in the mouse.
  • Method 200 then calculates a 3D representation of the fluorescent probe distribution internal to the mouse ( 210 ).
  • a fluorescent probe distribution refers to a description or mathematical depiction of fluorescent light emitters inside the mouse.
  • the fluorescent light corresponds to a fluorescent probe disposed inside the animal.
  • the fluorescent probe may include a fluorescent marker such as a dye molecule, or a fluorescent reporter that produces fluorescent light based on gene expression.
  • Light data internal to the mouse 2 surface generally refers to mathematical representation or approximation of light within the mouse 2 interior. This may include a set of points or volume elements, each characterized by 3D position and a source strength. In one embodiment, the present invention divides the mouse 2 interior into volume elements where each volume element is considered to contain a point light source at its center. A solid mesh of these volume elements then defines a collection of point sources used to approximate light data internal to the mouse and the actual probe distribution within mouse 2 . For example, a solid mesh of cubic volume elements may be used.
  • fluorescent probe 5 includes emits low-intensity light.
  • a low intensity fluorescent probe of the present invention emits light within mouse in the range of about 10 4 to about 10 14 photons/second, depending on probe concentration and excitation light intensity.
  • a fluorescent probe 5 that emits flux in the range of about 10 4 to about 10 10 photons/second is suitable.
  • Other light fluxes are permissible with the present invention.
  • Photons/second is one unit of measure suitable to quantify the amount of light produced by probe 5 .
  • Other units of measure are known to one of skill in the art, such as Watts.
  • the conversion of photons/second to Watts is 3.3 nanowatts equals about 10 10 photons/second at 600 nm.
  • probe 5 emits light between about 10 ⁇ 15 to 10 ⁇ 6 watts of light.
  • the amount of light produced by fluorescent probe 5 refers to the light emitted within mouse 2 —not necessarily the amount of light generated by excitation light source 4 (such as an LED) that generates the light incident on the fluorescent probe 5 .
  • Method 200 uses the fluorescent light emission data from the mouse surface, along with tomographic imaging software that models light propagation internal to the mouse and solves for fluorescent probe distribution.
  • the internal light propagation modeling includes both a) fluorescent excitation light propagation from the excitation light source 4 , and its entry points into the mouse, to the fluorescent probe 5 , and b) fluorescent emission light propagation from the fluorescent probe 5 to the surfaces captured in the fluorescent images.
  • step 210 Tomographic modeling, processing, and fluorescent probe determination of step 210 is described in further detail below with respect to FIGS. 9 and 10 .
  • the resultant 3D representation produced by method 200 may be expressed as a pictorial depiction, e.g., on a computer monitor.
  • FIGS. 13A and 13B show depictions of a mouse with a 3D surface topography and internal fluorescent probe 5 determined using method 200 .
  • Tomographic imaging in method 200 finds use in a wide array of imaging and research applications such as oncology, infectious disease research, gene expression research, and toxicology, for example.
  • the tomographic imaging is suitable for use with samples having a complex surface, such as a mouse.
  • a complex surface is any surface that cannot be described solely using a single polygonal description.
  • the reconstruction techniques described herein place no restrictions on the source distribution, such as the number of probes in the sample or the sizes and shapes of the sources, and no restrictions on the geometry, size or shape of the surface.
  • method 200 may occur in real time where image capture ( 202 ), topographic acquisition ( 204 ) and the data calculations ( 204 - 210 ) all occur without significant delays to a user. In other words, soon after all the images are obtained—e.g., the images are captured or previously captured images are selected and recalled from memory—and the user inputs desired parameters for the tomographic assessment, method 200 outputs 3D details for the internal fluorescent probe 5 . In one embodiment, mapping the fluorescent image data and determining the 3D fluorescent probe distribution (steps 206 - 210 ) finishes in less than about 5 minutes. In another embodiment, details of a fluorescent probe distribution are determined in less than about 1 minute.
  • a video display may then show a pictorial representation of the tomographic reconstruction output on a monitor to a user.
  • This quick processing allows a user to repeat process flow 200 —or change parameters in the tomographic assessment relatively easily. This increases researcher productivity. If the mouse is in an imaging box and under anesthesia, this also permits multiple imaging sessions, in efficient succession, without a need to handle the mouse between tomographic imaging sessions. This real time imaging permits the mouse to be anesthetized for shorter durations (despite permitting multiple imaging sessions).
  • FIG. 3 shows a process flow for fluorescent image capture 202 ( 202 in FIG. 2 ) according to a specific embodiment of the present invention.
  • image capture 202 occurs with the mouse resting or lying on a horizontal stage or flat surface in a resting or normal position, without the need for any straps holding the mouse.
  • the mouse may be anesthetized to prevent movement during imaging 202 .
  • the stage may then move between image captures, but the mouse typically remains stationary on the stage during image capture with the camera.
  • process flow 202 also captures photographic and structured light images.
  • Other embodiments of fluorescent light image capture 202 need not include photographic and structured light image capture.
  • Image capture 202 uses box 10 of FIGS. 14A and 14B .
  • image capture 202 begins be receiving a mouse in an imaging box ( 220 ). This often occurs when a user places the mouse on a stage within an imaging chamber for the imaging box. The user may also initiate image capture 202 using a computer associated with the imaging system.
  • the stage is movable.
  • the imaging system moves the stage to a desired position according to a control signal provided by a computer in the imaging system ( 222 ).
  • a control signal provided by a computer in the imaging system ( 222 ).
  • a user may input a desired image position via the computer user interface, and the imaging control system moves the stage accordingly.
  • a desired position for the stage may be pre-programmed based on an automated data collection routine that the user initiates.
  • a moving stage allows multiple positions and angles relative to a fixed camera.
  • the camera then captures a structured light image ( 224 ).
  • Structured light image capture may be accomplished using a structured light projection system.
  • the structured light projection system projects structured light down onto the mouse from an angle, and the camera (also above the mouse, or on the same side of the mouse as the projector) captures the altered structured light.
  • Suitable structured light generation systems are described in commonly owned and co-pending patent application Ser. No. 11/127,842.
  • the structured light image data is also transferred to an image processing unit and/or a processor in the imaging system for storage for further processing to build a 3D surface representation.
  • a camera then captures a photographic image ( 226 ).
  • the photographic image data is transferred to an image processing unit and/or a processor in the imaging system for storage.
  • the photographic image may be subsequently used for display.
  • the photographic image may be used in an overlay image that includes both the photographic image and fluorescent probe distribution (output from 210 ).
  • the overlay provides a simple pictorial view to facilitate user visualization of the internal fluorescent probe distribution.
  • the camera then captures a fluorescent light image ( 228 ).
  • Fluorescence imaging illuminates the mouse to excite fluorescence molecules in the internal fluorescent probe, and then captures an image of the mouse, or a portion thereof, as the internal probe fluoresces.
  • Fluorescent image capture provides incident light onto into the mouse with an illumination source. The incident light should be large enough in magnitude to elicit a fluorescent from the probe, but not too large so as to saturate a CCD camera. In response to the incident light, light emits from the “excited” fluorescent probe.
  • Trans-illumination and/or epi-illumination may be used.
  • FIG. 6A schematically shows trans-illumination in accordance with one embodiment.
  • Trans-illumination provides light from a side of the mouse opposite to the camera (e.g., incident light from below and a camera above), so that the light travels through the mouse. This provides lower levels of autofluorescence, which is useful for 3D tomographic reconstructions. Also, the ability to move the transillumination point relative to a fluorescent probe fixed within the animal, provide additional information that is use for 3D tomographic reconstructions.
  • the excitation light source 4 includes a lamp 90 that provides light that passes through a filter in excitation filter wheel 92 , which allows a user to change the spectrum of the incident excitation light.
  • a fiber bundle switch 94 directs the excitation light into one of two paths 95 and 97.
  • Path 95 is used for trans-illumination and directs the incident light along a fiber bundle or cable for provision towards a bottom surface of the mouse 2 .
  • the outlet position of path 95 can be moved or re-directed to create multiple incident excitation light locations of trans-illumination path 95 .
  • Epi-illumination provides the incident light from the same side of the animal that an image is captured (e.g., incident light from above, and a camera above the mouse), and is often referred to as reflection-based fluorescent imaging.
  • FIG. 6B schematically shows epi-illumination in accordance with one embodiment.
  • switch 94 directs the excitation light into path 97 , where it routs to a position above the mouse for provision towards a top surface of the mouse 2 on the same side of the mouse as camera 20 .
  • Epi-illumination provides a faster survey of the entire animal, but may be subject to higher levels of autofluorescence. Both trans-illumination and epi-illumination may be used. Epi-illumination avoids significant light attenuation through the mouse, and may help constrain volume elements near the camera-facing surface of the mouse. For example, the epi-illumination constraints may identify artifact voxels near the top surface, which are then removed by software.
  • an emission filter 98 allows a user to control a spectrum of light received by camera 20 .
  • This combination of excitation filter wheel 92 and emission filter 98 allows images to be captured with numerous combinations of excitation and emission wavelengths.
  • excitation filter wheel 92 includes twelve filters while emission filter 98 includes 24 positions.
  • Imaging may also capture both trans- and epi-illumination images, and combine the data.
  • the light takes a different path through mouse, which provides a different set of input criteria and internal light conditions for tomographic reconstruction calculations.
  • a structured light source 99 also provides structured light onto the top of the animal for structured light image capture by the camera 20 without moving the mouse 2 on the horizontal surface.
  • the stage is moveable, which allows camera 20 to capture images from multiple perspectives relative to the mouse 2 .
  • the stage may move in one dimension (e.g., up and down or side to side) or two dimensions for example.
  • the fluorescent excitation uses a different spectrum than the fluorescent emission.
  • the bandgap between excitation and emission filters will vary with the imaging system used to capture the images. A bandgap of at least 25 nm is suitable for many imaging systems.
  • the excitation spectrum may be achieved using any combination of lights and/or filters.
  • the emission spectrum will depend on a number of factors such as the fluorophore used, tissue properties, whether an emission filter is used before the camera, etc.
  • the transillumination location of the excitation light source is moved to capture multiple images of internal fluorescence and the same set of excitation and emission filters is used for the different excitation light source positions.
  • a camera then captures a fluorescent light image of at least a portion of the mouse ( 228 ).
  • the fluorescent image records fluorescence as a function of 2D position.
  • the image may include the entire mouse, or a portion of interest that has been zoomed in on (optically or digitally).
  • the image is transferred to the image processing unit and/or computer for subsequent processing.
  • Multiple fluorescent light images may be captured with the mouse in its current position ( 230 ). In one embodiment, this is done to facilitate spectral unmixing, where each image capture ( 228 ) uses a different excitation and/or emission spectrum. In another embodiment, multiple images are taken for differing trans-illumination positions of the excitation light source 4 ( FIG. 1 ). Each trans-illumination position provides a different set of input conditions to the tomographic reconstruction.
  • FIG. 7 illustrates 21 sample images 235 , each taken with a different trans-illumination position of the excitation light source.
  • the imaging system is configured to move the excitation light source (or has multiple excitation light sources that are controllably turned on/off) and captures an image of the mouse for each different trans-illumination position of the excitation light source.
  • All of the images 235 may be used in a tomographic reconstruction, or a subset can be used.
  • the subset may be selected based on a quality measure for the images, such as a threshold for number of fluorescent photons collected in each image. Other quality measures may be used to select the images.
  • the number of images captured may vary. In one embodiment, 1 to about 80 different trans-illumination positions and images are suitable for tomographic reconstruction. In a specific embodiment, from about 4 to about 50 images are suitable.
  • the images may be stored for tomographic assessment at a later time, e.g., the images—or a subset thereof—are recalled from memory during tomographic processing.
  • the stage and mouse may then be moved to a second position ( 232 ). While the stage is at the second position, one or more photographic, structured light, and/or fluorescent images of the mouse may be captured ( 224 - 230 ). Image collection may further continue by capturing images of the sample from additional positions and views. For example, image capture may occur at anywhere from 2 to 200 positions of the mouse within an imaging chamber. In general, as more images are captured, more information is gathered for tomographic reconstruction. Also, multiple structured light positions may be used to images more of the mouse in 3D. Eight positions, spaced every 45 degrees about a nose-to-tail axis of the mouse, is suitable in some 3D embodiments to build a stitched together surface representation for 360 degree viewing about the mouse.
  • image capture 202 is automated.
  • a user may initiate software included with an imaging system that controls components of the imaging system responsible for image capture.
  • the user may launch imaging and acquisition software on a computer associated with the imaging system that initializes the camera and carries out imaging automatically.
  • the software may then select a desired stage position if a moveable stage is used, prepare the system for photographic, structured light, and/or fluorescent image capture (e.g., turn on/off lights in the box), focus a lens, selectively position an appropriate excitation or emission filter, select an excitation fluorescent light source (one of many for example), set an f-stop, transfer and store the image data, build a reconstruction, etc.
  • fluorescent image capture software activates the camera to detect photons emitted from the mouse, which usually corresponds to absolute units from the surface. The camera may capture the fluorescent image quickly or over an extended period of time (up to several minutes).
  • Fluorescent imaging often captures image data with multiple reporters; each reporter may have its own wavelength spectrum.
  • a camera image of a mouse with multiple reporters has the spectral results of each reporter mixed together.
  • spectral unmixing is useful to clean fluorescent image data and separate the contributions from each source before tomographic processing.
  • the unmixing may also identify contributions from autofluorescence.
  • a spectral unmixing tool is employed in software to separate fluorescent contributions from multiple sources. This permits fluorescent tomography described herein to image multiple reporters in a mouse independently. For example, one reporter may be used in an imaging application to monitor cell death in the mouse, while the second reporter monitors cell propagation. A user may initiate the spectral unmixing tool and software with an appropriate user interface command.
  • FIG. 9 shows a process flow 300 for obtaining a 3D representation of a fluorescent probe distribution located inside mouse 2 in accordance with a specific embodiment of the present invention.
  • Process flow 300 expands upon method 200 of FIG. 2 , and converts surface light emission data to a mathematical representation of a fluorescent probe distribution within the mouse.
  • Process flow 300 first divides a surface representation for the mouse into a surface mesh that includes a set of surface elements ( 302 ). This may include obtaining a surface topography, if that has not already been done (see 204 in FIG. 2 ).
  • the number of surface elements will vary according to the mouse surface area and a desired solution accuracy for the tomographic reconstruction.
  • the number of surface elements in the set should be large enough to capture photon density details and variation across the mouse surface. For example, between about 100 and about 10,000 surface elements may be suitable for a mouse.
  • Process flow 300 selects a number of images for use in the tomographic assessment ( 304 ). As mentioned above in image capture, not all images previously captured and stored in memory need be used. For example, a user may select images that include a moving trans-illumination light source that is closer to a fluorescent probe compared to other images where the moving trans-illumination light source is farther away from the probe. Epi-illumination images may also be incorporated into process flow 300 .
  • Process flow 300 maps photon data from the images to the surface topography mesh ( 306 ). This may use the mapping techniques described above in 206 of FIG. 2 .
  • the mapping converts surface light data (excitation and/or emission) into light data internal to a surface. Notably, this relates surface emission intensity to photon density just inside the mouse surface.
  • process flow 300 converts values of light emission intensity for each surface element into photon density just inside the surface. Referring briefly to FIG. 5 , the value of emission intensity at a surface element, I( ⁇ 2 ), is related to the photon density ⁇ beneath the surface element. The exact form of the relationship depends on the model used to describe the transport of photons across the surface boundary.
  • I ⁇ ( ⁇ 2 ) c 4 ⁇ ⁇ ⁇ ⁇ n 2 ⁇ T ⁇ ( ⁇ ) ⁇ cos ⁇ ⁇ ⁇ 2 ⁇ d ⁇ ⁇ ⁇ 1 + 3 2 ⁇ 1 - R eff 1 + R eff ⁇ cos ⁇ ⁇ ⁇ ⁇ ⁇ ⁇ ( 1 )
  • c is the speed of light
  • n is the index of refraction of the sample medium
  • T is the transmission coefficient for light exiting the sample through the surface element
  • the internal reflectivity R eff depends on the index of refraction of the medium underneath a surface element.
  • R eff is typically in the range of 0.3-0.5.
  • Eqs. (1) and (2) may thus be used to convert surface emission data measured at each surface element to values of the photon density beneath the surface.
  • process flow 300 then divides the mouse interior volume into volume elements, or ‘voxels’ ( 310 ).
  • FIG. 8 shows sample side/lateral and top/dorsal view of a mouse 295 representation that includes volume elements 297 .
  • each volume element 297 is considered to contain a point light source at its center.
  • a solid mesh of volume elements 297 then defines a collection of point sources used to approximate light in the mouse.
  • Volume elements 297 may also be used as a framework to describe the fluorescent probe distribution 5 .
  • process flow 208 may use a volume element 297 resolution of about 0.5 to about 6 millimeters for a small mammal.
  • a volume element 297 resolution of about 1 millimeter is suitable for some mice.
  • Other volume element 297 sizes and densities may be used.
  • One or more early constraints may also be applied (312) to expedite or simplify the determination, such as applying one or more limits on the modeling and solution-space.
  • the internal light modeling solution space is spatially limited to within the boundaries of the mouse surface.
  • a volume space used within the reconstruction is limited by one or more practical considerations. For example, regions of the internal mouse volume far away from where fluorescent light emission takes place (e.g., the rear of the mouse when the head visibly shows the highest light emission density), as determined by a visual scan of the images, may be excluded from the solution space.
  • Process flow 300 then models light propagation. In the embodiment shown, this occurs in a three-step process where excitation light and emission light are each modeled separately and then the two are combined (314, 316, and 318).
  • tissue scattering and absorption parameters are known a priori, stored in memory, and recalled from memory when a reconstruction occurs. In another embodiment, tissue scattering and absorption parameters are calculated from trans-illumination measurements.
  • the condition for diffusive transport is that the scattering coefficient ⁇ s be greater than the absorption coefficient ⁇ a so that the change in the photon density is small between scattering events.
  • D c 3 ⁇ ( ⁇ A + ⁇ S ′ ) ( 5 )
  • An emission Green's function is a solution to Eq. (9) subject to the boundary condition imposed by the surface of the sample.
  • a Green's functions is used to model internal light propagation.
  • a Green's function mathematically describes light propagation through space, such as through tissue, from one location to another.
  • the Green's function uses volume elements 297 and surface mesh elements as vector spaces for its data elements.
  • an excitation Green's matrix models light propagation from a position of the excitation illumination source to the volume elements 297 ( 314 ).
  • An emission Green's matrix may also be used to model light propagation from the volume elements 297 to the surface elements ( 316 ).
  • the excitation and emission models are then combined (318).
  • the excitation and emission Green's function matrices are coupled together, along with a coupling constant, and form a single fluorescence Green's kernel matrix for the fluorescence forward model.
  • the excitation Green's function and emission Green's function matrices are composed using a hybrid Green's function expression which combines weighted terms of a radial partial current boundary condition and an extrapolated boundary condition. This coupled Green's function may be applied to fluorescence of the probe and/or autofluoresence.
  • Modeling may also include one or more of: a) establishing a relationship between the surface elements and volume elements, b) setting additional limits on the modeling and solution-space, c) deciding whether to use a homogeneous or non-homogeneous model for light propagation in tissue, and/or d) composing a mathematical representation of light internal to the mouse.
  • FIG. 10 describes modeling suitable for use with process flow 300 in more detail.
  • process flow 300 determines the light data internal to the mouse, including the desired fluorescent probe distribution that includes the fluorescent probe ( 320 ). For example, once the Green's function is determined, the distribution is obtained by solving the system of linear equations that relate the photon density at the surface to the source distribution inside the object. In one embodiment, process flow 300 solves for all the internal volume elements. Thus, once the Green's function is modeled and determined, it may be evaluated for every volume element-surface element pair, in order to obtain the system of linear equations (Eq. 7, below). Referring forward to Eq. (7), since ⁇ is known, and G ij can be determined as described below, the reconstruction method then solves the linear system, Eq. (7), for the source strengths S i .
  • process flow 300 uses the non-negative least squares algorithm to solve for the internal fluorescent probe distribution. Other techniques may be used. In some cases where the fluorescent probe distribution includes a spatially smoother solution, Eq. (7) can be augmented using a regularizing matrix in the first derivative.
  • the present invention relies on a simplified analytical approximation (planar boundary condition) for the Green's function as described above.
  • a look-up table can be used for the Green's function.
  • the look-up table may be created by previous measurements of photon transport in a sample (or similar sample approximated to be substantially equal to the current sample), or by computational simulations using techniques such as Monte Carlo or finite element modeling. This particular method is useful for samples consisting of inhomogeneous media, such as animal subjects.
  • the optical properties of the tissue, ⁇ a and ⁇ s may have spatial dependence or other heterogeneous properties.
  • FIG. 13A shows top and side views 380 and 382 of sample reconstructed results for a fluorescent probe distribution within a phantom mouse (e.g., plastic mouse having an embedded fluorescent probe).
  • the reconstructed source 383 shows a fluorescent dye in the phantom mouse whose fluorescent yield values are above 10% of the maximum light value in the reconstructed solution.
  • FIG. 13B shows reconstruction results for a fluorophore in a real mouse.
  • process flow 300 applies an iterative solution process. Iterative processing obtains multiple three-dimensional representations and compares them to improve the final output and assessment for the fluorescent probe distribution. In this case, process flow 300 varies the tomographic assessment or modeling, finds a potentially new of different solution in each iteration, and then selects one of the multiple solutions. Loop 328 , for example, varies the subset of images that were selected from a larger set of images.
  • iterative process flow 210 assesses the solution quality and assigns a quality to each iterative solution ( 322 ).
  • Equation 6 The value of ⁇ 2 measures the difference between the observed photon density ⁇ i and the calculated photon density ⁇ j ⁇ ⁇ G ij ⁇ S j over the surface of the sample.
  • Equation 6 Other terms shows in Equation 6 are described further below with respect to Equations 7-9.
  • loop 330 varies the number and/or size of volume elements.
  • volume element size is initially set, and changed as iteration proceeds.
  • the initial voxelation is relatively coarse and refined with successive iterations. For example, the volume element size may be reduced by a factor of two in a next iteration. If the solution quality improves after this second pass, then the volume element size may be again reduced by a factor of two in a third iteration. If the solution quality doesn't improve or gets worse, then the algorithm may have converged on a final solution and stop.
  • the initial volume element size may range from about 0.1 mm 3 to about 1 cm 3 , and subsequent and/or final volume element size for volume elements close to the source may reduce from about 1 mm 3 to about 10 mm 3 .
  • the initial volume element size may be about 200 mm 3 or about 1 cm 3
  • the final volume element size for volume elements close to the source may reduce to about 1 mm 3 .
  • adaptive meshing increases the density of the solid mesh near the probe to provide increased volumetric information in this space, while density of the solid mesh decreases in areas where no activity of interest is taking place (no light generation or transport).
  • a coarse volume element mesh is initially applied throughout the entire sample volume and the current solution is found, yielding an initial solution for S j .
  • Solution attainment and volume element mesh refinement may then be iterated repeatedly, producing a high-density volume element mesh localized around the fluorescent probe distribution. During each iteration, the quality of the current solution is assessed ( 322 ). In a specific embodiment, the iteration continues until further refinement produces no significant decrease in the assessment value.
  • An additional iterative improvement may be obtained by varying the number of surface elements, Ns, used in obtaining the three-dimensional representation (loop 326 ).
  • Ns used in obtaining the three-dimensional representation
  • the number of surface elements may be used to sample the surface uniformly.
  • process flow 300 iterates for different values of Ns corresponding to sampling the surface element mesh at different densities, and use the quality assessment ( 322 ) to determine the best solution among the different values of N S . For example, if the number of surface elements is between about 100 and about 300 surface elements for a small mouse, an iteration step size between about 10 and 50 may be suitable.
  • FIG. 10 expands modeling light propagation in a mouse in accordance with a specific embodiment of the present invention.
  • Modeling 340 establishes a relationship between the surface elements and volume elements.
  • the reconstruction uses a linear relationship between the source emission strength and the photon density at the surface.
  • the linear relationship is described by a Green's function.
  • the Green's function mathematically and numerically describes the transport of photons inside the mouse, and may accommodate for the effects of non-homogeneities in the volume and internal reflection at the boundary.
  • the Green's function also describes the transport of photons inside the sample from each point or volume element in the distribution to the inside of each surface element.
  • One useful form for the Green's function is a simplified approximation in which the surface of the sample is treated locally as a planar interface oriented tangent to a surface element, as shown in FIG. 12 .
  • the photon density at the surface is the analytical solution for a point source in a semi-infinite slab using the partial-current boundary condition. This allows the Green's function to be calculated with minimal computational expense. Other boundary conditions could be used.
  • Modeling 340 may incorporate additional information and constraints into the solution-space ( 342 ). Additional limits on a Green's function solution-space may apply the location of the input excitation signal at the surface. For example, if a trans-illumination fluorescent image is used, then the known (and potentially changing between images, see 230 in FIG. 3 ) bottom illumination source position may be included in the Green's function and modeling.
  • the model also selects or assigns homogeneous or non-homogeneous properties to the mammalian tissue ( 344 ).
  • a mouse includes a turbid interior.
  • a turbid interior refers to a volume that does not allow unimpeded transport of light.
  • the turbid interior may comprise one or more mediums, structures, solids, liquids, gases, etc.
  • the sample is modeled as homogeneous such that each representative volume element in the sample is characterized by identical light transport properties.
  • the sample is represented as heterogeneous such that various representative volume elements in the sample are characterized by different light transport properties.
  • the interior may comprise a mixture of tissues, bones, organs, etc., each of which may be characterized by separate light transport properties in a heterogeneous model.
  • animals are not fully homogeneous and that tissue absorption for living mammals varies with the type of tissue or tissue cells, and is generally affected by varying particles and quantities such as the presence of hemoglobin.
  • software run by an imaging system may implement homogeneous or heterogeneous assumptions on the optical behavior of mammalian tissue when imaging a living mouse.
  • Green's functions for homogeneous tissue models can be calculated analytically in real time for each imaging example, while more complex heterogeneous models require significant computational effort and may be saved in a look-up table.
  • Data for the fluorophore is then obtained 346 , such as data related to its emission spectrum.
  • an emission spectrum for the fluorophore at one or more wavelengths is provided as an input to the model.
  • the fluorophore(s) used in the fluorescent probe are typically known for an imaging application, and optical properties for the fluorophore wavelengths are also known and may be stored in software prior to imaging.
  • a user selects a wavelength filter, with its predetermined wavelength range, for image capture and the spectrum properties for the fluorophore at that wavelength range are input to the model.
  • the imaging process is automated and a computer recalls spectrum properties for the fluorophore from memory based on an automated wavelength filter selection.
  • a graphical user interface associated with the imaging system may also allow a user to select one or more fluorophores from a list, where information for each fluorophore is stored in a database.
  • Other fluorophore properties may include excitation spectrum and extinction coefficient and quantum efficiency, for example.
  • the photon density at the ith surface element may be approximated by the sum of the contributions from all the volume elements: ⁇ i ⁇ ⁇ j ⁇ ⁇ G ij ⁇ s j ( 7 )
  • ⁇ i photon density at the surface for the ith surface element.
  • ⁇ i is known from the camera image data after it is mapped onto the surface, while G ij is known from the modeling, leaving S j to be solved for.
  • S j refers to the amount of light in each volume element.
  • the Green's function in the linear system thus includes a) a first Green's function that describes the transport of the emission light from the volume elements to the sample surface and b) a second Green's function that describes the transport of the excitation light from the sample surface to the volume elements.
  • the first Green's function, G i E describes the transport of excitation light from the excitation source at the surface of the sample to the i th volume element.
  • the second Green's function, G ij F describes the transport of the fluorescent light from the i th volume element to the j th surface element.
  • Both Green's functions can be determined from analytical expressions, such as the simplified approximation described above in the case of a homogeneous medium, or from look-up tables in the case of an inhomogeneous medium.
  • the excitation and fluorescent light are typically at different wavelengths, and thus the fluorescence does not stimulate additional fluorescence.
  • the autofluorescence term in equation 10 can be subtracted from the measured photon density ( ⁇ j ), resulting in the following equation to be solved for the fluorophore concentration S j fluor : ⁇ i - ⁇ j ⁇ ⁇ G ij ⁇ s j autofluor ⁇ ⁇ j ⁇ ⁇ G ij ⁇ s j fluor ( 11 )
  • this equation can be solved by a non-negative least squares optimization method.
  • Other methods for solving systems of linear equations can also be used.
  • the reconstruction uses a tangential plane boundary approximation combined with a partial current boundary condition to model photon diffusion Green's function for each surface element.
  • FIG. 11 illustrates a schematic diagram showing this emission planar approximation.
  • a plane boundary 352 is drawn tangent to the ith surface element.
  • the photon density in the planar approximation is the solution for the point source at x j in a semi-infinite slab defined by the plane boundary, subject to the partial current boundary condition.
  • the boundary condition is simplified to the case of a plane boundary, although the orientation of the boundary may change for each surface element.
  • the simplified Green's function depends only on the distance between the volume element and the surface, and the angle of the surface element. This method of calculating the Green's function is fast and can be performed in real time for each mouse surface. It is understood that it is not necessary to use this simplified approximation to define the Green's function.
  • FIG. 12 illustrates a schematic diagram showing one suitable excitation approximation, which models light as it enters the mouse.
  • four fictitious point sources, S 0 -S 3 are used to model light onto a volume element.
  • Sources S 0 and S 1 are considered with a partial-current boundary condition, while sources S 2 and S 3 are considered with an extrapolated boundary model.
  • the boundary condition for excitation is simplified to the case of a plane boundary, whose orientation may change for each surface element.
  • planar boundary approximations discussed above work well for smooth surfaces with a large radius of curvature, and for cases where the absorption coefficient is not too small ( ⁇ a >0.1 cm ⁇ 1 ).
  • An advantage of the planar approximation technique described above is that it is computationally convenient for solving the diffusion equation with an arbitrary complex boundary such as a mouse. Areas with more structure, such as the head or the limbs of a mouse, may benefit from a more accurate model of the boundary.
  • Using a finite element modeling code to calculate the Green's functions is one option to obtain a more accurate boundary model. Finite element codes such as Flex PDE, from PDE Solutions, Inc. may be used for example.
  • Another option will be to extend the planar surface approximation to first order in curvature, which may allow continued use of analytic expressions for G ij .
  • process flow 300 has been described with many simplifications to the model to expedite processing, fluorescent tomographic reconstruction is not limited by these simplified computational methods.
  • the Green's Function may be calculated without many of the simplifications described above, even at the cost of increased computational requirements.
  • process flow 300 describes a specific method of obtaining measurements of light emission from the mouse, process flow 300 is not limited to how the light emission data is obtained or to the use of any particular apparatus.
  • light emission data may be obtained from an independent source and stored as data within a computer, and not necessarily produced as the result of imaging via a complementary or local imaging system.
  • process flow 300 may be used to obtain 3D reconstructions of any type of internal light source, including one or more bioluminescent sources.
  • FIGS. 14A and 14B illustrate an imaging system 10 configured to capture photographic, fluorescent and structured light images of a mouse in accordance with one embodiment of the present invention. While tomographic reconstruction will now be described with respect to imaging system 10 , it is understood that the tomographic reconstruction as described herein is well suited for use with other imaging systems.
  • Imaging system 10 may be used for imaging a low intensity fluorescent probe such as fluorescent molecules in a mouse and the like.
  • the low intensity fluorescent probe may be included in any of a variety of living or non-living light-emitting samples.
  • Non-living light-emitting samples may include calibration devices and phantom devices.
  • Living light-emitting samples may include, for example, animals or plants containing light-emitting molecules, tissue culture plates containing living organisms, and multi-well plates (including 96, 384 and 864 well plates) containing living organisms. Animals may include any mammal, such as a mouse or rat containing luciferase-expressing cells.
  • System 10 finds wide use in imaging and research.
  • the ability to track light-emitting cells in a small laboratory animal such as a mouse or rat opens up a wide range of applications in pharmaceutical and toxilogical research. These include in vivo monitoring of infectious diseases, tumor growth in metastases, transgene expression, compound toxicity, and viral infection or delivery systems for gene therapy.
  • the ability to detect signals in real-time and in living animals means that the progression of a disease or biological process can be studied throughout an experiment with the same set of animals without a need to sacrifice for each data point.
  • Imaging system 10 comprises an imaging box 12 having a door 18 and inner walls 19 ( FIG. 14B ) that define an interior cavity 21 that is adapted to receive a mouse 2 in which low intensity light is to be detected.
  • Imaging box 12 is suitable for imaging including the capture of low intensity light on the order of individual photons, for example.
  • Imaging box 12 is often referred to as “light-tight”. That is, box 12 seals out essentially all of the external light from the ambient room from entering the box 12 , and may include one or more seals that prevent light passage into the box when door 18 is closed.
  • door 18 comprises one or more light-tight features such as a double baffle seal, while the remainder of chamber 21 is configured to minimize any penetration of light into cavity 21 .
  • Mouse 2 is placed within box 12 for imaging by opening door 18 , inserting the mouse in chamber 21 , and closing door 18 .
  • Suitable imaging systems are available from Xenogen Corporation from Alameda, Calif., and include the IVIS® Spectrum, IVIS® 3D Series, IVIS® 200 Series, IVIS® 100 Series, and IVIS® Lumina. Further description of a suitable imaging box 12 is provided in commonly owned U.S. Pat. No. 7,113,217 entitled “3-D Imaging Apparatus for In-Vivo Representations”, which is incorporated by reference herein in its entirety for all purposes.
  • imaging system 10 is shown with a single cabinet design, other embodiments of the present invention include a disparate imaging box 12 and desktop computer that includes processing system 28 and a dedicated display.
  • Imaging box 12 includes an upper housing 16 adapted to receive a camera 20 ( FIG. 14B ).
  • a high sensitivity camera 20 e.g., an intensified or a charge-coupled device (CCD) camera, is mounted on top of upper housing 16 and positioned above imaging box 12 .
  • CCD camera 20 is capable of capturing luminescent, fluorescent, structured light and photographic (i.e., reflection based images) images of a living sample or phantom device placed within imaging box 12 .
  • One suitable camera includes a Spectral Instruments 620 Series as provided by Spectral Instruments of Arlington, Ariz.
  • CCD camera 20 is cooled by a suitable source thermoelectric chiller. Other methods, such as liquid nitrogen, may be used to cool camera 20 .
  • Camera may also be side-mounted, or attached to a moving chassis that moves the camera in cavity 21 .
  • Imaging system 10 may also comprise a lens (not shown) that collects light from the specimen or phantom device and provides the light to the camera 20 .
  • a stage 25 forms the bottom floor of imaging chamber 21 and includes motors and controls that allow stage 25 to move up and down to vary the field of view 23 for camera 20 .
  • a multiple position filter wheel may also be provided to enable spectral imaging capability.
  • Imaging box 12 may also include one or more light emitting diodes on the top portion of chamber 21 to illuminate a sample during photographic image capture. Other features may include a gas anesthesia system to keep the mouse anesthetized and/or a heated shelf to maintain an animal's body temperature during image capture and anesthesia.
  • Imaging box 12 also includes one or more fluorescent excitation light sources.
  • box 12 includes a trans-illumination device and an epi-illumination device.
  • the trans-illumination device is configured to direct light into a first surface of the mouse, where diffused light exits a second surface of the mouse.
  • the epi-illumination device is configured direct light onto a third surface of the specimen, where the diffused light exits the third surface of the mouse. Further description of fluorescent excitation light sources is provided in commonly owned and co-pending patent application Ser. No. 11/434,606, which is incorporated by reference in its entirety for all purposes.
  • a structured light source is included in imaging box.
  • the structured light source includes a mechanism for transmitting a set of lines onto the object from an angle.
  • the lines are displaced, or phase shifted relative to a stage, when they encounter an object with finite height, such as a mouse. This phase shift provides structured light information for the object.
  • Camera 20 then captures the structured light information.
  • surface topography data for the object is determined from the phase shift of the lines.
  • FIG. 14B shows system 10 with the removal of a side panel for imaging box 12 to illustrate various electronics and processing components included in system 10 .
  • Imaging system 10 comprises image processing unit 26 and processing system 28 .
  • Image processing unit 26 optionally interfaces between camera 20 and processing system 28 and may assist with image data collection and video data processing.
  • Processing system 28 which may be of any suitable type, comprises hardware including a processor 28 a and one or more memory components such as random-access memory (RAM) 28 b and read-only memory (ROM) 28 c.
  • RAM random-access memory
  • ROM read-only memory
  • Processor 28 a (also referred to as a central processing unit, or CPU) couples to storage devices including memory 28 b and 28 c .
  • ROM 28 c serves to transfer data and instructions uni-directionally to the CPU, while RAM 28 b typically transfers data and instructions in a bi-directional manner.
  • a fixed disk is also coupled bi-directionally to processor 28 a ; it provides additional data storage capacity and may also include any of the computer-readable media described below.
  • the fixed disk may be used to store software, programs, imaging data and the like and is typically a secondary storage medium (such as a hard disk).
  • Processor 28 a communicates with various components in imaging box 12 .
  • processing system 28 employs software stored in memory 28 c that is configured to permit communication with and/or control of components in imaging box 12 .
  • processing system 28 may include hardware and software configured to control camera 20 .
  • the processing hardware and software may include an I/O card, control logic for controlling camera 20 .
  • Components controlled by computer 28 may also include motors responsible for camera 20 focus, motors responsible for position control of a platform supporting the sample, a motor responsible for position control of a filter lens, f-stop, etc.
  • Processing system 28 may also interface with an external visual display (such as computer monitor) and input devices such as a keyboard and mouse.
  • a graphical user interface that facilitates user interaction with imaging system 10 may also be stored on system 28 , output on the visual display and receive user input from the keyboard and mouse.
  • the graphical user interface allows a user to view imaging results and also acts an interface to control the imaging system 10 .
  • One suitable imaging software includes “LivingImage” as provided by Xenogen Corporation of Alameda, Calif.
  • Processing system 28 may comprise software, hardware or a combination thereof.
  • System 28 may also include additional imaging hardware and software, tomographic reconstruction software that implements process flows and methods described above, and image processing logic and instructions for processing information obtained by camera 20 .
  • stored instructions run by processor 28 a may include instructions for i) receiving image data corresponding to light emitted from a mouse as described herein, ii) building a 3-D digital representation of a fluorescent probe internal to a mouse using data included in an image, and iii) outputting results of the tomographic reconstruction on a display such as a video monitor.
  • Imaging system 10 employs a quantitative model that estimates the diffusion of photons in tissue.
  • the model processes in vivo image data and in order to spatially resolve a 3D representation of the size, shape, and location of the light emitting source.
  • imaging apparatus 10 may employ one or more memories or memory modules configured to store program instructions for obtaining a 3D representation of a probe located inside a sample and other functions of the present invention described herein.
  • Such memory or memories may also be configured to store data structures, imaging data, or other specific non-program information described herein. Because such information and program instructions may be employed to implement the systems/methods described herein, the present invention relates to machine-readable media that include program instructions, state information, etc. for performing various operations described herein.
  • tangible machine-readable media include, but are not limited to, magnetic media such as hard disks, floppy disks, and magnetic tape; optical media such as CD-ROM disks; magneto-optical media such as floptical disks; and hardware devices that are specially configured to store and perform program instructions, such as read-only memory devices (ROM) and random access memory (RAM).
  • program instructions include both machine code, such as produced by a compiler, and files containing higher level code that may be executed by the computer using an interpreter.
  • the invention may also be embodied in a carrier wave traveling over an appropriate medium such as airwaves, optical lines, electric lines, etc.

Abstract

Described herein are systems and methods for obtaining a three-dimensional (3D) representation of the distribution of fluorescent probes inside a sample, such as a mammal. Using a) fluorescent light emission data from one or more images, b) a surface representation of the mammal, and c) computer-implemented photon propagation models, the systems and methods produce a 3D representation of the fluorescent probe distribution in the mammal. The distribution may indicate—in 3D—the location, size, and/or brightness or concentration of one or more fluorescent probes in the mammal.

Description

    CROSS-REFERENCE TO RELATED APPLICATIONS
  • This application is a continuation of U.S. application Ser. No. 11/733,358 filed Apr. 10, 2007, which claims priority under 35 U.S.C. §119(e) and is a non-provisional of U.S. Provisional Application No. 60/840,247, filed on Aug. 24, 2006 and titled “Fluorescent Imaging,” by Rice et al.; the Ser. No. 11/733,358 patent application also claims priority under 35 U.S.C. § 120 and is a continuation-in-part of U.S. patent application Ser. No. 10/606,976, filed Jun. 25, 2003 and titled “Method and Apparatus for 3-D Imaging of Internal Light Sources,” which claimed priority under 35 U.S.C. § 119(e) from a) U.S. Provisional Application No. 60/395,357, filed on Jul. 16, 2002 and titled “Method and Apparatus for 3-D Imaging of Internal Light Sources,” by Stearns et al., b) U.S. Provisional Application No. 60/396,458, filed on Jul. 16, 2002 and titled “In Vivo 3D Imaging of Light Emitting Reporters,” by Rice et al. and c) U.S. Provisional Application No. 60/396,313, filed on Jul. 16, 2002 and titled “3D in Vivo Imaging of Light Emitting Reporters,” by Rice et al.; each of the above listed patent applications is incorporated by reference in its entirety for all purposes.
  • FIELD OF THE INVENTION
  • The present invention relates to imaging with light. In particular, the present invention relates to systems and methods for obtaining a three-dimensional representation of a fluorescent probe distribution within a scattering medium.
  • BACKGROUND OF THE INVENTION
  • Imaging with light is gaining popularity in biomedical applications. One currently popular light imaging application involves the capture of low intensity light from a biological sample such as a mouse or other small animal. This technology is known as in vivo optical imaging. A light emitting probe inside the sample indicates where an activity of interest might be taking place. In one application, cancerous tumor cells are labeled with light emitting reporters or probes, such as fluorescent proteins or dyes.
  • Photons emitted by fluorescent cells scatter in the tissue of the mammal, resulting in diffusive photon propagation through the tissue. As the photons diffuse, many are absorbed, but a fraction reaches the surface of the mammal—and can be detected by a camera. Light imaging systems capture images that record the two-dimensional (2D) spatial distribution of the photons emitted from the surface.
  • However, the desirable imaging information often pertains to the location and concentration of the fluorescent source inside the subject, particularly a three-dimensional (3D) characterization of the fluorescent source. Reliable techniques to convert the 2D information in the camera images to a 3D characterization of the fluorescent probe concentration are desirable.
  • SUMMARY OF THE INVENTION
  • The present invention provides systems and methods for obtaining a representation of a fluorescent light distribution inside an animal. The fluorescent light distribution can be used to indicate the presence and location of a fluorescent probe in the animal. Using a) fluorescent light emission data from one or more images, b) a surface representation of at least a portion of the animal, and c) a computer-implemented model for photon propagation in the animal, the systems and methods determine a representation of the fluorescent light distribution inside the animal. The distribution may indicate the location, size, concentration and/or brightness of one or more fluorescent probes in the animal.
  • In one embodiment, the present invention relates to a method for obtaining a three-dimensional representation of a fluorescent probe distribution located inside an animal. The method includes obtaining one or more fluorescent images of at least a portion of the animal. The method also includes obtaining a three dimensional representation of a surface portion of the animal. The method further includes mapping fluorescent image data from the one or more fluorescent images to the three dimensional representation of the surface portion of the animal to create fluorescent light emission data from the surface portion of the animal. The method additionally includes determining a three-dimensional representation of the fluorescent probe distribution internal to the animal using the fluorescent light emission data from the surface portion of the animal.
  • In another embodiment, multiple fluorescent images are captured. A first fluorescent image includes a first trans-illumination position for an excitation light source relative to a camera; a second fluorescent image includes a second trans-illumination position for the excitation light source relative to the camera
  • In yet another embodiment, the three-dimensional imaging accommodates for autofluoresence in the animal being imaged. In this case, light imaging methods determine autofluoresence data in the animal and alter fluorescent light emission data from a surface of the animal with the autofluorescence data, before determining the three-dimensional representation of the fluorescent light distribution internal to the animal.
  • In still another embodiment, the present invention relates to an imaging system for obtaining a representation of a fluorescent probe distribution located inside an animal. The imaging system comprises an imaging chamber and a processing system. The imaging chamber includes a set of walls enclosing an interior cavity, a stage configured to support the animal within the interior cavity, a fluorescent excitation source, and a camera. The processing system includes a processor and memory. The memory includes instructions for obtaining one or more fluorescent images of at least a portion of the animal and instructions for obtaining a three dimensional representation of a surface portion of the animal. The memory also includes instructions for mapping fluorescent image data from the one or more fluorescent images to the three dimensional representation of the surface portion of the animal to create fluorescent light emission data from the surface portion of the animal. The memory further includes instructions for determining a three-dimensional representation of the fluorescent probe distribution internal to the animal using the fluorescent light emission data from the surface portion of the animal.
  • In another embodiment, the present invention relates to logic encoded in one or more tangible media for execution and, when executed, operable to obtain a three-dimensional representation of a fluorescent probe distribution located inside a mammal.
  • These and other features of the present invention will be described in more detail below in the detailed description of the invention and in conjunction with the following figures.
  • BRIEF DESCRIPTION OF THE DRAWINGS
  • The present invention is illustrated by way of example, and not by way of limitation, in the figures of the accompanying drawings and in which like reference numerals refer to similar elements and in which:
  • FIG. 1 shows a simplified pictorial of diffusive light propagation into, through, and out from, a mouse.
  • FIG. 2 illustrates a method for obtaining a 3D representation of a fluorescent probe distribution located inside a mammal in accordance with one embodiment of the present invention.
  • FIG. 3 shows a process flow for fluorescent image capture suitable for use in the method of FIG. 2.
  • FIG. 4 shows sample images of: autofluorescence of a mouse, fluorescence of an internal probe in the mouse; and a combination of the fluorescence and autofluorescence.
  • FIG. 5 illustrates a sample relationship for converting 2D camera data to surface data for a sample surface element.
  • FIG. 6A schematically shows trans-illumination in accordance with one embodiment.
  • FIG. 6B schematically shows epi-illumination in accordance with one embodiment.
  • FIG. 7 illustrates sample images, each taken with a different trans-illumination position of an excitation light source.
  • FIG. 8 shows sample side/lateral and top/dorsal view of a mouse representation that includes volume elements suitable for use in the method of FIG. 7.
  • FIG. 9 shows a process flow for obtaining a three-dimensional representation of a fluorescent probe distribution located inside mouse in accordance with a specific embodiment of the present invention.
  • FIG. 10 expands upon a model used in the method of FIG. 9.
  • FIG. 11 illustrates a schematic diagram of a planar approximation for light emitting from a mouse.
  • FIG. 12 illustrates a schematic diagram that models excitation light as it enters the mouse.
  • FIGS. 13A and 13B show sample reconstructed results for a fluorescent probe distribution within a phantom mouse and real mouse, respectively.
  • FIGS. 14A and 14B illustrate an imaging system configured to capture photographic, fluorescent and structured light images of a mouse in accordance with one embodiment of the present invention.
  • DETAILED DESCRIPTION OF THE PREFERRED EMBODIMENTS
  • In the following detailed description of the present invention, numerous specific embodiments are set forth in order to provide a thorough understanding of the invention. However, as will be apparent to those skilled in the art, the present invention may be practiced without these specific details or by using alternate elements or processes. In other instances well known processes, components, and designs have not been described in detail so as not to unnecessarily obscure aspects of the present invention.
  • Systems and methods described herein obtain a three-dimensional (3D) representation of a fluorescent probe distribution inside a sample. This is referred to as fluorescent tomographic imaging. FIG. 1 shows a simplified pictorial of fluorescent light imaging in a mouse 2.
  • The fluorescent probe distribution may include light data that describes one or more of: an estimated light intensity of one or more fluorescent probes 5 in mouse 2, an estimated location of the one or more fluorescent probes 5, an estimated size or shape of a distribution of fluorescent probes 5, and/or spectral characteristics of the one or more fluorescent probes 5. In one embodiment, a single fluorescent probe 5 is reconstructed as a point. In another embodiment, the probe 5 is reconstructed as a complex structure with dimensions characterized spatially in 3D. Although the fluorescent light distribution in FIG. 1 shows a single probe 5, the mouse 2 may include multiple sites and fluorescent probes 5. For simplicity, the remaining discussion will mainly refer to a single internal fluorescent probe 5; it is understood that tomographic reconstruction and processing described herein is well suited for finding and reconstructing multiple fluorescent probes 5 in a mouse 2 or other object being imaged.
  • Fluorescent probe 5 generally refers to any object or molecule that produces fluorescent light. The fluorescent probe 5 absorbs incident energy of a certain wavelength or wavelength range and, in response, emits light energy at a different wavelength or wavelength range. The absorption of light is often referred to as the “excitation”, while the emission of longer wave lights as the “emission”. The output wavelength range is referred to herein as ‘output spectrum’. Fluorescent probe 5 may include one or more fluorescent light emitting molecules, called ‘flourophores’. A flourophore refers to a molecule or a functional group in a molecule that absorbs energy of a specific wavelength and re-emits energy at a different wavelength. Many commercially available fluorophores are suitable for use with mouse 2. Suitable fluorophores include Qdot® 605, Qdot® 800, AlexaFluor® 680 and AlexaFluor® 750 as provided by Invitrogen of San Diego, Calif. Both organic and inorganic substances can exhibit fluorescent properties, and are suitable for use with fluorescent probe 5. In one embodiment, fluorescent probe 5 emits light in the range of about 400 nanometers to about 1300 nanometers.
  • The fluorescent probe distribution may be internal to any of a variety of light-emitting objects, animals or samples that contain light-emitting molecules. Objects may include, for example, tissue culture plates and multi-well plates (including 96, 384 and 864 well plates). Animals including a fluorescent probe distribution may include mammals such as a human, a small mammal such as a mouse, cat, primate, dog, rat or other rodent. Other animals may include birds, zebra-fish, mosquitoes and fruit flies, for example. Other objects and samples are also suitable for use herein, such as eggs and plants. For ease of discussion, the remaining disclosure will show and describe a mouse 2 as an imaging object that contains a fluorescent probe.
  • Animal tissue is a turbid medium, meaning that photons are both absorbed and scattered as they propagate through tissue. FIG. 1 shows a simplified pictorial of diffusive light propagation into, through, and out from, mouse 2.
  • An excitation light source 4 produces incident light 6 that enters a portion of mouse 2. The incident light 6 scatters in the mouse tissues and some of it eventually reaches an internal fluorescent probe 5. When excited by incident light 6, fluorescent probe 5 emits fluorescent light 7 from within mouse 2. The fluorescent photons 7 scatter and travel through tissue in the mouse to one or more surfaces 9; the light emitted from the surface may then be detected by a camera 20.
  • Thus, as light 6 and 7 diffuses through the mouse, some of the light is absorbed, but a fraction of the light propagates to a surface that faces the camera 20. For fluorescent imaging, there is a two-stage diffusion: a) incident light 6 from an incident surface to fluorescent probe 5, and b) emitted fluorescent light 7 from fluorescent probe 5 to the one or more surfaces 9. Methods described herein model the light propagation in mouse 2 to determine 3D parameters of fluorescent probe 5 and solve for the internal fluorescent probe distribution 5—given images captured by the camera and the model.
  • A difficulty in tomographic imaging mice is that the complex surface of the mouse will change with each mouse, and potentially each time the mouse is imaged (as its position and body shifts). The probe may also change each time the mouse is imaged—in position, size, strength, and spectral distribution. The difficulty, then, is determining the 3D parameters of an internal fluorescent probe distribution, such as the 3D location, size and brightness distribution of fluorescent probe 5, given that many parameters needed for tomographic imaging may change with each trial.
  • One distinguishing feature of methods described herein is that they use an actual surface topography of the mouse—as it rests under a camera at the time that light images are captured—or any other time. In this case, the methods also employ topographic determination tools. Topographic imaging determines a surface representation of an object, or a portion thereof. In one embodiment, the present invention uses structured light to determine a surface topography for at least a portion of the mouse. Tomographic imaging refers to information inside the mouse surface. An exemplary illustration of topographic vs. tomographic imaging uses a 2D planar slice through the mouse: topography gives the surface (the outer bounding line), while tomography provides information inside the bounding surface.
  • Another challenge to tomographic reconstruction that is overcome herein: the tissue in mouse 2 also autofluoresces. Tissue autofluorescence may act as a source of background or noise to tomographic imaging of a fluorescent probe distribution, and techniques described below also a) model autofluorescence and b) separate the contributions of tissue autofluorescence from light emitted from the mouse surface. This isolates light emitted from the mouse surface that corresponds to fluorescent probe 5.
  • The present invention overcomes these difficulties and permits real-time fluorescent tomographic imaging, despite variability and complexity of the mouse surface, the effects of autofluorescence, or internal fluorescent probe distribution.
  • In one embodiment, simplifying approximations to a photon diffusion model are implemented in order to expedite the computation time required to perform a reconstruction of the light corresponding to fluorescent probe 5. With the approximations described below, reconstruction times of less than 5 minutes may be achieved—compared with hours or days for methods that use FEM or Monte Carlo modeling.
  • FIG. 2 illustrates a method 200 for obtaining a 3D representation of a fluorescent light distribution located inside a mouse in accordance with one embodiment of the present invention. Processes in accordance with the present invention may include up to several additional steps not described or illustrated herein in order not to obscure the present invention.
  • Method 200 obtains one or more fluorescent images of at least a portion of the mouse (202). The images include fluorescent image data that describes fluorescent light emitted from the mouse. The images may be recalled from memory (previously captured) and/or captured in real time by a camera and imaging system, such as that described below with respect to FIGS. 14A and 14B. FIG. 3 describes one suitable method for image capture. In one embodiment, the fluorescent image data describes light that falls upon a camera or other photon detector that is distant from the mouse. In this case, the fluorescent image data is stored in the images in two-dimensions (2D).
  • Method 200 maps the 2D fluorescent image data onto a surface of the mouse (206). Before the mapping can occur, method 200 obtains a surface representation of at least a portion of the mouse (204). The surface portion may include all of the mouse, or a smaller portion. Typically, this portion includes parts of the mouse that the fluorescent image data will be mapped onto.
  • The surface representation refers to a mathematical description or approximation of the actual surface of the mouse, or a portion thereof. The surface representation need not include the entire mouse, and may include a portion of the mouse relevant to a particular imaging scenario. With a mouse for example, the surface representation might not necessarily include distal portions of the tail and distal portions of every foot. Thus, the surface representation is meant to broadly refer to any surface portion of the mouse and not necessarily the entire mouse. Typically, the surface representation includes one or more surface elements or regions of interest on the sample that produce surface light emission data related to the internal probe. For user convenience, the surface representation is often displayed in a pictorial depiction such as a 3D depiction (see FIGS. 13A and 13B).
  • Suitable techniques to obtain a surface representation include structured light, or another imaging modality such as computer tomography (CT) or magnetic resonance imaging (MRI), for example. The surface representation may be divided into a surface mesh comprising a set of surface elements, as will be described below.
  • In one embodiment, structured light is used to obtain a surface representation of the mouse. Structured light uses a set of lines of light that are projected down on the mouse at an angle (at about 30 degrees, for example) to the surface normal. The mouse generates structured light surface information as each light line reacts to the shape of the animal. Cumulatively, the lines of light each bend or alter in spacing as they pass over the mouse. The structured light surface information can be measured by a camera and used to determine the height of the surface at surface portions of the mouse that are illuminated by the structured light source. These surface portions are the portions of the mouse that face the camera (for a current position of the mouse relative to the camera). The position of the mouse relative to the camera may be changed to gain multiple structured light images and structured light information from multiple views.
  • A camera captures the structured light surface information, digitizes the information and produces one or more structured light images. A processor, operating from stored instructions, produces a 3D surface representation of the mouse—or a portion of the object facing the camera—using the structured light information. More specifically, a processing system, running on stored instructions for generating a topographic representation (a surface map) from the structured light surface information, builds a 3D topographic representation of the mouse using the structured light surface information. If multiple views are used, structured light topographies from these multiple views may be “stitched together” to provide a fuller surface representation from different angles. Structured light image capture, hardware and processing suitable for use with a mouse is described further in commonly owned and pending patent application Ser. No. 11/127,842 and entitled “Structured Light Imaging Apparatus”, which is incorporated herein by reference in its entirety.
  • Once the surface topography is determined, process flow 200 maps the fluorescent image data in the 2D fluorescent images to fluorescent image data at a surface of the mouse (206). This converts 2D light data collected at a camera to 3D light data at a 3D surface of the mouse. In one embodiment, the mapping converts radiance data from the fluorescent images to photon density just inside the surface.
  • The mapping manipulates 2D camera data according to the geometry between the mouse surface and the camera lens to derive values of the light emission intensity (or radiance) at the surface. A variety of techniques can be used to map camera light data to the mouse surface. In one embodiment, the mapping uses a simple 3D translation based on the relative position between the camera and mouse surface. When the mouse rests on a stage in an imaging box, four spatial relationships between a camera and a stage that supports the mouse include: the camera and stage both do not move, the camera moves relative to the stage, the stage moves relative to the camera, and the camera and stage both move relative to each other. In any of these cases, the position of the camera relative to the stage/mouse is known. This permits a 3D translation using the known coordinates from the camera to the mouse surface.
  • More sophisticated spatial relationships between a camera and mouse may be used. In another embodiment, the angle of the mouse surface is also accounted for in the mapping. FIG. 5 illustrates a relationship for converting 2D camera data to surface data for a sample surface element 402. FIG. 5 shows a relationship between surface element 402 (on the mouse), image 404, and an entrance pupil or camera 406 of an imaging system. Light emitted from surface element 402 passes through entrance pupil 406 and is recorded in image 404. The angle of emission with respect to the surface normal is θ, which is known from the surface topography determined in 204. The entrance pupil 406 subtends a small solid angle dΩ. The imaging system may collect light emitted from surface element 402 on the sample at an angle θ (measured with respect to the normal to surface element 402) into the solid angle dΩ subtended by the entrance pupil. This information may then be used to convert image data obtained by the camera into the surface emission intensity corresponding to the surface geometry.
  • Emission of light from a mouse surface may be specified in units of radiance, such as photons/sec/cm2/steradian. In one embodiment, an imaging system captures images of the mouse and reports surface intensity in units of radiance. Surface radiance can be converted to photon density just inside the mouse surface, using a model for photon propagation at the tissue-air interface, as described below. When the surface representation includes a set of surface elements, the mapping may produce a surface emission data vector that includes photon density at each surface element for the mouse topography. The photon density just inside the surface are then related to a light emitting probe distribution inside the mouse tissue using a diffusion model.
  • Returning back to FIG. 2, method 200 then models the contributions of tissue autofluorescence to the light emitted from the mouse (208). Autofluorescence refers to the natural fluorescence of substances within a material or organism. Mammalian tissue has autofluorescence properties that will affect fluorescent imaging. A camera receives image data that includes both: a) light escaping from the mouse surface due to autofluorescence of tissue in the mouse, and b) light escaping from the mouse surface due to fluorescent probe 5. From a camera's perspective, these two contributions are often mixed.
  • Multiple techniques are contemplated for determining autofluorescence and separating it from the surface emission for fluorescent probe 5. In one embodiment, autofluorescence is determined by measurements made in control animals (animals without a fluorescent probe). In this case, an average autofluorescence yield per unit volume of tissue can be derived from images of autofluorescence. The autofluorescence yield can then be used in a forward model of light propagation, e.g., see Eq. 7 below and its associated description.
  • Other autofluorescence determination techniques may be used. For example, when performing transillumination fluorescence measurements, the autofluorescence pattern on a surface of the animal typically matches a transillumination pattern observed when imaging with an emission filter and no excitation filter. This provides another way to determine and correct for autofluorescence. The process may include: 1) acquiring a normal fluorescence image with specified excitation filter and emission filter (image 250); 2) acquiring another image with excitation filter removed (image 252); multiplying image 252 by an appropriate scale factor and subtracting it from image A, thus reducing the contribution of autofluorescence from image A.
  • Another method to reduce autofluorescence is to use spectral unmixing. In this case, images are acquired with an array of different emission or excitation filters. Spectral unmixing software separates autofluorescence from the fluorescent probe emission, using known spectral characteristics of the images.
  • After autofluorescence has been determined and separated from the surface emission data, the remaining fluorescent probe emission contributions to the surface emission data can be used for tomographic processing (without the noise and contributions of tissue autofluorescence). Method 200 subtracts the modeled tissue autofluorescence from the light emitted from the surface (as calculated in 206), which isolates the light/signal due to the fluorescent probe 5. This is shown pictorially in FIG. 4 with two epi-illumination images: image 250 shows the autofluorescence of a mouse; image 252 shows the fluorescence of the internal probe in the mouse.
  • Method 200 then calculates a 3D representation of the fluorescent probe distribution internal to the mouse (210). As the term is used herein, a fluorescent probe distribution refers to a description or mathematical depiction of fluorescent light emitters inside the mouse. Typically, the fluorescent light corresponds to a fluorescent probe disposed inside the animal. As mentioned above, the fluorescent probe may include a fluorescent marker such as a dye molecule, or a fluorescent reporter that produces fluorescent light based on gene expression.
  • Light data internal to the mouse 2 surface generally refers to mathematical representation or approximation of light within the mouse 2 interior. This may include a set of points or volume elements, each characterized by 3D position and a source strength. In one embodiment, the present invention divides the mouse 2 interior into volume elements where each volume element is considered to contain a point light source at its center. A solid mesh of these volume elements then defines a collection of point sources used to approximate light data internal to the mouse and the actual probe distribution within mouse 2. For example, a solid mesh of cubic volume elements may be used.
  • In one embodiment, fluorescent probe 5 includes emits low-intensity light. In one embodiment, a low intensity fluorescent probe of the present invention emits light within mouse in the range of about 104 to about 1014 photons/second, depending on probe concentration and excitation light intensity. For some imaging systems, a fluorescent probe 5 that emits flux in the range of about 104 to about 1010 photons/second is suitable. Other light fluxes are permissible with the present invention. Photons/second is one unit of measure suitable to quantify the amount of light produced by probe 5. Other units of measure are known to one of skill in the art, such as Watts. For reference, the conversion of photons/second to Watts is 3.3 nanowatts equals about 1010 photons/second at 600 nm. In one embodiment, probe 5 emits light between about 10−15 to 10−6 watts of light. The amount of light produced by fluorescent probe 5 refers to the light emitted within mouse 2—not necessarily the amount of light generated by excitation light source 4 (such as an LED) that generates the light incident on the fluorescent probe 5.
  • Method 200 uses the fluorescent light emission data from the mouse surface, along with tomographic imaging software that models light propagation internal to the mouse and solves for fluorescent probe distribution. The internal light propagation modeling includes both a) fluorescent excitation light propagation from the excitation light source 4, and its entry points into the mouse, to the fluorescent probe 5, and b) fluorescent emission light propagation from the fluorescent probe 5 to the surfaces captured in the fluorescent images.
  • Tomographic modeling, processing, and fluorescent probe determination of step 210 is described in further detail below with respect to FIGS. 9 and 10. For user convenience, the resultant 3D representation produced by method 200 may be expressed as a pictorial depiction, e.g., on a computer monitor. FIGS. 13A and 13B show depictions of a mouse with a 3D surface topography and internal fluorescent probe 5 determined using method 200.
  • Tomographic imaging in method 200 finds use in a wide array of imaging and research applications such as oncology, infectious disease research, gene expression research, and toxicology, for example. The tomographic imaging is suitable for use with samples having a complex surface, such as a mouse. As the term is used herein, a complex surface is any surface that cannot be described solely using a single polygonal description. The reconstruction techniques described herein place no restrictions on the source distribution, such as the number of probes in the sample or the sizes and shapes of the sources, and no restrictions on the geometry, size or shape of the surface.
  • In some embodiments, method 200 may occur in real time where image capture (202), topographic acquisition (204) and the data calculations (204-210) all occur without significant delays to a user. In other words, soon after all the images are obtained—e.g., the images are captured or previously captured images are selected and recalled from memory—and the user inputs desired parameters for the tomographic assessment, method 200 outputs 3D details for the internal fluorescent probe 5. In one embodiment, mapping the fluorescent image data and determining the 3D fluorescent probe distribution (steps 206-210) finishes in less than about 5 minutes. In another embodiment, details of a fluorescent probe distribution are determined in less than about 1 minute. A video display may then show a pictorial representation of the tomographic reconstruction output on a monitor to a user. This quick processing allows a user to repeat process flow 200—or change parameters in the tomographic assessment relatively easily. This increases researcher productivity. If the mouse is in an imaging box and under anesthesia, this also permits multiple imaging sessions, in efficient succession, without a need to handle the mouse between tomographic imaging sessions. This real time imaging permits the mouse to be anesthetized for shorter durations (despite permitting multiple imaging sessions).
  • FIG. 3 shows a process flow for fluorescent image capture 202 (202 in FIG. 2) according to a specific embodiment of the present invention. In one embodiment, image capture 202 occurs with the mouse resting or lying on a horizontal stage or flat surface in a resting or normal position, without the need for any straps holding the mouse. The mouse may be anesthetized to prevent movement during imaging 202. The stage may then move between image captures, but the mouse typically remains stationary on the stage during image capture with the camera. In addition to fluorescent light image capture, process flow 202 also captures photographic and structured light images. Other embodiments of fluorescent light image capture 202 need not include photographic and structured light image capture.
  • Image capture 202 uses box 10 of FIGS. 14A and 14B. In this case, image capture 202 begins be receiving a mouse in an imaging box (220). This often occurs when a user places the mouse on a stage within an imaging chamber for the imaging box. The user may also initiate image capture 202 using a computer associated with the imaging system.
  • In one embodiment, the stage is movable. In this case, the imaging system moves the stage to a desired position according to a control signal provided by a computer in the imaging system (222). For example, a user may input a desired image position via the computer user interface, and the imaging control system moves the stage accordingly. Alternatively, a desired position for the stage may be pre-programmed based on an automated data collection routine that the user initiates. A moving stage allows multiple positions and angles relative to a fixed camera.
  • The camera then captures a structured light image (224). Structured light image capture may be accomplished using a structured light projection system. In a specific embodiment, the structured light projection system projects structured light down onto the mouse from an angle, and the camera (also above the mouse, or on the same side of the mouse as the projector) captures the altered structured light. Suitable structured light generation systems are described in commonly owned and co-pending patent application Ser. No. 11/127,842. The structured light image data is also transferred to an image processing unit and/or a processor in the imaging system for storage for further processing to build a 3D surface representation.
  • A camera then captures a photographic image (226). The photographic image data is transferred to an image processing unit and/or a processor in the imaging system for storage. The photographic image may be subsequently used for display. For example, the photographic image may be used in an overlay image that includes both the photographic image and fluorescent probe distribution (output from 210). The overlay provides a simple pictorial view to facilitate user visualization of the internal fluorescent probe distribution.
  • The camera then captures a fluorescent light image (228). Fluorescence imaging illuminates the mouse to excite fluorescence molecules in the internal fluorescent probe, and then captures an image of the mouse, or a portion thereof, as the internal probe fluoresces. Fluorescent image capture provides incident light onto into the mouse with an illumination source. The incident light should be large enough in magnitude to elicit a fluorescent from the probe, but not too large so as to saturate a CCD camera. In response to the incident light, light emits from the “excited” fluorescent probe.
  • Trans-illumination and/or epi-illumination may be used. FIG. 6A schematically shows trans-illumination in accordance with one embodiment. Trans-illumination provides light from a side of the mouse opposite to the camera (e.g., incident light from below and a camera above), so that the light travels through the mouse. This provides lower levels of autofluorescence, which is useful for 3D tomographic reconstructions. Also, the ability to move the transillumination point relative to a fluorescent probe fixed within the animal, provide additional information that is use for 3D tomographic reconstructions. In this case, the excitation light source 4 includes a lamp 90 that provides light that passes through a filter in excitation filter wheel 92, which allows a user to change the spectrum of the incident excitation light. A fiber bundle switch 94 directs the excitation light into one of two paths 95 and 97. Path 95 is used for trans-illumination and directs the incident light along a fiber bundle or cable for provision towards a bottom surface of the mouse 2. In one embodiment, the outlet position of path 95 can be moved or re-directed to create multiple incident excitation light locations of trans-illumination path 95.
  • Epi-illumination provides the incident light from the same side of the animal that an image is captured (e.g., incident light from above, and a camera above the mouse), and is often referred to as reflection-based fluorescent imaging. FIG. 6B schematically shows epi-illumination in accordance with one embodiment. In this case, switch 94 directs the excitation light into path 97, where it routs to a position above the mouse for provision towards a top surface of the mouse 2 on the same side of the mouse as camera 20.
  • Epi-illumination provides a faster survey of the entire animal, but may be subject to higher levels of autofluorescence. Both trans-illumination and epi-illumination may be used. Epi-illumination avoids significant light attenuation through the mouse, and may help constrain volume elements near the camera-facing surface of the mouse. For example, the epi-illumination constraints may identify artifact voxels near the top surface, which are then removed by software.
  • In either case, an emission filter 98 allows a user to control a spectrum of light received by camera 20. This combination of excitation filter wheel 92 and emission filter 98 allows images to be captured with numerous combinations of excitation and emission wavelengths. In a specific embodiment, excitation filter wheel 92 includes twelve filters while emission filter 98 includes 24 positions.
  • Imaging may also capture both trans- and epi-illumination images, and combine the data. In each view, the light takes a different path through mouse, which provides a different set of input criteria and internal light conditions for tomographic reconstruction calculations.
  • A structured light source 99 also provides structured light onto the top of the animal for structured light image capture by the camera 20 without moving the mouse 2 on the horizontal surface.
  • In another embodiment, the stage is moveable, which allows camera 20 to capture images from multiple perspectives relative to the mouse 2. The stage may move in one dimension (e.g., up and down or side to side) or two dimensions for example.
  • In one embodiment, the fluorescent excitation uses a different spectrum than the fluorescent emission. As one of skill in the art will appreciate, the bandgap between excitation and emission filters will vary with the imaging system used to capture the images. A bandgap of at least 25 nm is suitable for many imaging systems. The excitation spectrum may be achieved using any combination of lights and/or filters. The emission spectrum will depend on a number of factors such as the fluorophore used, tissue properties, whether an emission filter is used before the camera, etc. In one embodiment, the transillumination location of the excitation light source is moved to capture multiple images of internal fluorescence and the same set of excitation and emission filters is used for the different excitation light source positions.
  • A camera then captures a fluorescent light image of at least a portion of the mouse (228). The fluorescent image records fluorescence as a function of 2D position. The image may include the entire mouse, or a portion of interest that has been zoomed in on (optically or digitally). The image is transferred to the image processing unit and/or computer for subsequent processing.
  • Multiple fluorescent light images may be captured with the mouse in its current position (230). In one embodiment, this is done to facilitate spectral unmixing, where each image capture (228) uses a different excitation and/or emission spectrum. In another embodiment, multiple images are taken for differing trans-illumination positions of the excitation light source 4 (FIG. 1). Each trans-illumination position provides a different set of input conditions to the tomographic reconstruction. FIG. 7 illustrates 21 sample images 235, each taken with a different trans-illumination position of the excitation light source. In this case, the imaging system is configured to move the excitation light source (or has multiple excitation light sources that are controllably turned on/off) and captures an image of the mouse for each different trans-illumination position of the excitation light source.
  • All of the images 235 may be used in a tomographic reconstruction, or a subset can be used. The subset may be selected based on a quality measure for the images, such as a threshold for number of fluorescent photons collected in each image. Other quality measures may be used to select the images. The number of images captured may vary. In one embodiment, 1 to about 80 different trans-illumination positions and images are suitable for tomographic reconstruction. In a specific embodiment, from about 4 to about 50 images are suitable. The images may be stored for tomographic assessment at a later time, e.g., the images—or a subset thereof—are recalled from memory during tomographic processing.
  • In one embodiment, the stage and mouse may then be moved to a second position (232). While the stage is at the second position, one or more photographic, structured light, and/or fluorescent images of the mouse may be captured (224-230). Image collection may further continue by capturing images of the sample from additional positions and views. For example, image capture may occur at anywhere from 2 to 200 positions of the mouse within an imaging chamber. In general, as more images are captured, more information is gathered for tomographic reconstruction. Also, multiple structured light positions may be used to images more of the mouse in 3D. Eight positions, spaced every 45 degrees about a nose-to-tail axis of the mouse, is suitable in some 3D embodiments to build a stitched together surface representation for 360 degree viewing about the mouse.
  • In one embodiment, image capture 202 is automated. A user may initiate software included with an imaging system that controls components of the imaging system responsible for image capture. For example, the user may launch imaging and acquisition software on a computer associated with the imaging system that initializes the camera and carries out imaging automatically. According to stored instructions, the software may then select a desired stage position if a moveable stage is used, prepare the system for photographic, structured light, and/or fluorescent image capture (e.g., turn on/off lights in the box), focus a lens, selectively position an appropriate excitation or emission filter, select an excitation fluorescent light source (one of many for example), set an f-stop, transfer and store the image data, build a reconstruction, etc. For fluorescent image capture, software activates the camera to detect photons emitted from the mouse, which usually corresponds to absolute units from the surface. The camera may capture the fluorescent image quickly or over an extended period of time (up to several minutes).
  • Additional processing may occur on the fluorescent images. Fluorescent imaging often captures image data with multiple reporters; each reporter may have its own wavelength spectrum. A camera image of a mouse with multiple reporters has the spectral results of each reporter mixed together. In this case, spectral unmixing is useful to clean fluorescent image data and separate the contributions from each source before tomographic processing. The unmixing may also identify contributions from autofluorescence. In one embodiment, a spectral unmixing tool is employed in software to separate fluorescent contributions from multiple sources. This permits fluorescent tomography described herein to image multiple reporters in a mouse independently. For example, one reporter may be used in an imaging application to monitor cell death in the mouse, while the second reporter monitors cell propagation. A user may initiate the spectral unmixing tool and software with an appropriate user interface command.
  • FIG. 9 shows a process flow 300 for obtaining a 3D representation of a fluorescent probe distribution located inside mouse 2 in accordance with a specific embodiment of the present invention. Process flow 300 expands upon method 200 of FIG. 2, and converts surface light emission data to a mathematical representation of a fluorescent probe distribution within the mouse.
  • Process flow 300 first divides a surface representation for the mouse into a surface mesh that includes a set of surface elements (302). This may include obtaining a surface topography, if that has not already been done (see 204 in FIG. 2). The number of surface elements will vary according to the mouse surface area and a desired solution accuracy for the tomographic reconstruction. The number of surface elements in the set should be large enough to capture photon density details and variation across the mouse surface. For example, between about 100 and about 10,000 surface elements may be suitable for a mouse.
  • Process flow 300 then selects a number of images for use in the tomographic assessment (304). As mentioned above in image capture, not all images previously captured and stored in memory need be used. For example, a user may select images that include a moving trans-illumination light source that is closer to a fluorescent probe compared to other images where the moving trans-illumination light source is farther away from the probe. Epi-illumination images may also be incorporated into process flow 300.
  • Process flow 300 maps photon data from the images to the surface topography mesh (306). This may use the mapping techniques described above in 206 of FIG. 2.
  • Expanding upon the mapping described above with respect to 206 in FIG. 2, the mapping converts surface light data (excitation and/or emission) into light data internal to a surface. Notably, this relates surface emission intensity to photon density just inside the mouse surface. In one embodiment, process flow 300 converts values of light emission intensity for each surface element into photon density just inside the surface. Referring briefly to FIG. 5, the value of emission intensity at a surface element, I(θ2), is related to the photon density ρ beneath the surface element. The exact form of the relationship depends on the model used to describe the transport of photons across the surface boundary. One embodiment of this relationship, based on the partial-current boundary condition, is given by: I ( θ 2 ) = c 4 π n 2 T ( θ ) cos θ 2 Ω 1 + 3 2 1 - R eff 1 + R eff cos θ ρ ( 1 )
  • Here, c is the speed of light, n is the index of refraction of the sample medium, T is the transmission coefficient for light exiting the sample through the surface element, and θ is the internal emission angle, which is related to the external emission angle θ2 through Snell's law:
    n sin θ=sin θ2  (2)
  • The parameter Reff is the average internal reflection coefficient calculated from the following formulae: R eff = R ϕ + R j 2 - R ϕ + R j R ϕ = 0 π 2 2 sin θ cos θ R ( θ ) θ R j = 0 π 2 3 sin θ cos 2 θ R ( θ ) θ R ( θ ) = { 1 2 ( n cos θ 2 - cos θ n cos θ 2 + cos θ ) 2 + 1 2 ( n cos θ - cos θ 2 n cos θ + cos θ 2 ) 2 for θ < arcsin ( 1 / n ) 1 for θ > arcsin ( 1 / n ) ( 3 )
  • Thus, the internal reflectivity Reff depends on the index of refraction of the medium underneath a surface element. In tissue for example, Reff is typically in the range of 0.3-0.5.
  • Eqs. (1) and (2) may thus be used to convert surface emission data measured at each surface element to values of the photon density beneath the surface.
  • Autofluorescence is then modeled and subtracted from the surface emission data (308). Suitable techniques for doing so were described above with respect to 208 in FIG. 2.
  • Referring back to FIG. 9, process flow 300 then divides the mouse interior volume into volume elements, or ‘voxels’ (310). FIG. 8 shows sample side/lateral and top/dorsal view of a mouse 295 representation that includes volume elements 297. In one embodiment, each volume element 297 is considered to contain a point light source at its center. A solid mesh of volume elements 297 then defines a collection of point sources used to approximate light in the mouse. Volume elements 297 may also be used as a framework to describe the fluorescent probe distribution 5. In a specific embodiment, process flow 208 may use a volume element 297 resolution of about 0.5 to about 6 millimeters for a small mammal. A volume element 297 resolution of about 1 millimeter is suitable for some mice. Other volume element 297 sizes and densities may be used.
  • One or more early constraints may also be applied (312) to expedite or simplify the determination, such as applying one or more limits on the modeling and solution-space. In one embodiment, the internal light modeling solution space is spatially limited to within the boundaries of the mouse surface. In another embodiment, a volume space used within the reconstruction is limited by one or more practical considerations. For example, regions of the internal mouse volume far away from where fluorescent light emission takes place (e.g., the rear of the mouse when the head visibly shows the highest light emission density), as determined by a visual scan of the images, may be excluded from the solution space.
  • Process flow 300 then models light propagation. In the embodiment shown, this occurs in a three-step process where excitation light and emission light are each modeled separately and then the two are combined (314, 316, and 318).
  • Light transport in turbid media such as tissue is dominated by scattering and is essentially diffusive in nature. In one embodiment, tissue scattering and absorption parameters are known a priori, stored in memory, and recalled from memory when a reconstruction occurs. In another embodiment, tissue scattering and absorption parameters are calculated from trans-illumination measurements.
  • In many instances, the condition for diffusive transport is that the scattering coefficient μs be greater than the absorption coefficient μa so that the change in the photon density is small between scattering events. The photon density produced by a source power density, Ui, in a homogeneous medium may be represented by the diffusion equation:
    D∇ 2ρ−μa cρ=−U i( x )  (4)
  • where the diffusion coefficient D is, D = c 3 ( μ A + μ S ) ( 5 )
  • An emission Green's function is a solution to Eq. (9) subject to the boundary condition imposed by the surface of the sample.
  • In a specific embodiment, a Green's functions is used to model internal light propagation. A Green's function mathematically describes light propagation through space, such as through tissue, from one location to another. In one embodiment, the Green's function uses volume elements 297 and surface mesh elements as vector spaces for its data elements. In a specific embodiment, an excitation Green's matrix models light propagation from a position of the excitation illumination source to the volume elements 297 (314). An emission Green's matrix may also be used to model light propagation from the volume elements 297 to the surface elements (316).
  • The excitation and emission models are then combined (318). In a specific embodiment, the excitation and emission Green's function matrices are coupled together, along with a coupling constant, and form a single fluorescence Green's kernel matrix for the fluorescence forward model. In another specific embodiment, the excitation Green's function and emission Green's function matrices are composed using a hybrid Green's function expression which combines weighted terms of a radial partial current boundary condition and an extrapolated boundary condition. This coupled Green's function may be applied to fluorescence of the probe and/or autofluoresence.
  • Other modeling processing and factors are suitable for use. Modeling may also include one or more of: a) establishing a relationship between the surface elements and volume elements, b) setting additional limits on the modeling and solution-space, c) deciding whether to use a homogeneous or non-homogeneous model for light propagation in tissue, and/or d) composing a mathematical representation of light internal to the mouse. FIG. 10 describes modeling suitable for use with process flow 300 in more detail.
  • Referring back to FIG. 9, process flow 300 then determines the light data internal to the mouse, including the desired fluorescent probe distribution that includes the fluorescent probe (320). For example, once the Green's function is determined, the distribution is obtained by solving the system of linear equations that relate the photon density at the surface to the source distribution inside the object. In one embodiment, process flow 300 solves for all the internal volume elements. Thus, once the Green's function is modeled and determined, it may be evaluated for every volume element-surface element pair, in order to obtain the system of linear equations (Eq. 7, below). Referring forward to Eq. (7), since ρ is known, and Gij can be determined as described below, the reconstruction method then solves the linear system, Eq. (7), for the source strengths Si.
  • Typically, there is no exact solution to the linear system because the collection of point sources is only an approximation of the actual source distribution. One suitable reconstruction is then the best approximate solution of the linear system. In a specific embodiment, process flow 300 uses the non-negative least squares algorithm to solve for the internal fluorescent probe distribution. Other techniques may be used. In some cases where the fluorescent probe distribution includes a spatially smoother solution, Eq. (7) can be augmented using a regularizing matrix in the first derivative.
  • In one embodiment, the present invention relies on a simplified analytical approximation (planar boundary condition) for the Green's function as described above. In another embodiment, a look-up table can be used for the Green's function. The look-up table may be created by previous measurements of photon transport in a sample (or similar sample approximated to be substantially equal to the current sample), or by computational simulations using techniques such as Monte Carlo or finite element modeling. This particular method is useful for samples consisting of inhomogeneous media, such as animal subjects. In this case, the optical properties of the tissue, μa and μs may have spatial dependence or other heterogeneous properties.
  • FIG. 13A shows top and side views 380 and 382 of sample reconstructed results for a fluorescent probe distribution within a phantom mouse (e.g., plastic mouse having an embedded fluorescent probe). In this case, the reconstructed source 383 shows a fluorescent dye in the phantom mouse whose fluorescent yield values are above 10% of the maximum light value in the reconstructed solution. FIG. 13B shows reconstruction results for a fluorophore in a real mouse.
  • In one embodiment, process flow 300 applies an iterative solution process. Iterative processing obtains multiple three-dimensional representations and compares them to improve the final output and assessment for the fluorescent probe distribution. In this case, process flow 300 varies the tomographic assessment or modeling, finds a potentially new of different solution in each iteration, and then selects one of the multiple solutions. Loop 328, for example, varies the subset of images that were selected from a larger set of images.
  • To facilitate comparison between iterations, iterative process flow 210 assesses the solution quality and assigns a quality to each iterative solution (322). In one embodiment, the assessment measures a difference between the observed photon density and the calculated photon density. For example, a “chi squared” criteria may be used: x 2 = i [ ρ i - j G ij s j ρ i ] 2 ( 6 )
  • The value of χ2 measures the difference between the observed photon density ρi and the calculated photon density j G ij S j
    over the surface of the sample. Other terms shows in Equation 6 are described further below with respect to Equations 7-9.
  • In one embodiment, iterative process flow varies volume element configuration. Loop 330 varies the number and/or size of volume elements. In this case, volume element size is initially set, and changed as iteration proceeds. In some cases, the initial voxelation is relatively coarse and refined with successive iterations. For example, the volume element size may be reduced by a factor of two in a next iteration. If the solution quality improves after this second pass, then the volume element size may be again reduced by a factor of two in a third iteration. If the solution quality doesn't improve or gets worse, then the algorithm may have converged on a final solution and stop. In one embodiment, the initial volume element size may range from about 0.1 mm3 to about 1 cm3, and subsequent and/or final volume element size for volume elements close to the source may reduce from about 1 mm3 to about 10 mm3. In a specific example, the initial volume element size may be about 200 mm3 or about 1 cm3, and the final volume element size for volume elements close to the source may reduce to about 1 mm3.
  • In some cases, it is advantageous to reduce the number of volume elements in the problem while maintaining a high density of volume elements in the vicinity of the fluorescent probe. This can be achieved by using adaptive meshing. In one embodiment, adaptive meshing increases the density of the solid mesh near the probe to provide increased volumetric information in this space, while density of the solid mesh decreases in areas where no activity of interest is taking place (no light generation or transport). In one suitable adaptive meshing application, a coarse volume element mesh is initially applied throughout the entire sample volume and the current solution is found, yielding an initial solution for Sj. Next the volume elements that have source strengths greater than zero (Sj>0) are refined (i.e. subdivided) and those where the source strengths equal zero (Sj=0) are removed. Solution attainment and volume element mesh refinement may then be iterated repeatedly, producing a high-density volume element mesh localized around the fluorescent probe distribution. During each iteration, the quality of the current solution is assessed (322). In a specific embodiment, the iteration continues until further refinement produces no significant decrease in the assessment value.
  • An additional iterative improvement may be obtained by varying the number of surface elements, Ns, used in obtaining the three-dimensional representation (loop 326). Using a subset of the surface elements of the surface mesh reduces the number of constraints in the problem, which may simplify and expedite solution calculation. The number of surface elements may be used to sample the surface uniformly. In this case, process flow 300 iterates for different values of Ns corresponding to sampling the surface element mesh at different densities, and use the quality assessment (322) to determine the best solution among the different values of NS. For example, if the number of surface elements is between about 100 and about 300 surface elements for a small mouse, an iteration step size between about 10 and 50 may be suitable.
  • FIG. 10 expands modeling light propagation in a mouse in accordance with a specific embodiment of the present invention. Modeling 340 establishes a relationship between the surface elements and volume elements.
  • In one embodiment, the reconstruction uses a linear relationship between the source emission strength and the photon density at the surface. In a specific embodiment, the linear relationship is described by a Green's function. As mentioned before, the Green's function mathematically and numerically describes the transport of photons inside the mouse, and may accommodate for the effects of non-homogeneities in the volume and internal reflection at the boundary. In this case, the Green's function also describes the transport of photons inside the sample from each point or volume element in the distribution to the inside of each surface element. One useful form for the Green's function is a simplified approximation in which the surface of the sample is treated locally as a planar interface oriented tangent to a surface element, as shown in FIG. 12. The photon density at the surface is the analytical solution for a point source in a semi-infinite slab using the partial-current boundary condition. This allows the Green's function to be calculated with minimal computational expense. Other boundary conditions could be used.
  • Modeling 340 may incorporate additional information and constraints into the solution-space (342). Additional limits on a Green's function solution-space may apply the location of the input excitation signal at the surface. For example, if a trans-illumination fluorescent image is used, then the known (and potentially changing between images, see 230 in FIG. 3) bottom illumination source position may be included in the Green's function and modeling.
  • The model also selects or assigns homogeneous or non-homogeneous properties to the mammalian tissue (344). A mouse includes a turbid interior. A turbid interior refers to a volume that does not allow unimpeded transport of light. The turbid interior may comprise one or more mediums, structures, solids, liquids, gases, etc. In one embodiment, the sample is modeled as homogeneous such that each representative volume element in the sample is characterized by identical light transport properties. In another embodiment, the sample is represented as heterogeneous such that various representative volume elements in the sample are characterized by different light transport properties. In a mouse, the interior may comprise a mixture of tissues, bones, organs, etc., each of which may be characterized by separate light transport properties in a heterogeneous model. It is understood that animals are not fully homogeneous and that tissue absorption for living mammals varies with the type of tissue or tissue cells, and is generally affected by varying particles and quantities such as the presence of hemoglobin. However, software run by an imaging system may implement homogeneous or heterogeneous assumptions on the optical behavior of mammalian tissue when imaging a living mouse. Generally, Green's functions for homogeneous tissue models can be calculated analytically in real time for each imaging example, while more complex heterogeneous models require significant computational effort and may be saved in a look-up table.
  • Data for the fluorophore is then obtained 346, such as data related to its emission spectrum. In some embodiments, an emission spectrum for the fluorophore at one or more wavelengths is provided as an input to the model. The fluorophore(s) used in the fluorescent probe are typically known for an imaging application, and optical properties for the fluorophore wavelengths are also known and may be stored in software prior to imaging. In some cases, a user selects a wavelength filter, with its predetermined wavelength range, for image capture and the spectrum properties for the fluorophore at that wavelength range are input to the model. Alternately, the imaging process is automated and a computer recalls spectrum properties for the fluorophore from memory based on an automated wavelength filter selection. A graphical user interface associated with the imaging system may also allow a user to select one or more fluorophores from a list, where information for each fluorophore is stored in a database. Other fluorophore properties may include excitation spectrum and extinction coefficient and quantum efficiency, for example.
  • With a linear relationship between the source strength in each volume element and the photon density at each surface element described by a Green's function Gij, the photon density at the ith surface element may be approximated by the sum of the contributions from all the volume elements: ρ i j G ij s j ( 7 )
  • where ρi represents photon density at the surface for the ith surface element. Generally, ρi is known from the camera image data after it is mapped onto the surface, while Gij is known from the modeling, leaving Sj to be solved for. In other words, Sj refers to the amount of light in each volume element. For fluorescent tomographic reconstruction, Sj includes two components: one from the fluorescent probe and a second from the autofluorescence, which may be represented as:
    s j =[s fluor +s autofluor]j  (8)
  • For a fluorescent source, the relationship between the surface elements and the volume elements may accommodate both excitation and emission modeling. The Green's function in the linear system thus includes a) a first Green's function that describes the transport of the emission light from the volume elements to the sample surface and b) a second Green's function that describes the transport of the excitation light from the sample surface to the volume elements. In a specific approximation, the Green's function in the linear system (7) includes the product of two Green's functions: (need to fix eq 9) G ij = G ij em k G jk ex s k ext ( 9 )
  • The first Green's function, Gi E, describes the transport of excitation light from the excitation source at the surface of the sample to the ith volume element. The second Green's function, Gij F, describes the transport of the fluorescent light from the ith volume element to the jth surface element. Both Green's functions can be determined from analytical expressions, such as the simplified approximation described above in the case of a homogeneous medium, or from look-up tables in the case of an inhomogeneous medium. The excitation and fluorescent light are typically at different wavelengths, and thus the fluorescence does not stimulate additional fluorescence.
  • Combining equations 7, 8, and 9, gives: ρ i j [ G ij s j fluor + G ij s j autofluor ] ( 10 )
  • If autofluorescence is modeled according to one of the methods described earlier with respect to FIG. 2, then the autofluorescence term in equation 10 can be subtracted from the measured photon density (ρj), resulting in the following equation to be solved for the fluorophore concentration Sj fluor: ρ i - j G ij s j autofluor j G ij s j fluor ( 11 )
  • As described previously, this equation can be solved by a non-negative least squares optimization method. Other methods for solving systems of linear equations can also be used.
  • In one embodiment, the reconstruction uses a tangential plane boundary approximation combined with a partial current boundary condition to model photon diffusion Green's function for each surface element. FIG. 11 illustrates a schematic diagram showing this emission planar approximation. A plane boundary 352 is drawn tangent to the ith surface element. The photon density in the planar approximation is the solution for the point source at xj in a semi-infinite slab defined by the plane boundary, subject to the partial current boundary condition. Specifically, the boundary condition is simplified to the case of a plane boundary, although the orientation of the boundary may change for each surface element.
  • This simplified emission Green's function is the analytical solution for a point source in the semi-infinite slab using the partial-current boundary condition: G ij = 1 2 π D { exp ( - μ eff r ij ) r ij - 1 z b exp ( r ij / z b ) E 1 [ ( μ eff + 1 z b ) r ij ] } ( 12 )
  • Here rij=|xj−xi|, E1 is the first order exponential integral and
    μeff=[3μAA+μ′S)]  (13) z b = 2 D c 1 + R eff 1 - R eff ( 14 )
  • In the simplified model just described, the simplified Green's function depends only on the distance between the volume element and the surface, and the angle of the surface element. This method of calculating the Green's function is fast and can be performed in real time for each mouse surface. It is understood that it is not necessary to use this simplified approximation to define the Green's function.
  • A similar approximation solution for the excitation Green's function may be constructed. FIG. 12 illustrates a schematic diagram showing one suitable excitation approximation, which models light as it enters the mouse. In this case, four fictitious point sources, S0-S3, are used to model light onto a volume element. Sources S0 and S1 are considered with a partial-current boundary condition, while sources S2 and S3 are considered with an extrapolated boundary model. Again, the boundary condition for excitation is simplified to the case of a plane boundary, whose orientation may change for each surface element.
  • The planar boundary approximations discussed above work well for smooth surfaces with a large radius of curvature, and for cases where the absorption coefficient is not too small (μa>0.1 cm−1). An advantage of the planar approximation technique described above is that it is computationally convenient for solving the diffusion equation with an arbitrary complex boundary such as a mouse. Areas with more structure, such as the head or the limbs of a mouse, may benefit from a more accurate model of the boundary. Using a finite element modeling code to calculate the Green's functions is one option to obtain a more accurate boundary model. Finite element codes such as Flex PDE, from PDE Solutions, Inc. may be used for example. Another option will be to extend the planar surface approximation to first order in curvature, which may allow continued use of analytic expressions for Gij.
  • Although process flow 300 has been described with many simplifications to the model to expedite processing, fluorescent tomographic reconstruction is not limited by these simplified computational methods. For example, the Green's Function may be calculated without many of the simplifications described above, even at the cost of increased computational requirements. In addition, while process flow 300 describes a specific method of obtaining measurements of light emission from the mouse, process flow 300 is not limited to how the light emission data is obtained or to the use of any particular apparatus. For example, light emission data may be obtained from an independent source and stored as data within a computer, and not necessarily produced as the result of imaging via a complementary or local imaging system.
  • In addition, although the present invention has been described so far with respect to a fluorescent probe that emits light, process flow 300 may be used to obtain 3D reconstructions of any type of internal light source, including one or more bioluminescent sources.
  • The tomographic reconstruction techniques of the present invention are typically implemented by a suitable processor or computer-based apparatus. FIGS. 14A and 14B illustrate an imaging system 10 configured to capture photographic, fluorescent and structured light images of a mouse in accordance with one embodiment of the present invention. While tomographic reconstruction will now be described with respect to imaging system 10, it is understood that the tomographic reconstruction as described herein is well suited for use with other imaging systems.
  • Imaging system 10 may be used for imaging a low intensity fluorescent probe such as fluorescent molecules in a mouse and the like. The low intensity fluorescent probe may be included in any of a variety of living or non-living light-emitting samples. Non-living light-emitting samples may include calibration devices and phantom devices. Living light-emitting samples may include, for example, animals or plants containing light-emitting molecules, tissue culture plates containing living organisms, and multi-well plates (including 96, 384 and 864 well plates) containing living organisms. Animals may include any mammal, such as a mouse or rat containing luciferase-expressing cells.
  • System 10 finds wide use in imaging and research. The ability to track light-emitting cells in a small laboratory animal such as a mouse or rat opens up a wide range of applications in pharmaceutical and toxilogical research. These include in vivo monitoring of infectious diseases, tumor growth in metastases, transgene expression, compound toxicity, and viral infection or delivery systems for gene therapy. The ability to detect signals in real-time and in living animals means that the progression of a disease or biological process can be studied throughout an experiment with the same set of animals without a need to sacrifice for each data point.
  • Imaging system 10 comprises an imaging box 12 having a door 18 and inner walls 19 (FIG. 14B) that define an interior cavity 21 that is adapted to receive a mouse 2 in which low intensity light is to be detected. Imaging box 12 is suitable for imaging including the capture of low intensity light on the order of individual photons, for example. Imaging box 12 is often referred to as “light-tight”. That is, box 12 seals out essentially all of the external light from the ambient room from entering the box 12, and may include one or more seals that prevent light passage into the box when door 18 is closed. In a specific embodiment, door 18 comprises one or more light-tight features such as a double baffle seal, while the remainder of chamber 21 is configured to minimize any penetration of light into cavity 21.
  • Mouse 2 is placed within box 12 for imaging by opening door 18, inserting the mouse in chamber 21, and closing door 18. Suitable imaging systems are available from Xenogen Corporation from Alameda, Calif., and include the IVIS® Spectrum, IVIS® 3D Series, IVIS® 200 Series, IVIS® 100 Series, and IVIS® Lumina. Further description of a suitable imaging box 12 is provided in commonly owned U.S. Pat. No. 7,113,217 entitled “3-D Imaging Apparatus for In-Vivo Representations”, which is incorporated by reference herein in its entirety for all purposes. Although imaging system 10 is shown with a single cabinet design, other embodiments of the present invention include a disparate imaging box 12 and desktop computer that includes processing system 28 and a dedicated display.
  • Imaging box 12 includes an upper housing 16 adapted to receive a camera 20 (FIG. 14B). A high sensitivity camera 20, e.g., an intensified or a charge-coupled device (CCD) camera, is mounted on top of upper housing 16 and positioned above imaging box 12. CCD camera 20 is capable of capturing luminescent, fluorescent, structured light and photographic (i.e., reflection based images) images of a living sample or phantom device placed within imaging box 12. One suitable camera includes a Spectral Instruments 620 Series as provided by Spectral Instruments of Tucson, Ariz. CCD camera 20 is cooled by a suitable source thermoelectric chiller. Other methods, such as liquid nitrogen, may be used to cool camera 20. Camera may also be side-mounted, or attached to a moving chassis that moves the camera in cavity 21.
  • Imaging system 10 may also comprise a lens (not shown) that collects light from the specimen or phantom device and provides the light to the camera 20. A stage 25 forms the bottom floor of imaging chamber 21 and includes motors and controls that allow stage 25 to move up and down to vary the field of view 23 for camera 20. A multiple position filter wheel may also be provided to enable spectral imaging capability. Imaging box 12 may also include one or more light emitting diodes on the top portion of chamber 21 to illuminate a sample during photographic image capture. Other features may include a gas anesthesia system to keep the mouse anesthetized and/or a heated shelf to maintain an animal's body temperature during image capture and anesthesia.
  • Imaging box 12 also includes one or more fluorescent excitation light sources. In one embodiment, box 12 includes a trans-illumination device and an epi-illumination device. As mentioned above with respect to FIGS. 6A and 6B, the trans-illumination device is configured to direct light into a first surface of the mouse, where diffused light exits a second surface of the mouse. The epi-illumination device is configured direct light onto a third surface of the specimen, where the diffused light exits the third surface of the mouse. Further description of fluorescent excitation light sources is provided in commonly owned and co-pending patent application Ser. No. 11/434,606, which is incorporated by reference in its entirety for all purposes.
  • A structured light source is included in imaging box. The structured light source includes a mechanism for transmitting a set of lines onto the object from an angle. The lines are displaced, or phase shifted relative to a stage, when they encounter an object with finite height, such as a mouse. This phase shift provides structured light information for the object. Camera 20 then captures the structured light information. Using software that employs a structured light analysis, surface topography data for the object (over its entire surface or a portion) is determined from the phase shift of the lines.
  • FIG. 14B shows system 10 with the removal of a side panel for imaging box 12 to illustrate various electronics and processing components included in system 10. Imaging system 10 comprises image processing unit 26 and processing system 28. Image processing unit 26 optionally interfaces between camera 20 and processing system 28 and may assist with image data collection and video data processing. Processing system 28, which may be of any suitable type, comprises hardware including a processor 28 a and one or more memory components such as random-access memory (RAM) 28 b and read-only memory (ROM) 28 c.
  • Processor 28 a (also referred to as a central processing unit, or CPU) couples to storage devices including memory 28 b and 28 c. ROM 28 c serves to transfer data and instructions uni-directionally to the CPU, while RAM 28 b typically transfers data and instructions in a bi-directional manner. A fixed disk is also coupled bi-directionally to processor 28 a; it provides additional data storage capacity and may also include any of the computer-readable media described below. The fixed disk may be used to store software, programs, imaging data and the like and is typically a secondary storage medium (such as a hard disk).
  • Processor 28 a communicates with various components in imaging box 12. To provide communication with, and control of, one or more system 10 components, processing system 28 employs software stored in memory 28 c that is configured to permit communication with and/or control of components in imaging box 12. For example, processing system 28 may include hardware and software configured to control camera 20. The processing hardware and software may include an I/O card, control logic for controlling camera 20. Components controlled by computer 28 may also include motors responsible for camera 20 focus, motors responsible for position control of a platform supporting the sample, a motor responsible for position control of a filter lens, f-stop, etc.
  • Processing system 28 may also interface with an external visual display (such as computer monitor) and input devices such as a keyboard and mouse. A graphical user interface that facilitates user interaction with imaging system 10 may also be stored on system 28, output on the visual display and receive user input from the keyboard and mouse. The graphical user interface allows a user to view imaging results and also acts an interface to control the imaging system 10. One suitable imaging software includes “LivingImage” as provided by Xenogen Corporation of Alameda, Calif.
  • Processing system 28 may comprise software, hardware or a combination thereof. System 28 may also include additional imaging hardware and software, tomographic reconstruction software that implements process flows and methods described above, and image processing logic and instructions for processing information obtained by camera 20. For example, stored instructions run by processor 28 a may include instructions for i) receiving image data corresponding to light emitted from a mouse as described herein, ii) building a 3-D digital representation of a fluorescent probe internal to a mouse using data included in an image, and iii) outputting results of the tomographic reconstruction on a display such as a video monitor.
  • Imaging system 10 employs a quantitative model that estimates the diffusion of photons in tissue. In one embodiment, the model processes in vivo image data and in order to spatially resolve a 3D representation of the size, shape, and location of the light emitting source. Regardless of the imaging and computing system configuration, imaging apparatus 10 may employ one or more memories or memory modules configured to store program instructions for obtaining a 3D representation of a probe located inside a sample and other functions of the present invention described herein. Such memory or memories may also be configured to store data structures, imaging data, or other specific non-program information described herein. Because such information and program instructions may be employed to implement the systems/methods described herein, the present invention relates to machine-readable media that include program instructions, state information, etc. for performing various operations described herein. Examples of tangible machine-readable media include, but are not limited to, magnetic media such as hard disks, floppy disks, and magnetic tape; optical media such as CD-ROM disks; magneto-optical media such as floptical disks; and hardware devices that are specially configured to store and perform program instructions, such as read-only memory devices (ROM) and random access memory (RAM). Examples of program instructions include both machine code, such as produced by a compiler, and files containing higher level code that may be executed by the computer using an interpreter. The invention may also be embodied in a carrier wave traveling over an appropriate medium such as airwaves, optical lines, electric lines, etc.
  • While this invention has been described in terms of several preferred embodiments, there are alterations, permutations, and equivalents which fall within the scope of this invention which have been omitted for brevity's sake. For example, although the methods have primarily been discussed with respect to fluorescent light imaging, the present invention is also well-suited for use with other wavelength ranges and imaging modalities, such as near IR. In addition, although the methods have been described with respect to solving for autofluorescence separately from the tomographic reconstruction to expedite finding a solution, they may be combined to accommodate minor changes in tissue properties, albeit with less constrained computations and a need for more computational resources. It is therefore intended that the scope of the invention should be determined with reference to the appended claims.

Claims (1)

1. A method for obtaining a light distribution located inside an animal, the method comprising:
obtaining one or more fluorescent images of at least a portion of the animal;
obtaining a three dimensional representation of a surface portion of the animal;
dividing the three dimensional surface representation into a set of surface elements;
mapping fluorescent image data from the one or more fluorescent images to the set of surface elements to create fluorescent light emission data from the set of surface elements;
creating a set of volume elements within the animal;
converting the fluorescent light emission data from the set of surface elements into photon density internal to the animal using the set of surface elements and the set of volume elements; and
determining a three-dimensional representation of a fluorescent probe distribution internal to the animal using the photon density internal to the animal and the set of volume elements.
US11/829,927 2002-07-16 2007-07-29 Fluorescent light tomography Expired - Lifetime US7555332B2 (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
US11/829,927 US7555332B2 (en) 2002-07-16 2007-07-29 Fluorescent light tomography

Applications Claiming Priority (7)

Application Number Priority Date Filing Date Title
US39631302P 2002-07-16 2002-07-16
US39535702P 2002-07-16 2002-07-16
US39645802P 2002-07-16 2002-07-16
US10/606,976 US7616985B2 (en) 2002-07-16 2003-06-25 Method and apparatus for 3-D imaging of internal light sources
US84024706P 2006-08-24 2006-08-24
US11/733,358 US7599731B2 (en) 2002-07-16 2007-04-10 Fluorescent light tomography
US11/829,927 US7555332B2 (en) 2002-07-16 2007-07-29 Fluorescent light tomography

Related Parent Applications (2)

Application Number Title Priority Date Filing Date
US10/606,976 Continuation-In-Part US7616985B2 (en) 2002-07-16 2003-06-25 Method and apparatus for 3-D imaging of internal light sources
US11/733,358 Continuation US7599731B2 (en) 2002-07-16 2007-04-10 Fluorescent light tomography

Publications (2)

Publication Number Publication Date
US20080031494A1 true US20080031494A1 (en) 2008-02-07
US7555332B2 US7555332B2 (en) 2009-06-30

Family

ID=38648530

Family Applications (2)

Application Number Title Priority Date Filing Date
US11/733,358 Expired - Lifetime US7599731B2 (en) 2002-07-16 2007-04-10 Fluorescent light tomography
US11/829,927 Expired - Lifetime US7555332B2 (en) 2002-07-16 2007-07-29 Fluorescent light tomography

Family Applications Before (1)

Application Number Title Priority Date Filing Date
US11/733,358 Expired - Lifetime US7599731B2 (en) 2002-07-16 2007-04-10 Fluorescent light tomography

Country Status (3)

Country Link
US (2) US7599731B2 (en)
EP (1) EP2146752B1 (en)
WO (1) WO2008124640A1 (en)

Cited By (5)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20090041449A1 (en) * 2005-12-12 2009-02-12 Topins Co., Ltd. The testing matter photographing apparatus attached three dimensional image photographing device
US20090066960A1 (en) * 2007-09-10 2009-03-12 Biospace Lab Luminescence Imagining Installation and Method
WO2010127302A1 (en) * 2009-05-01 2010-11-04 The Trustees Of Columbia University In The City Of New York Systems, methods and computer-accessible media for obtaining three-dimensional information from two-dimensional fluorescence emission data
US20110272597A1 (en) * 2009-01-30 2011-11-10 Fujifilm Corporation Method for generating optical tomographic information, optical tomographic information generating apparatus, and storage medium
US9456789B2 (en) * 2011-10-26 2016-10-04 Canon Kabushiki Kaisha Object information acquisition apparatus

Families Citing this family (34)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
AU2002303819B2 (en) 2001-05-17 2007-03-01 Xenogen Corporation Method and apparatus for determining target depth, brightness and size within a body region
US10335038B2 (en) * 2006-08-24 2019-07-02 Xenogen Corporation Spectral unmixing for in-vivo imaging
US10775308B2 (en) 2006-08-24 2020-09-15 Xenogen Corporation Apparatus and methods for determining optical tissue properties
US20080218727A1 (en) * 2006-12-22 2008-09-11 Art, Advanced Research Technologies Inc. Method and apparatus for optical image reconstruction using contour determination
EP1955651A1 (en) * 2007-02-06 2008-08-13 Siemens Schweiz AG Device for the physical location of a moveable body part
US7738627B2 (en) * 2007-09-28 2010-06-15 The Invention Science Fund I, Llc Geometric X-ray fluorescence visualizer, imager, or information provider
US7664224B2 (en) * 2007-09-28 2010-02-16 Searete Llc X-ray fluorescence visualizing, imaging, or information providing of chemicals, compounds, or biological materials
US8041005B2 (en) * 2007-09-28 2011-10-18 The Invention Science Fund I, Llc X-ray fluorescence visualizer, imager, or information provider
US7660385B2 (en) 2007-09-28 2010-02-09 Searete Llc Time of flight aspects for X-Ray fluorescence visualizer, imager, or information provider
US7724867B2 (en) * 2007-09-28 2010-05-25 Invention Science Fund I, Llc Proximity-based X-Ray fluorescence visualizer, imager, or information provider
US7825376B2 (en) * 2007-09-28 2010-11-02 The Invention Science Fund I Scintillator aspects for X-ray fluorescence visualizer, imager, or information provider
US7773722B2 (en) * 2007-09-28 2010-08-10 The Invention Science Fund I, Llc Personal transportable X-ray fluorescence visualizing, imaging, or information providing
US7702066B2 (en) 2007-09-28 2010-04-20 Searete Llc Portable aspects for x-ray fluorescence visualizer, imager, or information provider
US20090086899A1 (en) * 2007-09-28 2009-04-02 Searete Llc, A Limited Liability Corporation Of The State Of Delaware Repositioning X-ray fluorescence visualizer, imager, or information provider
US7653173B2 (en) * 2007-09-28 2010-01-26 Searete Llc Combining X-ray fluorescence visualizer, imager, or information provider
US8000438B2 (en) * 2007-09-28 2011-08-16 The Invention Science Fund I, Llc Tool based X-ray fluorescence visualizing, imaging, or information providing
US20090086903A1 (en) * 2007-09-28 2009-04-02 Searete LLC, a limited liability corporation of Selective elemental color providing for X-ray fluorescence visualization, imaging, or information providing
US7649975B2 (en) * 2007-09-28 2010-01-19 Searete Llc X-ray fluorescence visualizing, imaging, or information providing of chemicals, compounds, or biological materials
EP2067432A3 (en) * 2007-12-05 2012-12-26 FUJIFILM Corporation Image processing system, image processing method, and program
US8390291B2 (en) * 2008-05-19 2013-03-05 The Board Of Regents, The University Of Texas System Apparatus and method for tracking movement of a target
US9042967B2 (en) 2008-05-20 2015-05-26 University Health Network Device and method for wound imaging and monitoring
EP2322101A4 (en) * 2008-07-08 2013-04-10 Hitachi Ltd Light meter
JP5885354B2 (en) 2010-06-08 2016-03-15 スミス アンド ネフュー インコーポレーテッド Implant components and methods
US9347894B2 (en) 2010-09-01 2016-05-24 Spectral Instruments Imaging, LLC Methods and systems for producing visible light and x-ray image data
JP2013538350A (en) 2010-09-01 2013-10-10 スペクトラル・インストゥルメンツ・イメージング・エルエルシー Excitation light source assembly
US8729502B1 (en) 2010-10-28 2014-05-20 The Research Foundation For The State University Of New York Simultaneous, single-detector fluorescence detection of multiple analytes with frequency-specific lock-in detection
JP5713387B2 (en) * 2011-01-28 2015-05-07 学校法人慈恵大学 Trace optical imaging apparatus and trace optical imaging method
US9314218B2 (en) 2011-06-20 2016-04-19 Caliper Life Sciences, Inc. Integrated microtomography and optical imaging systems
CN103356165A (en) * 2012-03-29 2013-10-23 锐珂(上海)医疗器材有限公司 Multispectral fluorescence three-dimensional tomography method and device
DE102012016122A1 (en) * 2012-08-15 2014-02-20 Westfälische Wilhelms-Universität Münster Method for detecting e.g. position of insect within observation arena during performing experiments for evaluating biological behavior, involves detecting light scattered due to interrupted reflectance in animal by camera
US8984800B2 (en) * 2013-03-15 2015-03-24 Technology Sg, L.P. Radiating systems for affecting insect behavior
CN115989999A (en) 2014-07-24 2023-04-21 大学健康网络 Collection and analysis of data for diagnostic purposes
WO2017173153A1 (en) 2016-03-30 2017-10-05 Ebay, Inc. Digital model optimization responsive to orientation sensor data
WO2022192242A1 (en) * 2021-03-10 2022-09-15 Designs For Vision, Inc. Fluorescence based image capture system

Citations (92)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US4687325A (en) * 1985-03-28 1987-08-18 General Electric Company Three-dimensional range camera
US4687352A (en) * 1984-12-29 1987-08-18 Brother Kogyo Kabushiki Kaisha Printer with an image reader
US4773097A (en) * 1984-05-31 1988-09-20 Omron Tateisi Electronics Co. Image analyzing apparatus
US5202091A (en) * 1985-03-01 1993-04-13 Lisenbee Wayne F Luminescence measurement arrangement
US5205291A (en) * 1988-11-08 1993-04-27 Health Research, Inc. In vivo fluorescence photometer
US5242441A (en) * 1992-02-24 1993-09-07 Boaz Avitall Deflectable catheter with rotatable tip electrode
US5319209A (en) * 1991-11-20 1994-06-07 Hammamatsu Photonics K.K. Method for measurement of intracellular ion concentration using fluorescence probe dyes
US5334193A (en) * 1992-11-13 1994-08-02 American Cardiac Ablation Co., Inc. Fluid cooled ablation catheter
US5414258A (en) * 1993-11-22 1995-05-09 Angstrom Technologies, Inc. Apparatus and method for calibration of fluorescence detectors
US5452723A (en) * 1992-07-24 1995-09-26 Massachusetts Institute Of Technology Calibrated spectrographic imaging
US5584872A (en) * 1992-11-13 1996-12-17 Scimed Life Systems, Inc. Electrophysiology energy treatment devices and methods of use
US5594253A (en) * 1994-12-28 1997-01-14 Lockheed Missiles And Space Company, Inc. Hybrid luminescent device for imaging of ionizing and penetrating radiation
US5636299A (en) * 1994-12-28 1997-06-03 Lockheed Missiles & Space Company, Inc. Hybrid luminescent device and method for imaging penetrating radiation
US5637874A (en) * 1994-12-20 1997-06-10 Biosensor Laboratories Co., Ltd. Apparatus and method for measuring chemiluminescence
US5650135A (en) * 1994-07-01 1997-07-22 The Board Of Trustees Of The Leland Stanford Junior University Non-invasive localization of a light-emitting conjugate in a mammal
US5661562A (en) * 1994-08-02 1997-08-26 Aharon; Oren Tomographic imaging system with scanning apertures
US5672881A (en) * 1994-09-14 1997-09-30 Glyko, Inc. Charge-coupled device imaging apparatus
US5705807A (en) * 1994-10-24 1998-01-06 Nissan Motor Co., Ltd. Photo detecting apparatus for detecting reflected light from an object and excluding an external light componet from the reflected light
US5738101A (en) * 1996-01-18 1998-04-14 The Regents Of The University Of California Optical imaging through turbid media with a degenerate four-wave mixing correlation time gate
US5746210A (en) * 1993-02-26 1998-05-05 David A. Benaron Device and method for detection, localization, and characterization of inhomogeneities in turbid media
US5807262A (en) * 1995-12-22 1998-09-15 U.S. Philips Corporation Device for localizing an object in a turbid medium using a modulated light source
US5812310A (en) * 1996-10-16 1998-09-22 Applied Precision, Inc. Orthogonal high accuracy microscope stage
US5818587A (en) * 1995-11-16 1998-10-06 Biophotonics Information Laboratories Ltd. Image measuring apparatus
US5835617A (en) * 1996-01-18 1998-11-10 Hamamatsu Photonics K.K. Optical computer tomographic apparatus and image reconstruction method using optical computer tomography
US5840572A (en) * 1994-10-11 1998-11-24 United States Of America As Represented By The Secretary Of The Navy Bioluminescent bioassay system
US5867250A (en) * 1996-05-03 1999-02-02 Baron; William S. Apparatus and method for optically mapping front and back surface topographies of an object
US5865754A (en) * 1995-08-24 1999-02-02 Purdue Research Foundation Office Of Technology Transfer Fluorescence imaging system and method
US5917190A (en) * 1993-10-29 1999-06-29 Trustees Of The University Of Pennsylvania Object imaging using diffuse light
US5943129A (en) * 1997-08-07 1999-08-24 Cambridge Research & Instrumentation Inc. Fluorescence imaging system
US5953446A (en) * 1996-05-17 1999-09-14 Therma-Wave, Inc. Method and apparatus for optical data analysis
US5963658A (en) * 1997-01-27 1999-10-05 University Of North Carolina Method and apparatus for detecting an abnormality within a host medium
US5970164A (en) * 1994-08-11 1999-10-19 Sophisview Technologies, Ltd. System and method for diagnosis of living tissue diseases
US6069698A (en) * 1997-08-28 2000-05-30 Olympus Optical Co., Ltd. Optical imaging apparatus which radiates a low coherence light beam onto a test object, receives optical information from light scattered by the object, and constructs therefrom a cross-sectional image of the object
US6108576A (en) * 1996-03-18 2000-08-22 The Research Foundation Of City College Of New York Time-resolved diffusion tomographic 2D and 3D imaging in highly scattering turbid media
US6175407B1 (en) * 1998-12-17 2001-01-16 Identix Incorporated Apparatus and method for optically imaging features on the surface of a hand
US6205347B1 (en) * 1998-02-27 2001-03-20 Picker International, Inc. Separate and combined multi-modality diagnostic imaging system
US6208886B1 (en) * 1997-04-04 2001-03-27 The Research Foundation Of City College Of New York Non-linear optical tomography of turbid media
US6219566B1 (en) * 1999-07-13 2001-04-17 Photonics Research Ontario Method of measuring concentration of luminescent materials in turbid media
US6217847B1 (en) * 1994-07-01 2001-04-17 The Board Of Trustees Of The Leland Stanford Junior University Non-invasive localization of a light-emitting conjugate in a mammal
US6242743B1 (en) * 1998-08-11 2001-06-05 Mosaic Imaging Technology, Inc. Non-orbiting tomographic imaging system
US6252623B1 (en) * 1998-05-15 2001-06-26 3Dmetrics, Incorporated Three dimensional imaging system
US6264610B1 (en) * 1999-05-05 2001-07-24 The University Of Connecticut Combined ultrasound and near infrared diffused light imaging system
US6267477B1 (en) * 1996-02-23 2001-07-31 Talia Technology Ltd. Three dimensional imaging apparatus and a method for use thereof
US6321111B1 (en) * 1995-02-21 2001-11-20 Massachusetts Institute Of Technology Optical imaging using time gated scattered light
US6332087B1 (en) * 1996-07-05 2001-12-18 The Carolinas Heart Institute Electromagnetic imaging and therapeutic (EMIT) systems
US20020001080A1 (en) * 1999-08-06 2002-01-03 Cambridge Research & Instrumentation, Inc., Massachusetts Corporation Spectral imaging system
US6364829B1 (en) * 1999-01-26 2002-04-02 Newton Laboratories, Inc. Autofluorescence imaging system for endoscopy
US6373557B1 (en) * 1997-12-23 2002-04-16 Siemens Aktiengesellschaft Method and apparatus for picking up a three-dimensional range image
US6377353B1 (en) * 2000-03-07 2002-04-23 Pheno Imaging, Inc. Three-dimensional measuring system for animals using structured light
US6381302B1 (en) * 2000-07-05 2002-04-30 Canon Kabushiki Kaisha Computer assisted 2D adjustment of stereo X-ray images
US6392241B1 (en) * 1996-07-10 2002-05-21 Packard Instrument Company, Inc. Fiber optic coupling device for detecting fluorescence samples
US6394965B1 (en) * 2000-08-15 2002-05-28 Carbon Medical Technologies, Inc. Tissue marking using biocompatible microparticles
US6403947B1 (en) * 1999-03-18 2002-06-11 Cambridge Research & Instrumentation Inc. High-efficiency multiple probe imaging system
US6415051B1 (en) * 1999-06-24 2002-07-02 Geometrix, Inc. Generating 3-D models using a manually operated structured light source
US6429943B1 (en) * 2000-03-29 2002-08-06 Therma-Wave, Inc. Critical dimension analysis with simultaneous multiple angle of incidence measurements
US20030002028A1 (en) * 2001-05-17 2003-01-02 Rice Bradley W. Method and apparatus for determining target depth, brightness and size within a body region
US6549288B1 (en) * 1998-05-14 2003-04-15 Viewpoint Corp. Structured-light, triangulation-based three-dimensional digitizer
US20030099329A1 (en) * 2001-10-23 2003-05-29 Schotland John Carl System and method for scanning near-field optical tomography
US6597931B1 (en) * 2000-09-18 2003-07-22 Photonify Technologies, Inc. System and method for absolute oxygen saturation
US6615061B1 (en) * 1998-11-23 2003-09-02 Abbott Laboratories Optical sensor having a selectable sampling distance for determination of analytes
US6615063B1 (en) * 2000-11-27 2003-09-02 The General Hospital Corporation Fluorescence-mediated molecular tomography
US6618463B1 (en) * 2002-07-09 2003-09-09 Washington University System and method for single-beam internal reflection tomography
US6618152B2 (en) * 2000-05-09 2003-09-09 Fuji Photo Film Co., Ltd. Optical coherence tomography apparatus using optical-waveguide structure which reduces pulse width of low-coherence light
US6628747B1 (en) * 2002-06-21 2003-09-30 Washington University In St. Louis System and method for dual-beam internal reflection tomography
US6628401B2 (en) * 2000-03-10 2003-09-30 Fuji Photo Film Co., Ltd. Optical tomography imaging method and apparatus
US6636755B2 (en) * 2000-09-26 2003-10-21 Fuji Photo Film Co., Ltd. Method and apparatus for obtaining an optical tomographic image of a sentinel lymph node
US6642953B1 (en) * 1997-12-02 2003-11-04 Francisco Soria Melguizo, S.A. System for analyzing images produced by bacterial reactions
US6646678B1 (en) * 1998-12-15 2003-11-11 Fuji Photo Film Co., Ltd. Photographing system
US6665072B2 (en) * 2000-02-25 2003-12-16 Cambridge Research & Instrumentation Inc. Instantaneous dual band fluorescence detection systems
US20040010192A1 (en) * 2000-06-15 2004-01-15 Spectros Corporation Optical imaging of induced signals in vivo under ambient light conditions
US20040015062A1 (en) * 2000-11-27 2004-01-22 Vasilis Ntziachristos Fluorescence-mediated molecular tomography
US20040021771A1 (en) * 2002-07-16 2004-02-05 Xenogen Corporation Method and apparatus for 3-D imaging of internal light sources
US6690520B1 (en) * 1996-05-15 2004-02-10 Sysmex Corporation Optical system for visualizing an object in a light scattering medium
US20040027659A1 (en) * 2002-08-08 2004-02-12 Messerschmidt Robert G. Sample holder
US6693710B1 (en) * 2000-06-16 2004-02-17 Cambridge Research & Instrumentation Inc. Polarization imaging system
US6710770B2 (en) * 2000-02-11 2004-03-23 Canesta, Inc. Quasi-three-dimensional method and apparatus to detect and localize interaction of user-object and virtual transfer device
US20040085536A1 (en) * 2002-11-01 2004-05-06 Schotland John Carl Tomography system and method using nonlinear reconstruction of scattered radiation
US6750964B2 (en) * 1999-08-06 2004-06-15 Cambridge Research And Instrumentation, Inc. Spectral imaging methods and systems
US6775567B2 (en) * 2000-02-25 2004-08-10 Xenogen Corporation Imaging apparatus
US6813030B2 (en) * 2000-02-18 2004-11-02 Japan Science And Technology Corporation Optical interference tomographic image observing apparatus
US20050149877A1 (en) * 1999-11-15 2005-07-07 Xenogen Corporation Graphical user interface for 3-D in-vivo imaging
US6919919B2 (en) * 2002-02-06 2005-07-19 Xenogen Corporation Light calibration device for use in low level light imaging systems
US6924893B2 (en) * 2002-05-13 2005-08-02 Marine Biological Laboratory Enhancing polarized light microscopy
US6963375B1 (en) * 2000-07-07 2005-11-08 Axis, Ab Image altering device for an image producing apparatus
US20060119865A1 (en) * 2004-12-06 2006-06-08 Hoyt Clifford C Systems and methods for in-vivo optical imaging and measurement
US20060173354A1 (en) * 2003-02-05 2006-08-03 Vasilis Ntziachristos Method and system for free space optical tomography of diffuse media
US20060203244A1 (en) * 2002-02-22 2006-09-14 Xenogen Corporation Dual illumination system for an imaging apparatus and method
US7113217B2 (en) * 2001-07-13 2006-09-26 Xenogen Corporation Multi-view imaging apparatus
US20060245631A1 (en) * 2005-01-27 2006-11-02 Richard Levenson Classifying image features
US20060268153A1 (en) * 2005-05-11 2006-11-30 Xenogen Corporation Surface contruction using combined photographic and structured light information
US7184047B1 (en) * 1996-12-24 2007-02-27 Stephen James Crampton Method and apparatus for the generation of computer graphic representations of individuals
US7263157B2 (en) * 2004-06-24 2007-08-28 Siemens Aktiengesellschaft Imaging tomography apparatus with two acquisition systems, and method for determining their system angles

Family Cites Families (7)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
JP3107914B2 (en) 1992-07-20 2000-11-13 浜松ホトニクス株式会社 Apparatus and method for measuring absorption information inside scattering absorber
JP3433534B2 (en) 1994-11-07 2003-08-04 浜松ホトニクス株式会社 Method and apparatus for measuring scattering and absorption characteristics in scattering medium
US7298415B2 (en) * 2001-07-13 2007-11-20 Xenogen Corporation Structured light imaging apparatus
EP1514093B1 (en) 2002-06-04 2021-04-28 Visen Medical, Inc. Imaging volumes with arbitrary geometries in non-contact tomography
US7034303B2 (en) 2003-06-27 2006-04-25 Washington University System and method of image reconstruction for optical tomography with limited data
US7190991B2 (en) * 2003-07-01 2007-03-13 Xenogen Corporation Multi-mode internal imaging
US20060016078A1 (en) * 2004-07-07 2006-01-26 Jeffrey Bladow Method for manufacturing a reinforced structural component, and article manufactured thereby

Patent Citations (99)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US4773097A (en) * 1984-05-31 1988-09-20 Omron Tateisi Electronics Co. Image analyzing apparatus
US4687352A (en) * 1984-12-29 1987-08-18 Brother Kogyo Kabushiki Kaisha Printer with an image reader
US5202091A (en) * 1985-03-01 1993-04-13 Lisenbee Wayne F Luminescence measurement arrangement
US4687325A (en) * 1985-03-28 1987-08-18 General Electric Company Three-dimensional range camera
US5205291A (en) * 1988-11-08 1993-04-27 Health Research, Inc. In vivo fluorescence photometer
US5319209A (en) * 1991-11-20 1994-06-07 Hammamatsu Photonics K.K. Method for measurement of intracellular ion concentration using fluorescence probe dyes
US5242441A (en) * 1992-02-24 1993-09-07 Boaz Avitall Deflectable catheter with rotatable tip electrode
US5452723A (en) * 1992-07-24 1995-09-26 Massachusetts Institute Of Technology Calibrated spectrographic imaging
US5334193A (en) * 1992-11-13 1994-08-02 American Cardiac Ablation Co., Inc. Fluid cooled ablation catheter
US5584872A (en) * 1992-11-13 1996-12-17 Scimed Life Systems, Inc. Electrophysiology energy treatment devices and methods of use
US5746210A (en) * 1993-02-26 1998-05-05 David A. Benaron Device and method for detection, localization, and characterization of inhomogeneities in turbid media
US5917190A (en) * 1993-10-29 1999-06-29 Trustees Of The University Of Pennsylvania Object imaging using diffuse light
US5414258A (en) * 1993-11-22 1995-05-09 Angstrom Technologies, Inc. Apparatus and method for calibration of fluorescence detectors
US5650135A (en) * 1994-07-01 1997-07-22 The Board Of Trustees Of The Leland Stanford Junior University Non-invasive localization of a light-emitting conjugate in a mammal
US6217847B1 (en) * 1994-07-01 2001-04-17 The Board Of Trustees Of The Leland Stanford Junior University Non-invasive localization of a light-emitting conjugate in a mammal
US5661562A (en) * 1994-08-02 1997-08-26 Aharon; Oren Tomographic imaging system with scanning apertures
US5970164A (en) * 1994-08-11 1999-10-19 Sophisview Technologies, Ltd. System and method for diagnosis of living tissue diseases
US5672881A (en) * 1994-09-14 1997-09-30 Glyko, Inc. Charge-coupled device imaging apparatus
US5840572A (en) * 1994-10-11 1998-11-24 United States Of America As Represented By The Secretary Of The Navy Bioluminescent bioassay system
US5705807A (en) * 1994-10-24 1998-01-06 Nissan Motor Co., Ltd. Photo detecting apparatus for detecting reflected light from an object and excluding an external light componet from the reflected light
US5637874A (en) * 1994-12-20 1997-06-10 Biosensor Laboratories Co., Ltd. Apparatus and method for measuring chemiluminescence
US5636299A (en) * 1994-12-28 1997-06-03 Lockheed Missiles & Space Company, Inc. Hybrid luminescent device and method for imaging penetrating radiation
US5594253A (en) * 1994-12-28 1997-01-14 Lockheed Missiles And Space Company, Inc. Hybrid luminescent device for imaging of ionizing and penetrating radiation
US6321111B1 (en) * 1995-02-21 2001-11-20 Massachusetts Institute Of Technology Optical imaging using time gated scattered light
US5865754A (en) * 1995-08-24 1999-02-02 Purdue Research Foundation Office Of Technology Transfer Fluorescence imaging system and method
US5818587A (en) * 1995-11-16 1998-10-06 Biophotonics Information Laboratories Ltd. Image measuring apparatus
US5807262A (en) * 1995-12-22 1998-09-15 U.S. Philips Corporation Device for localizing an object in a turbid medium using a modulated light source
US5835617A (en) * 1996-01-18 1998-11-10 Hamamatsu Photonics K.K. Optical computer tomographic apparatus and image reconstruction method using optical computer tomography
US5738101A (en) * 1996-01-18 1998-04-14 The Regents Of The University Of California Optical imaging through turbid media with a degenerate four-wave mixing correlation time gate
US6267477B1 (en) * 1996-02-23 2001-07-31 Talia Technology Ltd. Three dimensional imaging apparatus and a method for use thereof
US6108576A (en) * 1996-03-18 2000-08-22 The Research Foundation Of City College Of New York Time-resolved diffusion tomographic 2D and 3D imaging in highly scattering turbid media
US5867250A (en) * 1996-05-03 1999-02-02 Baron; William S. Apparatus and method for optically mapping front and back surface topographies of an object
US6690520B1 (en) * 1996-05-15 2004-02-10 Sysmex Corporation Optical system for visualizing an object in a light scattering medium
US5953446A (en) * 1996-05-17 1999-09-14 Therma-Wave, Inc. Method and apparatus for optical data analysis
US6332087B1 (en) * 1996-07-05 2001-12-18 The Carolinas Heart Institute Electromagnetic imaging and therapeutic (EMIT) systems
US6392241B1 (en) * 1996-07-10 2002-05-21 Packard Instrument Company, Inc. Fiber optic coupling device for detecting fluorescence samples
US5812310A (en) * 1996-10-16 1998-09-22 Applied Precision, Inc. Orthogonal high accuracy microscope stage
US7184047B1 (en) * 1996-12-24 2007-02-27 Stephen James Crampton Method and apparatus for the generation of computer graphic representations of individuals
US5963658A (en) * 1997-01-27 1999-10-05 University Of North Carolina Method and apparatus for detecting an abnormality within a host medium
US6208886B1 (en) * 1997-04-04 2001-03-27 The Research Foundation Of City College Of New York Non-linear optical tomography of turbid media
US5943129A (en) * 1997-08-07 1999-08-24 Cambridge Research & Instrumentation Inc. Fluorescence imaging system
US6069698A (en) * 1997-08-28 2000-05-30 Olympus Optical Co., Ltd. Optical imaging apparatus which radiates a low coherence light beam onto a test object, receives optical information from light scattered by the object, and constructs therefrom a cross-sectional image of the object
US6642953B1 (en) * 1997-12-02 2003-11-04 Francisco Soria Melguizo, S.A. System for analyzing images produced by bacterial reactions
US6373557B1 (en) * 1997-12-23 2002-04-16 Siemens Aktiengesellschaft Method and apparatus for picking up a three-dimensional range image
US6205347B1 (en) * 1998-02-27 2001-03-20 Picker International, Inc. Separate and combined multi-modality diagnostic imaging system
US6549288B1 (en) * 1998-05-14 2003-04-15 Viewpoint Corp. Structured-light, triangulation-based three-dimensional digitizer
US6252623B1 (en) * 1998-05-15 2001-06-26 3Dmetrics, Incorporated Three dimensional imaging system
US6242743B1 (en) * 1998-08-11 2001-06-05 Mosaic Imaging Technology, Inc. Non-orbiting tomographic imaging system
US6615061B1 (en) * 1998-11-23 2003-09-02 Abbott Laboratories Optical sensor having a selectable sampling distance for determination of analytes
US6646678B1 (en) * 1998-12-15 2003-11-11 Fuji Photo Film Co., Ltd. Photographing system
US6175407B1 (en) * 1998-12-17 2001-01-16 Identix Incorporated Apparatus and method for optically imaging features on the surface of a hand
US6364829B1 (en) * 1999-01-26 2002-04-02 Newton Laboratories, Inc. Autofluorescence imaging system for endoscopy
US6403947B1 (en) * 1999-03-18 2002-06-11 Cambridge Research & Instrumentation Inc. High-efficiency multiple probe imaging system
US6264610B1 (en) * 1999-05-05 2001-07-24 The University Of Connecticut Combined ultrasound and near infrared diffused light imaging system
US6529627B1 (en) * 1999-06-24 2003-03-04 Geometrix, Inc. Generating 3D models by combining models from a video-based technique and data from a structured light technique
US6415051B1 (en) * 1999-06-24 2002-07-02 Geometrix, Inc. Generating 3-D models using a manually operated structured light source
US6219566B1 (en) * 1999-07-13 2001-04-17 Photonics Research Ontario Method of measuring concentration of luminescent materials in turbid media
US6373568B1 (en) * 1999-08-06 2002-04-16 Cambridge Research & Instrumentation, Inc. Spectral imaging system
US6750964B2 (en) * 1999-08-06 2004-06-15 Cambridge Research And Instrumentation, Inc. Spectral imaging methods and systems
US20020001080A1 (en) * 1999-08-06 2002-01-03 Cambridge Research & Instrumentation, Inc., Massachusetts Corporation Spectral imaging system
US6690466B2 (en) * 1999-08-06 2004-02-10 Cambridge Research & Instrumentation, Inc. Spectral imaging system
US20050149877A1 (en) * 1999-11-15 2005-07-07 Xenogen Corporation Graphical user interface for 3-D in-vivo imaging
US6710770B2 (en) * 2000-02-11 2004-03-23 Canesta, Inc. Quasi-three-dimensional method and apparatus to detect and localize interaction of user-object and virtual transfer device
US6813030B2 (en) * 2000-02-18 2004-11-02 Japan Science And Technology Corporation Optical interference tomographic image observing apparatus
US6665072B2 (en) * 2000-02-25 2003-12-16 Cambridge Research & Instrumentation Inc. Instantaneous dual band fluorescence detection systems
US6775567B2 (en) * 2000-02-25 2004-08-10 Xenogen Corporation Imaging apparatus
US6377353B1 (en) * 2000-03-07 2002-04-23 Pheno Imaging, Inc. Three-dimensional measuring system for animals using structured light
US6628401B2 (en) * 2000-03-10 2003-09-30 Fuji Photo Film Co., Ltd. Optical tomography imaging method and apparatus
US6429943B1 (en) * 2000-03-29 2002-08-06 Therma-Wave, Inc. Critical dimension analysis with simultaneous multiple angle of incidence measurements
US6618152B2 (en) * 2000-05-09 2003-09-09 Fuji Photo Film Co., Ltd. Optical coherence tomography apparatus using optical-waveguide structure which reduces pulse width of low-coherence light
US20040010192A1 (en) * 2000-06-15 2004-01-15 Spectros Corporation Optical imaging of induced signals in vivo under ambient light conditions
US6693710B1 (en) * 2000-06-16 2004-02-17 Cambridge Research & Instrumentation Inc. Polarization imaging system
US6381302B1 (en) * 2000-07-05 2002-04-30 Canon Kabushiki Kaisha Computer assisted 2D adjustment of stereo X-ray images
US6963375B1 (en) * 2000-07-07 2005-11-08 Axis, Ab Image altering device for an image producing apparatus
US6394965B1 (en) * 2000-08-15 2002-05-28 Carbon Medical Technologies, Inc. Tissue marking using biocompatible microparticles
US6597931B1 (en) * 2000-09-18 2003-07-22 Photonify Technologies, Inc. System and method for absolute oxygen saturation
US6636755B2 (en) * 2000-09-26 2003-10-21 Fuji Photo Film Co., Ltd. Method and apparatus for obtaining an optical tomographic image of a sentinel lymph node
US20040015062A1 (en) * 2000-11-27 2004-01-22 Vasilis Ntziachristos Fluorescence-mediated molecular tomography
US6615063B1 (en) * 2000-11-27 2003-09-02 The General Hospital Corporation Fluorescence-mediated molecular tomography
US20030002028A1 (en) * 2001-05-17 2003-01-02 Rice Bradley W. Method and apparatus for determining target depth, brightness and size within a body region
US7113217B2 (en) * 2001-07-13 2006-09-26 Xenogen Corporation Multi-view imaging apparatus
US6775349B2 (en) * 2001-10-23 2004-08-10 Washington Univ. In St. Louis System and method for scanning near-field optical tomography
US20030099329A1 (en) * 2001-10-23 2003-05-29 Schotland John Carl System and method for scanning near-field optical tomography
US6919919B2 (en) * 2002-02-06 2005-07-19 Xenogen Corporation Light calibration device for use in low level light imaging systems
US20060203244A1 (en) * 2002-02-22 2006-09-14 Xenogen Corporation Dual illumination system for an imaging apparatus and method
US6924893B2 (en) * 2002-05-13 2005-08-02 Marine Biological Laboratory Enhancing polarized light microscopy
US6628747B1 (en) * 2002-06-21 2003-09-30 Washington University In St. Louis System and method for dual-beam internal reflection tomography
US6618463B1 (en) * 2002-07-09 2003-09-09 Washington University System and method for single-beam internal reflection tomography
US20040021771A1 (en) * 2002-07-16 2004-02-05 Xenogen Corporation Method and apparatus for 3-D imaging of internal light sources
US20040027659A1 (en) * 2002-08-08 2004-02-12 Messerschmidt Robert G. Sample holder
US20040085536A1 (en) * 2002-11-01 2004-05-06 Schotland John Carl Tomography system and method using nonlinear reconstruction of scattered radiation
US20060173354A1 (en) * 2003-02-05 2006-08-03 Vasilis Ntziachristos Method and system for free space optical tomography of diffuse media
US7263157B2 (en) * 2004-06-24 2007-08-28 Siemens Aktiengesellschaft Imaging tomography apparatus with two acquisition systems, and method for determining their system angles
US20060118742A1 (en) * 2004-12-06 2006-06-08 Richard Levenson Systems and methods for in-vivo optical imaging and measurement
US20060146346A1 (en) * 2004-12-06 2006-07-06 Hoyt Clifford C Systems and methods for in-vivo optical imaging and measurement
US20060119865A1 (en) * 2004-12-06 2006-06-08 Hoyt Clifford C Systems and methods for in-vivo optical imaging and measurement
US20070016078A1 (en) * 2004-12-06 2007-01-18 Hoyt Clifford C Systems and methods for in-vivo optical imaging and measurement
US20060245631A1 (en) * 2005-01-27 2006-11-02 Richard Levenson Classifying image features
US20060268153A1 (en) * 2005-05-11 2006-11-30 Xenogen Corporation Surface contruction using combined photographic and structured light information

Cited By (8)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20090041449A1 (en) * 2005-12-12 2009-02-12 Topins Co., Ltd. The testing matter photographing apparatus attached three dimensional image photographing device
US7747163B2 (en) * 2005-12-12 2010-06-29 Topins Co., Ltd. Testing matter photographing apparatus attached three dimensional image photographing device
US20090066960A1 (en) * 2007-09-10 2009-03-12 Biospace Lab Luminescence Imagining Installation and Method
US7834989B2 (en) 2007-09-10 2010-11-16 Biospace Lab Luminescence imagining installation and method
US20110272597A1 (en) * 2009-01-30 2011-11-10 Fujifilm Corporation Method for generating optical tomographic information, optical tomographic information generating apparatus, and storage medium
US8742371B2 (en) * 2009-01-30 2014-06-03 Fujifilm Corporation Method for generating optical tomographic information, optical tomographic information generating apparatus, and storage medium
WO2010127302A1 (en) * 2009-05-01 2010-11-04 The Trustees Of Columbia University In The City Of New York Systems, methods and computer-accessible media for obtaining three-dimensional information from two-dimensional fluorescence emission data
US9456789B2 (en) * 2011-10-26 2016-10-04 Canon Kabushiki Kaisha Object information acquisition apparatus

Also Published As

Publication number Publication date
EP2146752B1 (en) 2022-05-18
US7555332B2 (en) 2009-06-30
US7599731B2 (en) 2009-10-06
WO2008124640A1 (en) 2008-10-16
EP2146752A4 (en) 2012-07-25
US20070253908A1 (en) 2007-11-01
EP2146752A1 (en) 2010-01-27

Similar Documents

Publication Publication Date Title
US7555332B2 (en) Fluorescent light tomography
US11730370B2 (en) Spectral unmixing for in-vivo imaging
US9080977B2 (en) Apparatus and methods for fluorescence guided surgery
US7797034B2 (en) 3-D in-vivo imaging and topography using structured light
US8180435B2 (en) Method and apparatus for determining target depth, brightness and size within a body region
US10130318B2 (en) Integrated microtomography and optical imaging systems
US20110098575A1 (en) Method and an arrangement for the determination of the optical properties of a multi-layered tissue
US20090240138A1 (en) Diffuse Optical Tomography System and Method of Use
AU2002303819A1 (en) Method and apparatus for determining target depth, brightness and size within a body region
US10775308B2 (en) Apparatus and methods for determining optical tissue properties
EP2068714A2 (en) Spectral unmixing for in-vivo imaging
EP1707944A2 (en) Method and apparatus for determining target depth, brightness and size within a body region

Legal Events

Date Code Title Description
AS Assignment

Owner name: XENOGEN CORPORATION, CALIFORNIA

Free format text: ASSIGNMENT OF ASSIGNORS INTEREST;ASSIGNORS:RICE, BRADLEY W.;KUO, CHAINCY;STEARNS, DANIEL G.;AND OTHERS;REEL/FRAME:020017/0517;SIGNING DATES FROM 20070927 TO 20071001

STCF Information on status: patent grant

Free format text: PATENTED CASE

CC Certificate of correction
FPAY Fee payment

Year of fee payment: 4

FPAY Fee payment

Year of fee payment: 8

MAFP Maintenance fee payment

Free format text: PAYMENT OF MAINTENANCE FEE, 12TH YEAR, LARGE ENTITY (ORIGINAL EVENT CODE: M1553); ENTITY STATUS OF PATENT OWNER: LARGE ENTITY

Year of fee payment: 12