The system uses a technique called fluorescence lifetime imaging, which has applications in DNA sequencing and cancer diagnosis, among other things. So the new work could have implications for both biological research and clinical practice.
“The theme of our work is to take the electronic and optical precision of this big expensive microscope and replace it with sophistication in mathematical modeling,” says Ayush Bhandari, a graduate student at the MIT Media Lab and one of the system’s developers. “We show that you can use something in consumer imaging, like the Microsoft Kinect, to do bioimaging in much the same way that the microscope is doing.”
MIT researchers have developed a new biomedical imaging system that harnesses an off-the-shelf depth sensor such as Microsoft’s Kinect. The coloration of these images depicts the phase information contained in six of the 50 light frequencies the system analyzes.
Fluorescence lifetime imaging, as its name implies, depends on fluorescence, or the tendency of materials known as fluorophores to absorb light and then re-emit it a short time later. For a given fluorophore, interactions with other chemicals will shorten the interval between the absorption and emission of light in a predictable way. Measuring that interval — the “lifetime” of the fluorescence — in a biological sample treated with a fluorescent dye can reveal information about the sample’s chemical composition.
In traditional fluorescence lifetime imaging, the imaging system emits a burst of light, much of which is absorbed by the sample, and then measures how long it takes for returning light particles, or photons, to strike an array of detectors. To make the measurement as precise as possible, the light bursts are extremely short.
The fluorescence lifetimes pertinent to biomedical imaging are in the nanosecond range. So traditional fluorescence lifetime imaging uses light bursts that last just picoseconds, or thousandths of nanoseconds.
Off-the-shelf depth sensors like the Kinect, however, use light bursts that last tens of nanoseconds. That’s fine for their intended purpose: gauging objects’ depth by measuring the time it takes light to reflect off of them and return to the sensor. But it would appear to be too coarse-grained for fluorescence lifetime imaging.
The Media Lab researchers, however, extract additional information from the light signal by subjecting it to a Fourier transform. The Fourier transform is a technique for breaking signals — optical, electrical, or acoustical — into their constituent frequencies. A given signal, no matter how irregular, can be represented as the weighted sum of signals at many different frequencies, each of them perfectly regular.
The Media Lab researchers represent the optical signal returning from the sample as the sum of 50 different frequencies. Some of those frequencies are higher than that of the signal itself, which is how they are able to recover information about fluorescence lifetimes shorter than the duration of the emitted burst of light.
For each of those 50 frequencies, the researchers measure the difference in phase between the emitted signal and the returning signal. If an electromagnetic wave can be thought of as a regular up-and-down squiggle, phase is the degree of alignment between the troughs and crests of one wave and those of another. In fluorescence imaging, phase shift also carries information about the fluorescence lifetime.
Not all of the light that strikes the biological sample is absorbed; some of it is reflected back. The MIT researchers’ system takes the measurements of incoming light and fits them to a mathematical model of the overlapping intensity profiles of both reflected and re-emitted light.
Once it’s deduced the intensity profile of the reflected light, it can calculate the distance between the emitter and the sample. So unlike conventional fluorescence lifetime imaging, the researchers’ approach doesn’t require distance calibration.
According to Bhandari, some of his colleagues were skeptical that the returning light signal contained enough information to produce accurate models of the intensity profiles. “They were not convinced that the precision of Kinect-like sensors is enough,” he says.
“But lifetime and distance are two numbers. If you have two numbers, then 50 measurements is a lot. The desired information is two points, but the measurement is 50 points, so you have a ratio of one to 25. It’s enough to give you the intuition that it should be workable.”
The depth sensors that the researchers used in their experiments — the Kinect and others — had arrays of roughly 20,000 light detectors each, and the most accurate results came when the detector was 2.5 meters away from the biological sample. That setup doesn’t afford the image resolution that existing fluorescence lifetime imaging microscopes do. But while denser arrays of detectors and optics that better control the emission and gathering of light would inflate the cost of the researchers’ system beyond the $100 that a Kinect costs, it still shouldn’t be nearly as expensive as current fluorescence lifetime imaging systems.
“If I had one of these devices, what I would do is just go looking around the world at stuff,” says Adam Cohen, a professor of chemistry and chemical biology and of physics at Harvard University. “There might be all sorts of interesting things. It’s a new modality for looking around the world. It’s just like, humans can’t see polarization in the light — we only see color — and there’s all this structure that insects can see because they can see polarization that we’re just blind to. To my knowledge, there’s no living organism that can see excited-state lifetime, and so there’s this contrast modality that has not been explored by life that nonetheless might be very interesting. So I’d look at it.”
Fluorescence lifetime imaging (FLI) is a popular method for extracting useful information that is otherwise unavailable
from a conventional intensity image. Usually, however, it requires expensive equipment, is often limited to either
distinctly frequency- or time-domain modalities, and demands calibration measurements and precise knowledge of the
illumination signal. Here, we present a generalized time-based, cost-effective method for estimating lifetimes by repurposing
a consumer-grade time-of-flight sensor. By developing mathematical theory that unifies time- and frequencydomain
approaches, we can interpret a time-based signal as a combination of multiple frequency measurements. We
show that we can estimate lifetimes without knowledge of the illumination signal and without any calibration. We
experimentally demonstrate this blind, reference-free method using a quantum dot solution and discuss the method’s
implementation in FLI applications.
SOURCES – MIT News, Optica