G01N21/314—Investigating relative effect of material at wavelengths characteristic of specific elements or molecules, e.g. atomic absorption spectrometry with comparison of measurements at specific and non-specific wavelengths

G01N2021/3144—Investigating relative effect of material at wavelengths characteristic of specific elements or molecules, e.g. atomic absorption spectrometry with comparison of measurements at specific and non-specific wavelengths for oxymetry

G—PHYSICS

G01—MEASURING; TESTING

G01N—INVESTIGATING OR ANALYSING MATERIALS BY DETERMINING THEIR CHEMICAL OR PHYSICAL PROPERTIES

G01N21/00—Investigating or analysing materials by the use of optical means, i.e. using infra-red, visible or ultra-violet light

G01N21/17—Systems in which incident light is modified in accordance with the properties of the material investigated

G01N21/25—Colour; Spectral properties, i.e. comparison of effect of material on the light at two or more different wavelengths or wavelength bands

G01N21/314—Investigating relative effect of material at wavelengths characteristic of specific elements or molecules, e.g. atomic absorption spectrometry with comparison of measurements at specific and non-specific wavelengths

G01N2021/3155—Measuring in two spectral ranges, e.g. UV and visible

Abstract

A portable multi-spectral imaging system and device is disclosed. The system includes at least one image acquisition device for acquiring an image from a subject, a filtering device to filter the light received by the image acquisition device, a processor for processing the image acquired by the image acquisition device, and a display. There is software running on the processor that determines oxgynation values of the subject based on the processed image.

The invention is directed to multi-spectral imaging. Specifically, the invention is directed to a portable multi-spectral imager.

2. Background of the Invention Spectroscopy, whether it is visible, near infrared, infrared or Raman, is an enormously powerful tool for the analysis of biomedical samples. The medical community, however, has a definite preference for imaging methods, as exemplified by methods such as MRI and CT scanning as well as standard X-ray photography and ultrasound imaging. This is entirely understandable as many factors need to be taken into account for a physician to make a clinical diagnosis. Imaging methods potentially can provide far more information to a physician than their non-imaging counterparts. With this medical reality in mind, there has been considerable effort put into combining the power and versatility of imaging method with the specificity of spectroscopic methods. Near-infrared (near-IR) spectroscopy and spectroscopic imaging can measure the balance between oxygen delivery and tissue oxygen utilization by monitoring the hemoglobin oxygen saturation in tissues (Sowa, M. G. et ah, 1998, Proc. SPIE 3252, pp. 199-207; Sowa, G. W. et al,, 1999, Journal of Surgical Research. 86:62-29; Sow, G. W. et al., 1999, Journal of Biomedical Optics, 4:474-481; Mansfield, J. R., et al., 2000. International Society of Optical Engineers, 3920:99-197). For in-vivo human studies, the forearm or leg has been the investigational site for many of the noninvasive near-IR studies. Non-imaging near-IR applications have examined the local response of tissue to manipulations of blood flow (De-Blasi. R. A. et al., 1992, Adv. Exp. Med. Biol. 317:771-
777). Clinically, there are situations where the regional variations in oxygenation saturation are of interest (Stranc, M. F, et al, 1998, British Journal of Plastic Surgery, 51 :210-218). Near-IR imaging offers a means of accessing the spatial heterogeneity of the hemoglobin oxygenation saturation response to tissue perfusion. (Mansfield, J. R. et al., 1997, Analytical Chemistry, 69:3370-3374; Mansfield, J. R.. et al., 1997,

Spectroscopic imaging methodologies and data are becoming increasingly common in analytical laboratories, whether it be magnetic resonance (MRI), mid-IR, Raman, fluorescence and optical microscopy, or near-lR/visible-based imaging. However, the volume of information contained in spectroscopic images can make standard data processing techniques cumbersome. Furthermore, there are few techniques
that can demarcate which regions of a spectroscopic image contain similar spectra without a priori knowledge of either the spectral data or the sample's composition. The objective of analyzing spectroscopic images is not only to determine what the spectrum is at any particular pixel in the sample, but also to determine which regions of the sample contain similar spectra; i.e., what regions of the sample contain chemically related compounds. Multivariate analysis methodologies can be used to determine both the spectral and spatial characteristics of a sample within a spectroscopic imaging data set. These techniques can also be used to analyze variations in the temporal shape of a time series of images either derived for extracted from a time series of spectroscopic images. There are few techniques that can demarcate which regions of a sample contain similar substances without a priori knowledge of the sample's composition. Spectroscopic imaging provides the specificity of spectroscopy while at the same time relaying spatial information by providing images of the sample that convey some chemical meaning. Usually the objective in analyzing heterogeneous systems is to identify not only the components present in the system, but their spatial distribution. The true power of this technique relative to traditional imaging methods lies in its inherent multivariate nature. Spatial relationships among many parameters can be assessed simultaneously. Thus, the chemical heterogeneity or regional similarity within a sample is captured in a high dimensional representation which can be projected onto a number of meaningful low dimensional easily interpretable representations which typically comprise a set of composite images each having a specific meaning.

While it is now clear that both spectroscopy and spectroscopic imaging can play roles in providing medically relevant information, the raw spectral or imaging measurement seldom reveals directly the property of clinical interest. For example using spectroscopy, one cannot easily determine whether the tissue is cancerous, or determine blood glucose concentrations and the adequacy of tissue perfusion. Instead, pattern recognition algorithms, clustering methods, regression and other theoretical methods provide the means to distill diagnostic information from the original analytical measurements. There are however various methods for the collection of spectroscopic images. In all such cases, the result of a spectroscopic imaging experiment is something termed a
spectral image cube, spectroscopic imaging data cube or just hypercube. This is a three dimensional array of data, consisting of two spatial dimensions (the imaging component), and one spectral dimension, It can be thought of as an array of spatially resolved individual spectra, with every pixel in the first image consisting of an entire spectrum, or as a series of spectrally resolved images. In either representation, the 3D data cube can be treated as a single entity containing enormous amounts of spatial and spectral information about the sample from which it was acquired.

As an extension of the three dimensional array acquired in a spectroscopic imaging experiment, one can collect data cubes as a function of additional parameters such as time, temperature or pH. Numerous algorithms can be used to analyze these multi-dimensional data sets so that chemical and spectral variations can be studied as additional parameters. However, taken together, they can allow one to more fully understand the variations in the data. This can be done in a gated or sequential fashion. Multi -modal image fusion, or image registration, is an important problem frequently addressed in medical image analysis. Registration is the process of aligning data that arise from different sources into one consistent coordinate frame. For example, various tissues appear more clearly in different types of imaging methods. Soft tissue, for example, is imaged well in MR scans, while bone is more easily discernible in CT scans. Blood vessels are often highlighted better in an MR angiogram than in a standard IvIR scan. Multiple scans of the same patient will generally be unregistered when acquired, as the patient may be in different positions in each scanner, and each scanner has its own coordinate system. In order to fuse the information from all scans into one coherent frame, the scans must be registered. The very reason why multiple scans are useful is what makes the registration process difficult. As each modality images tissue differently and has its own artifacts and noise characteristics, accurately modeling the intensity relationship between the scans, and subsequently aligning them, is difficult.

The registration of two images consists of finding the transformation that best maps one image into the other. IfI1 and I2 are two images of the same tissue and T is the correct transformation, then the voxel Ij (x) corresponds to the same position in the sample as the voxel I2 (T(x)). In the simplest case. T is a rigid transformation consisting of three degrees of freedom of rotation and three degrees of freedom of translation. The
need for rigid registration arises primarily from the patient being in different positions in the scanning devices used to image the anatomy. The information from all the images is best used when presented in one unified coordinate system. Without such image fusion, the clinician must mentally relate the information from the disparate coordinate frames. One method of aligning the two images is to define an intermediate, patient- centered coordinate system, instead of trying to directly register the images to one another. An example of a patient-centered reference frame is the use of fiducial markers attached to a patient throughout the various image acquisitions. The fiducial markers define a coordinate system specific to the patient, independent of the scanner or choice of imaging modality. If the markers remain fixed and can be accurately localized in all the images, then the volumes can be registered by computing the best alignment of the corresponding fiducials (Horn. B. K. P.. 1987, Journal of the Optical Society of America A, 4:629-642; Mandava, V. R., et al., Proc SPIE. 1992, 1652:271-282; Haralick, R. M., et at, 1993, Computer and Robot Vision). The main drawback of this method is that the markers must remain attached to the patient at the same positions throughout all image acquisitions. For applications such as change detection over months or years, this registration method is not suitable. Fiducial registration is typically used as ground-truth to evaluate the accuracy of other methods as careful placement and localization of the markers can provide very accurate alignment (West, J. et al, 1996, Proc SPlE. Newport Beach, Calif.).

When fiducial markers are not available to define the patient coordinate frame, corresponding anatomical feature points can be extracted from the images and used to compute the best alignment (Maintz, J. B. Antione, et al., 1995 Computer Vision, Virtual Reality and Robotics in Medicine, pp. 219-228; Maguire. Jr., G., et al., 1991, IEEE Computer Graphics Applications, 11 :20-29). This approach depends greatly on the ability to automatically and accurately extract reliable image features. In general, methods of feature extraction such as intensity thresholding or edge detection do not work well on medical scans, due to non-linear gain fields and highly textured structures. Even manual identification of corresponding 3D anatomical points can be uireliable. Without the ability to accurately locali?e corresponding features in the images, alignment in this manner is difficult.
Instead of localizing feature points in the images, richer structures such as object surfaces can be extracted and used as a basis of registration. A common method of registering MR and CT of the head involves extracting the skin (or skull) surfaces from both images, and aligning the 3D head models (Jiang, H., et aL 1992 Proc. SPIE, 1808:196-213; Lemoine, D. et aL, 1994, Proc. SPIE, 2164:46-56), For PET/MR registration, the brain surface is typically used since the skull is not clearly visible in PET (Pelizzari, C, et aL, J Comput Assist. Toraogr., 1989, 13:20-26). The 3D models are then rigidly registered using surface-based registration techniques (Ettinger, G., 1997, MIT Ph.D Thesis). The success of such methods relies on the structures being accurately and consistently segmented across modalities and the surfaces having rich enough structure to be unambiguously registered.

Voxel-based approaches to registration do not extract any features from the images, but use the intensities themselves to register the two images. Such approaches model the relationships between intensities of the two images when they are registered, and then search through the transformation space to find an alignment that best agrees with the model. Various intensity models are discussed, including correlation, mutual information, and joint intensity priors.

Correlation is a measure commonly used to compare two images or regions of images for computer vision problems such as alignment or matching. Given the intensity values of two image patches stacked in the vectors u and v, the normalized correlation measure is the dot product of unit vectors in the directions of u and v:

(wvV(|[u||v||)

An advantage of correlation-based methods is that they can be computed quite efficiently using convolution operators. Correlation is applicable when one expects a linear relationship between the intensities in the two images. In computer vision problems, normalized correlation provides some amount of robustness to lighting variation over a measure such as sum of square differences (SSD), ||u-v||2. The primary reason for acquiring more than one medical scan of a patient stems from the fact that each scan provides different information to the clinician. Therefore, two images that have a simple
linear intensity relationship may be straightforward to register, but do not provide any additional information than one scan by itself. On the other hand, if the images are completely independent (e.g. no intensity relationship exists between them), then they cannot be registered using voxel -based methods. In general, there is some dependence between images of different modalities and each modality does provide additional information.

One simplified model of the medical imaging process is that an internal image is a rendering function R of underlying tissue properties, P(x), over positions x. An image of modality A could be represented as a function RA (P) and a registered image of modality B of the same patient would be another function, say RB (P). Suppose a function F(x) could be computed relating the two rendering functions such that the following is true (with the possible addition of some Gaussian noise, N):

The function F would predict the intensity at a point in Image A given the intensity at the corresponding point in Image B. Such a function could be used to align a pair of images that are initially in different coordinate systems using SSD:

T*=A- tgnύnrΣ.k{F(Rs(P(X} 5t>~RΛ(P(xi\) '

where T is the transformation between the two sets of image coordinates. Van den Elsen et al. compute such a mapping that makes a CT image appear more like an MR, and then register the images using correlation (van den Elsen, P., et al., 1994. "Visualization in Biomedical Computing," 1994 Proc SPIE, 2359:227-237). In general, explicitly computing the function F that relates two imaging modalities is difficult and under- constrained.

Maximization of mutual information (MI) is a general approach applicable to a wide range of multi-modality registration applications (Bell, A. J., et al., 1995 Advances in Neural Information Processing 7; Collignon, D., et al., 1995, First Conf. on Computer Vision, Virtual Reality and Robotics in Medicine Springer; Maes, F. et al, 1996,
Mathematical Methods in Biomedical Image Analysis; Wells, W. M., et al, 1996, Medical Image Analysis, I(l ):35-51 ). One of the strengths of using mutual information is that MI does not use any prior information about the relationship between joint intensity distributions. While mutual information does not explicitly model the function F that relates the two imaging modalities, it assumes that when the images are aligned, each image should explain the other better than when the images are not aligned.

Given two random variables U and V, mutual information is defined as (Bell, 1995):

MI(U, ¥\=H(ir}+H\V}~H(U V)

where H(U) and H(V) arc the entropies of the two variables, and H(U1V) is the joint entropy. The entropy of a discrete random variable is defined as:

H(Uy-SP^n) \o& Pu{u)

where P.&ub.u (u) is the probability mass function associated with U. Similarly, the expression for joint entropy entropy operates over the joint PDF:

H(U,V)^ΣΣPiU (u,v\ log Fll v(utv}

When U and V are independent, H(U.V)=H(U)+H(V), which implies the mutual information is 7ero. When there is a one-to-one functional relationship between U and V, (i.e. they are completely dependent), the mutual information is maximized as:

To operate on images over a transformation, we consider the two images, I.sub.l (x) and
I.sub.2 (x) to be random variables under a spatial parameterization, x. We seek to find the value of the transformation T that maximizes the mutual information (Wells, 1996):

'P-

The entropies of the two images encourage transformations that project Il onto complex parts of 12. The third term, the (negative) joint entropy of Ij and I2, takes on large values when X explains Y well. Derivatives of the entropies with respect to the pose parameters can be calculated and used to perform stochastic gradient ascent (Wells, 1996). West et al. compare many multi-modal registration techniques and find mutual information to be one of the most accurate across all pairs of modalities (West 1996).

Leventon et al. introduced an approach to multi-modal registration using statistical models derived from a training set of images (Leventon. M., et al., 1998, Medical Image Computing and Computer-assisted Intervention). The method involved building a prior model of the intensity relationship between the two scans being registered. The method requires a pair of registered training images of the same modalities as those to be registered in order to build the joint intensity model. To align a novel pair of images, the likelihood of the two images given a certain pose based on our model by sampling the intensities at corresponding points is computed. This current hypothesis can be improved by ascending the log likelihood function. In essence, one computes a probabilistic estimate of the function F (that relates the two imaging modalities) based on intensity co-occurrence. To align the novel images, the pose is found that maximizes the likelihood that those images arose from the same relation F.

Building a joint-intensity model does require having access to a registered pair of images of the same modality and approximately the same coverage as the novel pair to be registered. Mutual information approaches do not need to draw upon previously registered scans. However, when this information is available, the prior joint intensity model provides the registration algorithm with additional guidance which results in
convergence on the correct alignment more quickly, more reliably and from more remote initial starting points.

Summary of the Invention The present invention overcomes the problems and disadvantages associated with current designs of existing hyperspectral imaging devices. In particular, the device provides for real-time measurements of patients oxygenation levels and other components which are specific to a disease condition. The present invention expands on the invention disclosed in U.S. Patent No, 6,640,130, herein incorporated by reference in its entirety. Furthermore, it is an improvement over the existing device in that this is done in real-time.

One embodiment of the invention is directed to a portable multi-spectral imaging system. The system includes at least one image acquisition device for acquiring an image from a subject, a filtering device to filter the light received by the image acquisition device, a processor for processing the image acquired by the image acquisition device, and a display. There is software running on the processor that determines oxygenation values or other relevant components of the subject based on the processed image. The oxygenation values may be based on at least one of oxyhemoglobin, deoxyhemoglobin and oxygen saturation levels or on other relevant components. The software may filter the image to reduce noise, correct at least one of the images to account for motion of the subject, eliminates extraneous objects from the acquired image, and/or compare the data of the acquired image to stored data.

The system may also include an illumination source and may communicate, through wireless or wired channels, with an analysis device. The analysis device may compare multiple images, store acquired images, and/or store acquired oxgynation values.

Another embodiment of the invention is directed to a portable multi-spectral imaging device. The device includes at least one image acquisition device for acquiring a multi-spectral image from a subject, a filtering device to filter the light received by the image acquisition device, an analogue front end module to convert the image to a digital image, a microprocessor to temporarily store at least one image and control the analogue
front end module, and. a communications module for communicating the acquired image to an analysis device.

The image may be used to determine oxygenation values of the subject. The oxygenation values may be based on at least one of oxyhemoglobin, de-oxyhemoglobin and/or oxygen saturation levels.

The device may also include an illumination source that may produce filtered light. The communication may either be wired or wirelessly. The device may have a maximum diameter of two inches, and/or be handheld. The device may further include a power source. Other embodiments and advantages of the invention are set forth in part in the description, which follows, and in part, may be obvious from this description, or may be learned from the practice of the invention.

Description of the Drawings The invention is described in greater detail by way of example only and with reference to the attached drawings, in which:

Figure 1 is a 3-D view of an embodiment of a multi-spectra imaging device.

Figure 2 is an exploded view of Figure 1.

Figure 3 is a schematic block diagram of the multi-spectral imaging device. Figures 4a-4 show a second embodiment of a multi-spectra imaging device.

Description of the Invention

As embodied and broadly described herein, the disclosures herein provide detailed embodiments of the invention. However, the disclosed embodiments are merely exemplary of the invention that may be embodied in various and alternative forms. Therefore, there is no intent that specific structural and functional details should be limiting, but rather the intention is that they provide a basis for the claims and as a representative basis for teaching one skilled in the art to variously employ the present invention. A problem in the art capable of being solved by the embodiments of the present invention is producing a miniaturized medical multi-spectral imaging (MSIj sensor
capable of providing real-time measurements of oxygen saturation StO? in skin, serving as an excellent indicator of oxygenation status in patients with multiple medical conditions, including but not limited to diabetes, wound care, vascular disease and pressure ulcers as well as providing an early warning for the onset of shock. It has been surprisingly discovered that a highly sensitive miniature sensor may be capable of precise early measurements of oxygen levels in the skin of patients at risk of diabetes, wound care, vascular disease, pressure ulcers or other disease states., The device may be optimized to support field-deployable, portable operational scenarios for remote and extreme environments, both military and civilian. Furthermore, by reducing costs, the hand-held device may be used by patients at risk for developing diabetic foot ulcers with the aim of warning for early tissue breakdown, thereby preventing ulceration.

The miniature multi-spectral imaging device is likely to have superior qualities in comparison to other currently existing point measuring near-infrared spectroscopy (NtR) devices. This is due to the fact that the device predominantly measures SO2 in the skin capillary bed. Therefore, skin thickness and fat layers have less affect on the returned signal whereas NIR based technologies need to actively correct for these factors.

The device's sensor has cross polarized illumination detection which is less sensitive to surface glare and superficial scattering. Furthermore, it generates a more advanced hemoglobin decomposition algorithm. Crosstalk between de-oxyhemoglobin and other background terms was noted in previous algorithms leading to high variability in the de-oxyhemoglobin signal and in the oxygen saturation value determined from it.

It is possible to correlate the spectrum of each pixel with the presence and concentration of various chemical species. This data can then be interpreted as a "gradient map" of these species in a surface. HSI for medical applications (MHSl) has been shown to accurately predict viability and survival of tissue deprived of adequate perfusion, and to differentiate diseased tissue (e.g. tumor) and ischemic tissue from normal tissue. MHSI analysis uses spatial and spectral characteristics obtained from the skin to develop indices for shock prediction based primarily on oxy- and deoxy-Hb signals including (1 ) average index (mean across image), (2) heterogeneity index (inter- quartile range), (3) mottling index (analysis of spatial features) and (4) temporal shift index (change in mottling pattern from one image to the next). Using a combination of
these techniques, a hyperspectral index (HSI) as a simple numerical reading has been developed. HSI may serve as an early indicator of the many disease states such as vascular disease, diabetes, pressure ulcers and shock..

The handheld multi-spectral imaging system is designed for tissue optical imaging for non-contact evaluation of tissue oxygenation, over an extended area, without harmful radiation, and without the need of using any agent into the tissue. Due to its compact design with a fast image sensor, high-efficient illuminator and high-speed wavelength selection, it may be entirely portable and self-contained to acquire image data and provide hyperspectral information of tissue oxygenation to the user. The system includes an imaging acquisition parameter dynamics module, This module is responsible for on-the-fly adjustment of image sensor parameters in order to ensure that SNR (Signal-to-Noise) requirements are met for a given target (i.e. skin type, surface type, etc.). In addition, accommodations are made to handle various lighting conditions present during acquisition phase. The module is responsible for choosing appropriate wavelengths and controlling high-speed wavelength selector. High-efficiency illumination may also be controlled and is adjusted in real-time based on surface reflectance characteristics used as feedback at a given wavelength of light.

It is important to enforce acquisition parameters such that a general imaging problem is constrained. A fiducial metrics module is responsible for locating fiducial marks and executing the device's spatial positioning/stabilization logic relative to surface to be imaged in real-time. Thus, optimizing illumination delivery as well as positioning repeatability.

In parallel with other tasks, an acquisition subsystem continuously monitors quality of the data with respect to achieved SNR. Since sought imaging modality is unconstrained relative to patient's motion (by design, to deliver robust usage model) motion tracking is also performed in real-time to ensure that data is not corrupted during imaging.

Overall, the acquisition platform is designed to deliver accurate real-time performance due to custom hardware (truly parallelized microprocessors. FPGAs)/software implementation (embedded) that incorporates observed feedback with a priori knowledge about imaging modality to produce unparalleled performance and data
quality. The major image processing software is designed and developed to achieve adaptive filtering, fast and accurate image registration, fast and effective tissue/obstruction masking and high performance algorithm for tissue oxygenation values such as oxyhemoglobin, deoxyhemoglobin and oxygen saturation. The software includes adaptive spatial filtering, which will take place in order to ensure that noise is minimized without compromising informative data. The software further includes spatial registration, which is employed to correct for target motion during acquisition phase. Registration is robust as to handle translational as well as rotational motion components. This ensures proper spectral composition. It is advantageous to be able to discern useful spectral data from extraneous data in the field of view. Methods for classifying useful data from all other data have been developed based on both spatial as well as spectral features. Such methods allow for higher accuracy and increased performance by eliminating from acquired dataset extraneous objects like hair, non-skin material (bandages, clothing, etc.), or any other objects that do not carry useful information about tissue hemoglobin content.

The system may extract tissue oxygenation values for oxyhemoglobin, deoxyhemoglobin and oxygen saturation, using the data acquired from a set of predetermined wavelengths via one or more spectral classification methods- Apart from making classification decisions based only on acquired data, a priori knowledge integration will take place effectively fusing an ensemble of classifiers to boost resulting clinical efficacy. Such data is extracted from clinical studies engaging diverse patient populations. Extracting robust "historical" (information across multiple visits) patient information and effectively presenting it for doctor's use can be achieved using well- established statistical techniques such as PCA (Principal Components Analysis), ICA (Independent Components Analysis) and LDA (partial least square) to isolate the most informative features therefore to improve the robustness and measurement accuracy of the system.

Figure 1 is a 3-D view of the multi-spectra imaging device 100. Preferably, device 100 is a handheld device having a maximum diameter of less than 20 inches. More preferably the maximum diameter is less than 10 inches and even more preferably the diameter is less than 5 inches. Device 100 is preferably self contained, and can be
mounted on a tripod or arm extending from the wall of a hospital or clinical setting. Device 100 may be in wired or wireless communication with a user. Device 100 may have an internal or external power source.

Figure 2 is an exploded view of device 100, The device may include a lens polarizer assembly 105, a lens 110, an illumination polarizer assembly 115, an illumination module 120, an image sensor 125, and a lens mount 130. Device 100 may optionally include a power input 135 and/or a wired communications interface 140.

The handheld multi-spectral imaging system, is designed for tissue optical imaging for non-contact evaluation of tissue oxygenation, over an extended area, without harmful radiation, and without the need of using any agent into the tissue. The compact design with fast image sensor, high efficient illuminator and high-speed wavelength selection allows the device to be entirely portable and self-contained to acquire image data and provide hyperspectral information of tissue oxygenation to the user. The device may employ fast image acquisition and data processing with on-line processing using a combination of parallel processing circuits achievable with Field-Programmable Gate Arrays (FPGAs). The device may be capable of on-line display of the hyperspectral image on the handheld device via simple and intuitive graphic user interface for rapid view and sending image data to a remote computer for further processing and manipulation. Custom electronics and software allow for fast signal/imaging processing, analysis and display, and sending image data through wired and/or wireless connection for storage and may extract tissue oxygenation values for oxyhemoglobin, deoxyhemoglobin and oxygen saturation, using the data from a set of predetermined wavelengths use one or more tissue classification methods.

Several tissue masks may be applied to improve system performance by just focusing on the tissue of interest. Some tissue masks are spectral based and others are spatial based.

Since the device uses noncoπtact measurements, hemoglobin oxygenation status is not affected by how much pressures is placed on skin as with NIR probes. Measurement can also be made at a reasonably remote distance and through optical face shields if necessary. There is no need to disinfect system between patients as would be required for NIR systems.
The multi-spectral system uses visible wavelengths rather than NlR wavelengths which are more effectively absorbed by hemoglobin, In addition, because the photon pathlength is more superficial (~2mm), the multi-spectral imager predominantly measures hemoglobin in skin capillaries. As a result skin and fat layer thicknesses have less influence on the optical signals.

The multi -spectral imaging system captures hemoglobin oxygen saturation measurements over a reasonably wide field of view enabling the spatial variation to be measured. For example, subclinical skin mottling prior to the onset of shock, diabetic foot ulcers, claudication or other disease states can be measured with the spectral imager. Figure 3 is a schematic block diagram of a multi-spectral imager. The image sensor is used for gathering multi -spectral data. While the Analog Front End (AFE) is a chipset that handles all functions related to conversion of analog signal data from image sensor to digital images. The digital image is then sent to the microprocessor, which is used to control all AFE parameters (gain, DC offset, brightness, exposure, etc.), temporarily store one or more images (RAM/Flash), interface with other modules via GPIO (illumination module), and interface with FGPA for image acquisition control and image dumping.

The Field Programmable Gate Array (FPGA) is used to control high level image acquisition and frame transfer (from RAM/Flash), interface with other modules via GPIO, processing of hypercube data, algorithm implementation, and interface with integrated wired and wireless communications modules. The illumination module is used to emit light of specified wavelengths during the image acquisition cycle.

Images collected at each wavelength may be spatially filtered to improve signal to noise ratios. The images may also be shifted accordingly so the pixels in each image represent the same site of the object plane. Spectral and spatial algorithms may be used to mask anything in the object plane that does not resemble tissue (e.g. the patients' clothing, hat or any other accessory, hair, dirt or grim, etc.). Oxyhemoglobin, deoxyhemoglobin and oxygen saturation may be extracted from data at each pixel identified as representing tissue using standard hemoglobin decomposition algorithms. Figure 4a is a front view of another embodiment of the multi-spectral imager 600, while Figure 4b is a rear view of the embodiment. The multi-spectral imager 600 may be
fitted with a disposable illumination cartridge. M ulti -spectral imager 600 has a circuit board including at least one image sensor 605. Preferably there are between two and twenty image sensors. Sensors 605 may sense any wavelength including visible, color, near infrared, far infrared, or any combination thereof. Imager 600 may also include a lens for each image sensor 605. Between each lens and each image sensor 605 may be a filter. The filters may be set to filter out specific wavelengths. The filters may be optimized to produce the best detection. Additionally imager 600 may include an illumination source 610. The image sensors 605 and the illumination source 610 may all be located on the same circuit board. The circuit board may further include at least one Field-Programmable Gate Array (FPGA).

Imager 600 may include a display 615 to display a captured image. Display 615 may be a touch screen so that information can be entered through display 615 into imager 600. Display 615 may be of any size. Imager 600 may Interface with an analysis device via an Ethernet connection, USB connection or wirelessly. Analysis device may be used for image to image comparisons, storage, and review of images. Imager 600 may be made of any material, including but not limited to, plastic and metal.

Other embodiments and uses of the invention will be apparent to those skilled in the art from consideration of the specification and practice of the invention disclosed herein. All references cited herein, including all publications, U.S. and foreign patents and patent applications, are specifically and entirely incorporated by reference. It is intended that the specification and examples be considered exemplary only with the true scope and spirit of the invention indicated by the following claims. Furthermore, the term "comprising of includes the terms "'consisting of and "consisting essentially of.''

Claims

Claims

1. A portable multi -spectral imaging system, comprising: at least one image acquisition device for acquiring an image from a subject; a filtering device to filter the light received by the image acquisition device; a processor for processing the image acquired by the image acquisition device; wherein there is software executing on the processor that determines oxgyπation values of the subject based on the processed image; and a display for displaying at least one of the processed image and the oxgynation values.

2. The system of claim 1, wherein the oxygenation values are based on at least one of oxyhemoglobin, deoxyhemoglobin, oxygen saturation levels, and component analysis,

3. The system of claim 1, wherein the software filters the image to reduce noise.

4. The system of claim 1, wherein multiple images are acquired of a subject.

5. The system of claim 4, wherein the software corrects at least one of the images to account for motion of the subject.

6. The system of claim L wherein the software eliminates extraneous objects from the acquired image.

7. The system of claim 1 , wherein the software compares the data of the acquired image to stored data.

8. The system of claim 1 , further comprising an illumination source.

9. The system of claim L wherein the system communicates with an analysis device.

10. The system of claim 9, wherein the communication is one of wired and wirelessly.

11. The system of claim 9, wherein the analysis device at least one of compares multiple images, stores acquired images, and stores acquired oxgynation values.

12. A portable multi-spectral imaging device, comprising: at least one image acquisition device for acquiring a multi-spectral image from a subject; a filtering device to filter the light received by the image acquisition device; a analogue front end module to convert the image to a digital image; a microprocessor to temporarily store at least one image and control the analog front end module; and a communications module for communicating the acquired image to an analysis device.

13. The device of claim 12, wherein the image is used to determine oxgynation values of the subject.

14. The device of claim 13, wherein the oxygenation values are based on at least one of oxyhemoglobin, deoxyhemoglobin and oxygen saturation levels.

15. The device of claim 13, wherein the oxgynation values are based on component analysis.

16. The device of claim 12. further comprising an illumination source.