ECE345, Visual-to-Audio Electronic Travel Aid
Code for TM320C54x (v2a.asm) download
This project involves the design and implementation of a audio synthesis device that converts moving IMAGEs into audio signals. The system is built on a TM320C54x DSP with interface to an IMAQ camera module via the serial port on a PC. Brief description: A LabVIEW VI acquires an image from the IMAQ camera module. It quantizes the image into a 5x5, 3-bit image, and sends the data to the TM320C54x DSP via a serial port. The TM320C54x DSP constructs a 64-tap FIR by combining a series of 64-tap head related transfer functions (HRTF) according to the incoming data, and then filters an input audio signal with this FIR filter, in effect creating a correspondence between the filtered signal and the original image.
The existence of numerous imaging modalities makes it possible to present different data present in different modalities together thus forming multimodal IMAGEs. Component IMAGEs forming multimodal IMAGEs should be aligned, or registered so that all the data, coming from the different modalities, are displayed in proper locations. The term image registration is most commonly used to denote the process of alignment of IMAGEs , that is of transforming them to the common coordinate system. This is done by optimizing a similarity measure between the two IMAGEs. A widely used measure is Mutual Information (MI). This method requires estimating joint histogram of the two IMAGEs. Experiments are presented that demonstrate the approach. The technique is intensity-based rather than feature-based. As a comparative assessment the performance based on normalized mutual information and cross correlation as metric have also been presented.
-The existence of numerous imaging modalities makes it possible to present different data present in different modalities together thus forming multimodal IMAGEs. Component IMAGEs forming multimodal IMAGEs should be aligned, or registered so that all the data, coming from the different modalities, are displayed in proper locations. Mutual Information is the similarity measure used in this case for optimizing the two IMAGEs. This method requires estimating joint histogram of the two IMAGEs. The fusion of IMAGEs is the process of combining two or more IMAGEs into a single image retaining important features from each. The Discrete Wavelet Transform (DWT) has become an attractive tool for fusing multimodal IMAGEs. In this work it has been used to segment the features of the input IMAGEs to produce a region map. Features of each region are calculated and a region based approach is used to fuse the IMAGEs in the wavelet domain.
In computer vision, sets of data acquired by sampling the same scene or object at different times, or from different perspectives, will be in different coordinate systems. Image registration is the process of transforming the different sets of data into one coordinate system. Registration is necessary in order to be able to compare or integrate the data obtained from different measurements. Image registration is the process of transforming the different sets of data into one coordinate system. To be precise it involves finding transformations that relate spatial information conveyed in one image to that in another or in physical space. Image registration is performed on a series of at least two IMAGEs, where one of these IMAGEs is the reference image to which all the others will be registered. The other IMAGEs are referred to as target IMAGEs.
Java Media APIs: Cross-Platform Imaging, Media, and Visualization presents integrated Java media solutions that demonstrate the best practices for using this diverse collection. According to Sun MicroSystems, "This set of APIs supports the integration of audio and video clips, animated presentations, 2D fonts, graphics, and IMAGEs, as well as speech input/output and 3D models." By presenting each API in the context of its appropriate use within an integrated media application, the authors both illustrate the potential of the APIs and offer the architectural guidance necessary to build compelling programs.
Chinese Remainder theorem implementation. Its done in php. Its implemented with interactive gui.index.php contain all the code.Rest folders contain the css IMAGEs and js scripts that enhance the gui
This project features a complete JPEG Hardware Compressor (standard Baseline DCT, JFIF header) with 2:1:1 subsampling, able to compress at a rate of up to 24 IMAGEs per second (on XC2V1000-4 @ 40 MHz with resolution 352x288).
Image resolution is not limited. It takes an RGB input (row-wise) and outputs to a memory the compressed JPEG image. Its quality is comparable to software solutions.
The GIFT (the GNU Image-Finding Tool) is a Content Based Image Retrieval System (CBIRS: http://en.wikipedia.org/wiki/CBIR). It enables you to do Query By Example (QBE: http://en.wikipedia.org/wiki/QBE) on IMAGEs, giving you the opportunity to improve query results by relevance feedback. For processing your queries the program relies entirely on the content of the IMAGEs, freeing you from the need to annotate all IMAGEs before querying the collection.
OTSU Gray-level image segmentation using Otsu s method.
Iseg = OTSU(I,n) computes a segmented image (Iseg) containing n classes
by means of Otsu s n-thresholding method (Otsu N, A Threshold Selection
Method from Gray-Level Histograms, IEEE Trans. Syst. Man Cybern.
9:62-66 1979). Thresholds are computed to maximize a separability
criterion of the resultant classes in gray levels.
OTSU(I) is equivalent to OTSU(I,2). By default, n=2 and the
corresponding Iseg is therefore a binary image. The pixel values for
Iseg are [0 1] if n=2, [0 0.5 1] if n=3, [0 0.333 0.666 1] if n=4, ...
[Iseg,sep] = OTSU(I,n) returns the value (sep) of the separability
criterion within the range [0 1]. Zero is obtained only with IMAGEs
having less than n gray level, whereas one (optimal value) is obtained
only with n-valued IMAGEs.