August 3 - 12, 2011
Figure 1. Segmentation of the internal carotid artery (left).
Vessel tree with the common, internal and external carotid
Accurate vessel segmentation is required in many clinical
applications, such as identifying the degree of stenosis
(narrowing) of a vessel to assess if blood flow to an organ is
sufficient, quantification of plaque buildup (to determine the
risk of stroke, for example), and in detecting aneurisms which
pose severe risks if ruptured. Proximity to bone can pose
segmentation challenges due to the similar appearance of bone
and contrasted vessels in CT (Figure 1 – the internal carotid
has to cross the skull base); other challenges are posed by low
X-ray dose images, and pathology such as stenosis and
Figure 2. Cross section of vessel segmentation from CT data,
shown with straightened centerline.
A typical segmentation consists of a centerline that tracks the
length of the vessel, lumen surface and vessel wall surface.
Since for performance reasons most clinical applications use
only local vessel models for detection, tracking and
segmentation, in the presence of noise the results can become
physiologically unrealistic – for example in the figure above,
the diameter of the lumen and wall cross-sections vary too
Figure 3. Vessel represented as a centerline with periodically
sampled cross-sections in the planes orthogonal to the
centerline. Note that some planes intersect, which makes this
representation problematic. The in-plane cross-sections of the
vessel are shown on the right.
The goal of this project is to design a method for refining a
vessel segmentation based on the following general approach:
- Choose an appropriate geometric representations for vessel
segmentation (e.g., generalized cylinders) and derive the
equations and methods necessary to manipulate it as required
and to convert to and from the representation. One common, but
sometimes problematic representation is shown in Figure 3.
- Learn a geometric model for vessels based on the
representation from a set of training data (for example
segmentations obtained from low-noise clinical images). Example
- Relative rate of vessel diameter change as a function of
- Typical wall thickness as a function of lumen cross-section area
Learn an appearance model for the vessels that captures
details about how vessels appear in a clinical imaging modality
such as CT. For example:
- Radial lumen intensity profile in Hounsfeld units
- Rate of intensity change along the centerline
Compute the most
likely vessel representation given a starting segmentation and
the learned geometric and appearance models.
The project will use real clinical data and many different
types of vessels.References:
- C. Kirbas and F. Quek. “A review of vessel extraction
techniques and algorithms”. ACM Computing Surveys, vol. 36, pp.
- T. McInerney and D. Terzopoulos. “Deformable models in
medical image analysis: A survey”. Medical Image Analysis, vol.
1, pp. 91 – 108, 1996.
Optimization, Statistics and Estimation, Differential Equations
and Geometry. MATLAB programming.
Vessel segmentation, shape statistics, appearance models
Modern commercial airplanes are assembled out of millions of different parts. While many of these parts are rigid, many of them are not. For example, the hydraulic lines and flexible electrical conduits that supply an airplane's landing gear change their shape as the landing gear goes through its motion (you can see some of these lines in the accompanying photograph). These shapes can be modeled by minimizing the potential energy of the rest state of one of these flexible lines as the ends of the lines are moved by the landing gear. While this problem is amenable to solution through direct optimization of individual finite elements, the method often proves to be slow and unreliable. In this investigation, we will explore the use of variational methods (i.e. the calculus of variations) in an attempt to discover a more elegant and robust approach to modeling these flexible airplane parts. Reference:
Any textbook on the calculus of variations. My favorite is The Variational Principles of Mechanics, by Cornelius Lanczos.
Geometrical modeling, calculus of variations, boundary value problemsPrerequisites:
Calculus of variations, optimization, numerical methods for ODEs and 2-point boundary value problems, Matlab
Driven by rapid advances in many fields including Biology, Finance and Web Services, applications involving millions or even billions of data items such as documents, user records, reviews, images or videos are not that uncommon. Given a query from a user, fast and accurate retrieval of relevant items from such massive data sets is of critical importance. Each item in a data set is typically represented by a feature vector, possibly in a very high dimensional space. Moreover, such a vector tends to be sparse for many applications. For instance, text documents are encoded as a word frequency vector. Similarly, images and videos are commonly represented as sparse histograms of a large number of visual features. Many techniques have been proposed in the past for fast nearest neighbor search. Most of these can be divided in two paradigms: Specialized data structures (e.g., trees), and hashing (representing each item as a compact code). Tree-based methods scale poorly with dimensionality, typically reducing to worst case linear search. Hashing based methods are popular for large-scale search but learning accurate and fast hashes for high-dimensional sparse data is still an open question.
In this project, we aim to focus on fast approximate nearest neighbor search in massive databases by converting each item as a binary code. Locality Sensitive Hashing (LSH) is one of the most prominent methods that uses randomized projections to generate simple hash functions. However, LSH usually requires long codes for good performance. The main challenge of this project is how to learn appropriate hash functions that take input data distribution into consideration. This will lead to more compact codes, thus reducing the storage and computational needs significantly. The project will focus on understanding and implementing a few state-of-the-art hashing methods, developing the formulation for learning data-dependent hash functions assuming a known data density, and experimenting with medium to large scale datasets.Keywords:
Approximate Nearest Neighbor (ANN) search, Hashing, LSH, Sparse data, High-dimensional hashing References:
For a quick overview of ANN search, review the following tutorials (more references are given at the end of the tutorials):
- Good computing skills (Matlab or C/C++)
- Strong background in optimization, linear algebra and calculus
- Machine learning and computational geometry background preferred but not necessary
Keywords of the presentation: EUV lithograpy, diffraction
A PC sold in 2010 had billions of transistors with 32 nm gate-length. In a year, that dimension will shrink to 22 nm. Light is essential to fabrication and quality control of such small semiconductor devices.
Integrated circuits are manufactured by repeatedly depositing a film of material and etching a pattern in the deposited film. The pattern is formed by a process called photo lithography. An optical image of a photomask is projected on to a silicon wafer on which the integrated circuits will be formed. Presently, 193 nm wavelength deep UV light is used to project the image. Photomasks are inspected for defects by deep UV microscopy, which is subject to the same resolution limit as lithography. Manufacturing next generation integrated circuits and inspecting their photomasks require higher-resolution imaging. The leading candidate for next generation lithography, EUV (extreme ultraviolet) lithography uses 13.5 nm wavelength. At that wavelength, no material is highly reflective or transparent. The photomask and mirrors of the projector are coated with a multi-layer Bragg reflector to achieve sufficient reflectance. The photomask has a patterned absorber film over the Bragg reflector. The thickness of the absorber and the lateral dimensions of the mask pattern are on the order of four EUV wavelengths.
Rapid yet accurate simulation of EUV image formation is essential to optimizing the photomask pattern and interpreting the images of an inspection microscope. This project concerns the key step of image simulation, which is calculating the diffracted near-field of the EUV photomask.
- G. Kirchhoff, Vorlesungen uber Mathematiche Physik, Zweiter Band, Mathematiche Optik, Leipzig, Druck und Verlag, 1891, also books.google.com, p 80 states the Kirchhoff approximation.
H. H. Hopkins, “On the diffraction theory of optical images,” Proc. Roy. SOC. London, Ser. A 217, pp. 408-432, 1953.
- C. A. Mack, Inside PROLITH: A Comprehensive Guide to Optical Lithography Simulation, FINLE Technologies (Austin, TX: 1997).
K. Adam, A. R. Neureuther, “Domain decomposition methods for the rapid electromagnetic simulation of photomask scattering,” Journal of Micro/Nanolithography, MEMS, and MOEMS, Vol. 1, No. 3, p. 253-269, 2002, SPIE, Bellingham, WA.
P. Evanschitzky, F. Shao, A. Erdmann, “Simulation of larger mask areas using the waveguide method with fast decomposition technique”, Proc. SPIE Vol. 6730, 2007, SPIE, Bellingham, WA.
Diffraction, Computational Lithography, Partial Coherence ImagingPrerequisites:
Basic optics, electromagnetics, computational methods for wave equations
In the next generation electrical grid, or "smart grid", there will be many heterogeneous power generators, power storage devices and power consumers. This will include residential customers who traditionally are only part of the ecosystem as consumers, but will in the foreseeable future increasingly provide renewable energy generation through photovoltaics and wind energy and provide energy storage through plug-in hybrid vehicles. What makes this electrical grid "smart" is the capability to insert a vast number of sensors and actuators into the system. This allows a wide variety of information about all the constituents to be collected and various aspects of the electrical grid to be controlled via advanced electric meters, smart appliances, etc. Information gathered consists of e.g. amount of energy use, planned energy consumption, efficiency and status of equipment, energy generation costs, etc and this information is then used by all constituents to optimize certain objectives. This necessitates communication and information technology to transmit and process this information. The goal of this project is to focus on the optimization of local objectives in a smart grid. In particular, we study various centralized and decentralized optimization algorithms to determine the optimal matching and maintain stability between energy producers, energy storage, and energy consumers all connected in a complex and dynamic network. Technical prerequisites:
scripting languages (Matlab, python), optimization, linear and nonlinear programming. Preferred but not necessary:
graph theory, combinatorics, computer programming, experience with CPLEX, R.