A Physically-based Method for 2D and 3D Similarity and Affine Invariant Alignments

Similar documents
2 Michael E. Leventon and Sarah F. F. Gibson a b c d Fig. 1. (a, b) Two MR scans of a person's knee. Both images have high resolution in-plane, but ha

A Study of Medical Image Analysis System

Reconstruction of complete 3D object model from multi-view range images.

L1 - Introduction. Contents. Introduction of CAD/CAM system Components of CAD/CAM systems Basic concepts of graphics programming

Thin Plate Spline Feature Point Matching for Organ Surfaces in Minimally Invasive Surgery Imaging

Course Review. Computer Animation and Visualisation. Taku Komura

Nonrigid Surface Modelling. and Fast Recovery. Department of Computer Science and Engineering. Committee: Prof. Leo J. Jia and Prof. K. H.

A Survey of Volumetric Visualization Techniques for Medical Images

User-Defined B-Spline Template-Snakes

Sensor-aided Milling with a Surgical Robot System

Accurate Reconstruction by Interpolation

Volumetric Deformable Models for Simulation of Laparoscopic Surgery

An Adaptive Eigenshape Model

2D-3D Registration using Gradient-based MI for Image Guided Surgery Systems

Surgery Simulation and Planning

coding of various parts showing different features, the possibility of rotation or of hiding covering parts of the object's surface to gain an insight

Recognition (Part 4) Introduction to Computer Vision CSE 152 Lecture 17

2D rendering takes a photo of the 2D scene with a virtual camera that selects an axis aligned rectangle from the scene. The photograph is placed into

CSC Computer Graphics

Human pose estimation using Active Shape Models

Announcements. Recognition (Part 3) Model-Based Vision. A Rough Recognition Spectrum. Pose consistency. Recognition by Hypothesize and Test

Lecture 6 Geometric Transformations and Image Registration. Lin ZHANG, PhD School of Software Engineering Tongji University Spring 2013

Hand Gesture Extraction by Active Shape Models

Research Article Polygon Morphing and Its Application in Orebody Modeling

Project Updates Short lecture Volumetric Modeling +2 papers

Computer Graphics I Lecture 11

Clipping. CSC 7443: Scientific Information Visualization

Segmentation and Modeling of the Spinal Cord for Reality-based Surgical Simulator

Curves & Surfaces. Last Time? Progressive Meshes. Selective Refinement. Adjacency Data Structures. Mesh Simplification. Mesh Simplification

Image Analysis, Geometrical Modelling and Image Synthesis for 3D Medical Imaging

Open-Curve Shape Correspondence Without Endpoint Correspondence

Depth-Layer-Based Patient Motion Compensation for the Overlay of 3D Volumes onto X-Ray Sequences

Simulation in Computer Graphics. Deformable Objects. Matthias Teschner. Computer Science Department University of Freiburg

Open Topology: A Toolkit for Brain Isosurface Correction

Deforming Objects. Deformation Techniques. Deforming Objects. Examples

Volume Rendering. Computer Animation and Visualisation Lecture 9. Taku Komura. Institute for Perception, Action & Behaviour School of Informatics

3D Volume Mesh Generation of Human Organs Using Surface Geometries Created from the Visible Human Data Set

Calculating the Distance Map for Binary Sampled Data

Volume Illumination and Segmentation

INTERACTIVE CUTTING OF THE SKULL FOR CRANIOFACIAL SURGICAL PLANNING

On 3D Shape Synthesis

2008 International ANSYS Conference

Image Coding with Active Appearance Models

Image Segmentation and Registration

A Sketch Interpreter System with Shading and Cross Section Lines

3D Statistical Shape Model Building using Consistent Parameterization

Interpreter aided salt boundary segmentation using SVM regression shape deformation technique

MODELING AND HIERARCHY

Face Recognition At-a-Distance Based on Sparse-Stereo Reconstruction

Free-Form Deformation and Other Deformation Techniques

Contours & Implicit Modelling 4

Tetrahedral Mesh Modeling of Density Data for Anatomical Atlases and Intensity-Based Registration

Elastic registration of medical images using finite element meshes

Face Alignment Under Various Poses and Expressions

Volume visualization. Volume visualization. Volume visualization methods. Sources of volume visualization. Sources of volume visualization

Occluded Facial Expression Tracking

Scalar Algorithms: Contouring

Transformations Week 9, Lecture 18

SIMULATION OF ELASTIC SOFT TISSUE DEFORMATION IN ORTHODONTICS BY MASS-SPRING SYSTEM

3D Surface Reconstruction of the Brain based on Level Set Method

Internal Organ Modeling and Human Activity Analysis

Computer Vision cmput 428/615

Navigation System for ACL Reconstruction Using Registration between Multi-Viewpoint X-ray Images and CT Images

A Haptic VR Milling Surgery Simulator Using High-Resolution CT-Data.

Scalar Data. Visualization Torsten Möller. Weiskopf/Machiraju/Möller

This week. CENG 732 Computer Animation. Warping an Object. Warping an Object. 2D Grid Deformation. Warping an Object.

Accurate 3D Face and Body Modeling from a Single Fixed Kinect

Colorado School of Mines. Computer Vision. Professor William Hoff Dept of Electrical Engineering &Computer Science.

An Automatic Face Identification System Using Flexible Appearance Models

A Survey of Light Source Detection Methods

UNIT 2 2D TRANSFORMATIONS

Image Thickness Correction for Navigation with 3D Intra-cardiac Ultrasound Catheter

Contours & Implicit Modelling 1

Mathematical Tools in Computer Graphics with C# Implementations Table of Contents

EE795: Computer Vision and Intelligent Systems

Image Registration. Prof. Dr. Lucas Ferrari de Oliveira UFPR Informatics Department

EE795: Computer Vision and Intelligent Systems

Iterative Estimation of 3D Transformations for Object Alignment

Linear and Affine Transformations Coordinate Systems

Video Alignment. Final Report. Spring 2005 Prof. Brian Evans Multidimensional Digital Signal Processing Project The University of Texas at Austin

Three-Dimensional Computer Vision

Biomedical Image Processing

Development of 3D Model-based Morphometric Method for Assessment of Human Weight-bearing Joint. Taeho Kim

DISTANCE MAPS: A ROBUST ILLUMINATION PREPROCESSING FOR ACTIVE APPEARANCE MODELS

3D Mathematics. Co-ordinate systems, 3D primitives and affine transformations

Registration of Deformable Objects

CS 4620 Midterm, March 21, 2017

3D Active Appearance Model for Aligning Faces in 2D Images

Exploiting Typical Clinical Imaging Constraints for 3D Outer Bone Surface Segmentation

A New Shape Matching Measure for Nonlinear Distorted Object Recognition

Image warping , , Computational Photography Fall 2017, Lecture 10

Shape Classification and Cell Movement in 3D Matrix Tutorial (Part I)

Image-Space-Parallel Direct Volume Rendering on a Cluster of PCs

Basic principles of MR image analysis. Basic principles of MR image analysis. Basic principles of MR image analysis

3-Dimensional Object Modeling with Mesh Simplification Based Resolution Adjustment

Model-based segmentation and recognition from range data

Physically-Based Modeling and Animation. University of Missouri at Columbia

Towards full-body X-ray images

Outline. Reconstruction of 3D Meshes from Point Clouds. Motivation. Problem Statement. Applications. Challenges

Processing 3D Surface Data

Transcription:

A Physically-based Method for 2D and D Similarity and Affine Invariant Alignments Jim X. Chen and Harry Wechsler Department of Computer Science, George Mason University {jchen, wechsler}@cs.gmu.edu Abstract We present a physically-based alignment method for two point sets, which are invariant to either similarity or affine transformations. Here we assume a known correspondence between the point sets. We then employ the method to align 2D face images, find a mean-face, which is an average of all the aligned faces, and transform all face images to mean-face images for statistical face recognition. After that, we employ the same method to align D MRI slices with a reference model to reconstruct a patient s knee surface model for surgery assistance. Our method is relatively easy to implement, has no analytical deductions, and allows interactive visualizations of the transformation process. Background When comparing images to other images or models, one would like to cancel camera transformations. Werman and Weinshall developed 2D metrics to measure the similarity and affine distances between two point sets [20]. Their transformation matrices and other similar analytical methods are widely employed in 2D image alignments and recognitions [][5][4][8]. The analytical methods are concise and direct solutions to some 2D alignment problems. However, there is no known closed form for the alignment of two D point sets. For knee surgery assistance [2], we need to align two D point sets when we try to reconstruct a patient s D knee model from the MRI slices. Here we introduce a physically-based iterative method for the alignments of both 2D and D point sets with weighted extremes, which is an alternative method different from the analytical methods. Compared to the analytical methods, our method avoids complex analytical deductions needed for different weighted extremes, solves similar problems both in 2D and D, and is relatively easier to implement. 2 Similarity Invariant Alignment Given two point sets, P = {( x i, y i, z i )} n i = and Q = {( u i, v i, w i )} n i =, where P and Q are two n matrices, our goal is to find a similarity transformation that will align the first point set with the second point set, so that the Euclidean distance between points will be at minimum. That is, after translation, rotation, and uniform scale along x, y, and z axes, we have P' = srp + T, ()

n so that k i d i is at minimum, where k i is the extreme emphasis and d i = P' [] i Q[] i, namely d i = ( x i ' u i ) 2 + ( y i ' v i ) 2 + ( z i ' w i ) 2. We employ a physical-based iterative method to achieve the goal, as shown in Figure. First, we assume that the points in P are unit mass particles subject to two constraints: ) P[i] is connect to Q[i] by a spring with a force subject to Hooke s law. That is, F[i] = k i (Q[i] - P[i]), (2) where Q[i] is fixed and k i is the spring stiffness; 2) P is deformable in the sense that it can be uniformly scaled along x, y, and z axes. In other words, the size of P as an object can be changed, but the shape remains. Then, we allow P to move and deform (through translation, rotation, and uniform scale) until it achieves complete equilibrium, so that F n = F[] i = 0, () n M c = r i F[] i = 0, (4) n and F[] i is at minimum, (5) where r i is the vector from the center of mass C to P[i]. The result, P, is the similarity invariant alignment we are looking for. P[] F[] Q[] z y O x Q[n] F[n] r c P[n] P[n-] Q[n-] M c F[2] Q[2] P[2] Q[] C F F[] Q[4] P[] P[4] Fig. : Transformation according to a physically-based model From the total force, we can calculate the translation of P over time according to Newton s second law: 2

a c F F = ----- = --, V, and (6) n c = V old + a c dt P = P old + V c dt m c t t 0 t t 0 Similarly, from the total moment in Equation 4, we can calculate the rotation of P over time from H c t = M d c t = r i ( ϖ r i ) m i, (7) t 0 n where ϖ = ( ω x, ω y, ω z ). For scale, we can calculate F s = n ( P[] i C Q[] i C ) F[] i, (8) where positive F s is the shrinking force and negative F s is the expanding force of object P. Therefore, according to Equation 6 to Equation 8, for each time frame, we can find the corresponding translation, rotation, and scale to calculate new P, which will be better aligned with Q. After a period of time, the point set P will be in alignment with point set Q. Instead of strictly follow the physics equations, we further simplify and use an iterative method to find the solution. For example, in order to satisfy Equation, we translate P along x axis an initial distance tx. After the translation, if F is getting smaller, we are in the right direction and the iteration continue, otherwise we reverse the direction and reduce the translation distance (i.e., tx=-tx/2 in programming code), and continue to converge to the final solution. Similar iterations are carried out for translations along other axes, rotations, and scales along x, y, and z axes at the center of mass, respectively. The algorithm is as follows: Alignment (P, Q) { while (tx+ty+tz+wx+wy+wz+s > threshold) { translatex (P, tx); Fnew = CalculateCurrentForce(); // if force not reduced, reverse direction of convergence if (Fnew > F) tx = - tx/2; F = Fnew; translatey (P, ty); Fnew = CalculateCurrentForce(); if (Fnew > F) ty = - ty/2; F = Fnew; translatez (P, tz); Fnew = CalculateCurrentForce(); if (Fnew > F) tz = - tz/2; F = Fnew; translatexyz(p, -C); // rotate and scale along the centroid rotatex(p, wx); Fnew = CalculateCurrentForce(); if (Fnew > F) wx = - wx/2; F = Fnew; rotatey(p, wy); Fnew = CalculateCurrentForce(); if (Fnew > F) wy = - wy/2; F = Fnew; rotatez(p, wz); Fnew = CalculateCurrentForce(); if (Fnew > F) wz = - wz/2; F = Fnew;

} scalexyz(p, s); Fnew = CalculateCurrentForce(); if (Fnew > F) s = - s/2; F = Fnew; translatexyz(p, C); } output(p) The algorithm can be improved in a number of ways in the implementation, but the idea is the same. The time complexity of the algorithm depends mainly on the number of points in P and the threshold. It is an O(n) algorithm. Given initial value of q = tx+ty+tz+wx+wy+wz+s, where q<<n, it takes about log 2 q iterations to get to a threshold of one. In practice, the method is fast enough to achieve real-time animation of face model alignment. Finding 2D Shape-Free Face Images In face image categorization and recognition, we need to cancel camera transformations and normalize the orientation and shape of the images [][4][4][5]. Our goal is to model grey-level appearance independently of shape. Given the N original face images which may have different shapes and orientations, first we register the landmark points (or simply points) so that each image corresponds to a 2D point set, which is also called the shape of the face. Then, we transform the point sets to align with an predefined shape model according to similarity invariant alignment. After that, we average all the point sets to find a mean shape. Finally, we warp each face image according to its shape and the mean shape to find its corresponding mean-shape face image, or shape-free image. The resulting face images have the same shape, but different textures. The process is shown in Figure 2. face image face image 2...... face image N shape shape 2...... shape N mean shape mean face image mean face image 2...... mean face image N. Registration and Alignment Fig. 2: Process of finding shape-free face images First, we automatically superimpose a point set geometry on a face image, as shown in Figure. The landmark points, assistant lines, and the Bezier curve from point 9 to 27 are all displayed with the image. Then, we use hand-tracing to adjust those points that are not at the corresponding face image landmarks. The assistant lines and Bezier curve will interactively move according to the modified control points to assist finding the boundary of the face outline. The corresponding angles and relative distances will remain fixed to constrain the hand-tracing. This process is repeated for all N faces so we have all the face shapes, which is registered as point sets P[i], where i 20. 4

Then, alignment function is called to match with a given normalized shape, which can be an arbitrary shape in the face image set or the same as in Figure. The 2D alignment algorithm is a simplification of the algorithm given in Section 2 without translation and scale along z axis, and rotation along x and y axes. The result after alignment is saved to a file faceshape.j.txt, where j N for all face images. 9 28 7 8 6 8 0 2 2 29 9 27 20 4 5 26 2 6 5 4 25 45 7 45 45 45 22 2 24.2 Mean-Face Shape and Triangulation Fig. : Geometry of the landmark points After the alignment, we average all the faces corresponding landmark points to find a mean-face N N shape: x i = --- and. The mean-face shape is then triangulated according N x j y i i = --- N y j i to the minimum spanning algorithm, which repeats connecting the current shortest points pairs that has no cross intersection with other edges until there is no connection possible. The result is shown in Figure 4. 28 9 7 6 8 0 2 8 2 9 27 29 20 4 5 26 6 2 5 4 25 7 22 2 24 Fig. 4: Triangulation according to the growing minimum spans 5

. Warping to Shape-Free Face Images First, we apply the triangulation topology of the mean-face shape to all the face shapes. Then, according to the corresponding triangles, we warp the original face images into their corresponding mean face images, or shape-free images [][5][9]. We employ a scan-line algorithm to scan convert and fill the triangles in the output image. Given a pixel address (x, y) in the shape-free image, the inverse mapping to the corresponding point (u, v) in the original face image is shown in Figure 5 (Equation (g) and (h) are used to find the corresponding point.) Here u and v are floating-point numbers. The objective is to obtain the intensity at (u, v), which will be used as the intensity at (x, y). This is also called the resampling of the point (u, v). There are many different ways to calculate the pixel intensity of (x, y) from an interpolation of intensity at point (u, v) according to its surrounding pixel intensities [9]. Here bicubic interpolations are employed to find the corresponding intensities. P 2 (u 2,v 2 ) (u, v) (u 2,v 2 ) P (u,v ) Input Triangle P (u,v ) (u,v ) Inverse Mapping P 2 (x 2,y 2 ) e e P (x,y ) (x (x, y) 2, y) (x,y) P (x, y ) e 2 Output Triangle x 2 = x + (x 2 - x ) y - y y 2 - y (a) x = x + (x - x ) y - y y - y (b) u 2 = u + (u 2 - u ) y - y y 2 - y (c) v 2 = v + (v 2 - v ) y - y y 2 - y (d) u = u + (u - u ) y - y y - y (e) v = v + (v - v ) y - y y - y (f) u = u 2 + (u - u 2 ) x - x 2 x - x 2 (g) v = v 2 + (v - v 2 ) x - x 2 x - x 2 (h) Fig. 5: Mapping from output triangle to input triangle.4 Results Figure 6 shows how shape-free images are generated: first, the control points are collected. Then, the control points of face shapes are aligned, and mean-face shape is calculated. After that, triangulation is applied and each face is transformed into a shape-free image by mapping the corresponding triangle textures (warping). Fig. 6: Collecting control points, triangulation after alignment, and warping 6

Figure 7 shows the results of 20 shape-free images. The first (upper-left corner) face is used as the template for alignment. 4 Generating D Patient-Specific Knee Model Fig. 7: Results of 20 shape-free images Longevity of a knee joint is dependent upon keeping inside certain physiological limits for those stresses (pressures) that occur on both the convex and concave contact surfaces during normal activities. These limits include total load per unit area for the joint and a specific pattern for distributing the load on the curved surfaces. Alignment is critical for satisfying these conditions. Knee osteotomy is a kind of orthopedic surgery to re-align the lower limb by opening or cutting a bone wedge from the leg. It is a better alternative than other types of knee replacement surgeries, especially for young people. However, an osteotomy requires understanding of the imbalance of pressures at the knee joint, predicting abnormal walking gait cycle, and cutting the bone wedge precisely. Therefore, knee osteotomy is extremely difficult and prone to further damage despite the fact that it is simply a bone cut. As a result, many surgeons favor other kinds of knee replacement surgeries. Our objective is to use computer graphics to assist the knee surgery study and operation. Current static examinations of the knee such as X-rays and MRI s will be replaced by interactive visualization, surgery study, planning, exercise, and results predictions. In recent years, some computer-based surgical simulation systems are developed to help surgeons carry out knee surgeries [2][8]. However, the knee models are not patient-specific data [9]. Here we present a method of reconstructing a D patient-specific knee model from a set of MRI slices. After receiving a set of MRI slices of the knee, we employ image segmentation, alignment, surface reconstruction, as well as deformation or warping to reconstruct the patient s D knee surface model in computer. 7

4. Image Segmentation The purpose of image segmentation is to find the contour line of certain 2D object within a 2D image (or contour surface of certain D object within a D image). Manual segmentation (hand tracing), which many exiting knee reconstruction methods employ [9][], is time consuming. On the other hand, the clinical knee MRI slices usually have high noise/signal ratio and low resolution. Therefore, fully automatic segmentation is not possible and certain human interventions have to be employed to set the initial conditions, guide the segmentation process, or correct errors. Our solution to this problem uses a graphical user interface to highlight a few control points of one MRI slice (Figure 8_a), employs a deformable template (snake) method [2] to find the bone contour points, and exploits the previous contour points to serve as starting control points for the current slice, and so forth for the rest of the slices. A contour editor is constructed to modify the error output from the algorithm (Figure 8_b). The modified points will serve as input control points back to the algorithm, which will generate modified contour interactively. After this, an MRI model is constructed (Figure 8_c). 4.2 Alignment The purpose of image alignment is to find a D similarity invariant transformation so that the points in one image can be related to their corresponding points in another image or a surface model [][][7][8][20]. Surface reconstruction rebuilds an object s surface from MRI s, CT scan images, or other volumetric data gathered from different sources [6][][6]. Conventional D reconstruction methods are based on hand-traced contours [0][7], which are widely used in many commercial reconstruction packages. However, they require too much manual work and the result may not be realistic if the number of contours is not enough. We have an existing computer-based reference model a high fidelity cadaver knee surface model, which is what an ordinary knee should be. Despite small variances, ordinary human knees have same appearance and features. Our method uses physics-based similarity transformation, as discussed in Section 2, to align the MRI model (a set of image contours) with the reference model to reconstruct the patient s knee model. Specifically, we fix the MRI model, define feature control points in the reference model and the MRI model, and build up constraint forces according to the distances of the corresponding control points so that the reference model will be transformed to align with the MRI model as close as possible. The reference model only goes through an affine transformation so that its orientation, location, and size may change, but its shape will be preserved. After this, we will deform the reference model and recover the lost features between the MRI slices, as described in the next section. 4. Warping In order to recover the lost features between MRI slices, after we align the corresponding features of the MRI model with the reference model, we deform the reference model according to surface warping interpolations. First, each MRI model s vertex is paired with a reference model s vertex as shown in Figure 8_d. Because there are more vertices in the reference model than in the MRI model, the search process for vertex pairs is reversed for a given reference model s vertex, the MRI model s vertices are searched for the corresponding vertex. This way many reference model s vertices can be eliminated immediately after a few comparisons. The following function is maximized at the paired vertices: ( r n) d r, where r is the vector from 8

the MRI model s vertex to the reference model s vertex, n is the normalized surface normal at the reference model s vertex, and d is a factor to attenuate the emphasis on distance. When the function is at maximum, the distance between the paired vertices ( r ) and the angle between r and n are at minimum. A porcupine spine will be visible in D from the MRI model s vertex pointing in the same direction as r. The user can interactively modify poorly chosen paired vertices by manually rotate the spine around the MRI model s vertex, and the rest of the spines (i.e., all other paired vertices) will be adjusted automatically. After this, each of the reference model s paired vertices (that are on the spines) is translated to the corresponding contour point, and each of the reference model s other vertices (that are not on the spines) will be interpolated to its corresponding position by bi-linear interpolation according to its nearest paired neighbor vertices. After this stage, we have a patient s D surface model. a) Trace control points on one slice c) MRI contours an MRI model b) Use previous contour, modify errors, and generate current contour automatically Normal n Spine Patient MRI Model r Ref. model vertex MRI contour vertex + Reference Surface Model Patient s = D Bone Model d) Deform the reference model e) Reconstruct a patient s knee surface model Fig. 8: Construct a patient s knee MRI model and D surface model 9

5 Affine Invariant Alignment Affine transformation includes rotation, translation, as well as shears and non-uniform scales along x, y, and z axes. In general, the following matrix is used for shear transformation: shear = ab0 c d 0 e f 0 000 (9) We can modify the alignment algorithm in Section 2 to include the scales and shears: AffineAlignment (P, Q) { Alignment (P, Q); while (a+b+c+d+e+f+sx+sy+sz > threshold) { translatexyz(p, -C); sheara(p, a); Fnew = CalculateCurrentForce(); if (Fnew > F) a = - a/2; F = Fnew; shearb(p, b); Fnew = CalculateCurrentForce(); if (Fnew > F) b = - a/2; F = Fnew; shearc(p, c); Fnew = CalculateCurrentForce(); if (Fnew > F) c = - a/2; F = Fnew; sheard(p, d); Fnew = CalculateCurrentForce(); if (Fnew > F) d = - a/2; F = Fnew; sheare(p, e); Fnew = CalculateCurrentForce(); if (Fnew > F) e = - a/2; F = Fnew; shearf(p, f); Fnew = CalculateCurrentForce(); if (Fnew > F) f = - a/2; F = Fnew; scalex(p, sx); Fnew = CalculateCurrentForce(); if (Fnew > F) sx = - sx/2; F = Fnew; scaley(p, sy); Fnew = CalculateCurrentForce(); if (Fnew > F) sy = - sy/2; F = Fnew; scalez(p, sz); Fnew = CalculateCurrentForce(); if (Fnew > F) sz = - sz/2; F = Fnew; translatexyz(p, C); } output(p) } Acknowledgement Thanks to Xusheng Wang, Bobbie Pew, Douglas Pew, Ying Zhu, and Xiaodong Fu for helping with the programming. References. F. L. Bookstein, Principal Warps: Thin-Plate Splines and the Decomposition of Deformations, IEEE Transactions on Pattern Analysis and Machine Intelligence, Vol., No. 6, June 989, pp. 567-585. 0

2. E. Y. S. Chao, et al., Computer-aided Preoperative Planning in Knee Osteotomy, The Iowa Orthopedic Journal, Vol. 5, 995, pp. 4-8.. T. F. Cootes, C. J. Taylor, D. H. Cooper, and J. Graham, Active Shape Models their Training and Applicaiton, Computer Vision and Image Understanding, Vol. 6, No., January, 995, pp. 8-59. 4. I. Craw, N. Costen, and T. Kato, Automatic Face Recognition, http:// www.maths.abdn.ac.uk/~igc/papers/rafr/rafr.html. 5. A. D. J. Cross and E. R. Hancock, Holistic Matching, 5th European Conference on Computer Vision (ECCV 98), Springer-Verlag, Berlin Heidelberg, 998, pp. 40-55. 6. M. Eck and H. Hoppe, Automatic reconstruction of B-spline surfaces of arbitrary topological type, Proceedings of SIGGRAPH 96, pp. 25-2. 7. P. A. Elsen, E. D. Pol, and M. A. Viergever, Medical image matching - a review with classification, IEEE Engineering in Medicine and Biology, March 99, pp. 26-9. 8. M. Fadda, et al., Computer assisted planning for total knee arthroplasty, Proc. of st Joint Conf. of Computer Vision, Virtual Reality and Robotics in Medicine and Medical Robotics and Computer-assisted Surgery (CVRMed-MRCAS 97), Grenoble, France, Mar. 997, pp. 69-627. 9. S. Gibson, et al., Simulating arthroscopic knee surgery using volumetric object representation, real-time volume rendering and haptic feedback, Proc. of st Joint Conf. of Computer Vision, Virtual Reality and Robotics in Medicine and Medical Robotics and Computer-assisted Surgery (CVRMed-MRCAS 97), Grenoble, France, Mar. 997, pp. 69-78. 0. D. P. Huijsmans and G. J. Jense, Representation of D objects, reconstructed from series of parallel 2D slices, Theoretical Foundations of Computer Graphics and CAD. 988, pp. 02-07.. T. Kapur, et al., Model based segmentation of clinical knee MRI,, International Conference on Computer Vision (ICCV 98), Bombay, India, January, 998. 2. M. Kass, A. Witkin, and D. Terzopoulos, Snakes - Active contour models, International Journal of Computer Vision, Vol., No. 2, 987, pp. 259-268.. M. Levoy, Display of surface from volume data, IEEE Computer Graphics & Applications, May 988, pp. 29-7. 4. K. M. Lam and H. Yan, An Analytic-to-Holistic Approach for Face Recognition Based on a Single Frontal View, IEEE Transactions on Pattern Analysis and Machine Intelligence, Vol. 20, No. 7, July, 998, pp. 67-686. 5. A. Lanitis, C. J. Taylor, and T. F. Cootes, A Unified Approach to Coding and Interpreting Face Images, International Conference on Computer Vision (ICCV 95), 995, pp. 68-7. 6. W. E. Lorensen and H. E. Cline, Marching cubes: a high resolution D surface reconstruction algorithm, Proceedings of SIGGRAPH 87, 987, pp. 6-69. 7. D. Meyers, S. Skinner and K. Sloan, Surfaces from contours, ACM Transactions on Graphics, Vol., No., 992, pp. 228-258. 8. K. Nagao and W. E. L. Grimson, Affine Matching of Planar Sets, Computer Vision and Image Understanding, Vol. 70, No., April, 998, pp. -22. 9. W. T. Vetterling, et al., Numerical Recipes in C: The Art of Scientific Computing, 99. 20. M. Werman and D. Weinshall, Similarity and Affine Invariant Distances between 2D Point Sets, IEEE Transactions on Pattern Analysis and Machine Intelligence, Vol. 7, No. 8, August, 995, pp. 80-84. 2. Y. Zhu, J. X. Chen, S. Xiao, and E. B. MacMahon, D Knee Modeling and Biomechanical Simulation, to appear in Computing in Science and Engineering, May/June 999.