While current face recognition algorithms have provided convincing performance on frontal face poses, recognition is far less effective when the pose and illumination conditions vary. Here the authors show how compound image transforms can be used for face recognition in various poses and illumination conditions. The method works by first dividing each image into four equal-sized tiles. Then, image features are extracted from the face images, transforms of the images, and transforms of transforms of the images. Finally, each image feature is assigned with a Fisher score, and test images are classified by using a simple Weighted Nearest Neighbor rule such that the Fisher scores are used as weights. Experimental results using the full color FERET dataset show that with no parameter tuning, the accuracy of rank-10 recognition for frontal, quarter-profile, and half-profile images is ~98%, ~94% and ~91%, respectively. The proposed method also achieves perfect accuracy on several other face recognition datasets such as Yale B, ORL and JAFFE. An important feature of this method is that the recognition accuracy improves as the number of subjects in the dataset gets larger.