MF3D Stimulus Set 1¶
The MF3D stimulus set (Release 1) consist of 14,000 high resolution images and animated video sequences. The images and animations are hosted on Figshare and can be downloaded separately ():
What’s in MF3D R1?¶
MF3D R1 is a publicly released stimulus set for the macaque research community that consists of 14,000 static renders of the macaque avatar, saved as high resolution (3840 x 2160 pixels, 32-bit) RGBA images in .png format (Figure 1A, i) and a smaller sample of animated video clips. The inclusion of the alpha transparency channel allows for compositing of multiple images into a frame, including backgrounds, as well as making it easy to generate control stimuli with identical silhouettes. The high resolution permits down-sampling or cropping as appropriate for the display size being used.
The virtual scene was configured such that the avatar will appear at real-world retinal size when images are presented at full-screen on a 27” monitor with 16:9 aspect ratio, at 57cm viewing distance from the subject. For each 2D colour image, we additionally provide a label map image, which is an indexed image that assigns each pixel an integer value depending on the anatomical region of the avatar it belongs to (Figure 1A, ii). Label maps can be used to analyse subjects’ gaze in free-viewing paradigms (Figure 1A, iii-iv), for gaze-contingent reward schedules, or for generating novel stimuli by masking specific structures in the corresponding colour image.
The static stimuli of MF3D release 1 are divided into two collections:
variable expressions with fixed identity (corresponding to real individual M02); and
variable identities with fixed expression (neutral).
MF3D R1 Subsets¶
For the expression set, we varied head orientation (±90° azimuth x ±30° elevation in 10° increments = 133 orientations; Figure 1B, i), facial expression type (neutral plus bared-teeth ‘fear grimace’, open-mouthed threat, coo, yawn, and tongue-protrusion = 5) and the intensity of the expression (25, 50, 75 and 100% = 4; Figure 1B, ii). We additionally include the neutral expression with open and closed eyes, as well as azimuth rotations beyond 90° (100 to 260° in 10° increments) for a total of 2,926 colour images. In order to maintain naturalistic poses, head orientation was varied through a combination of neck (±30° azimuth and elevation) and body (±60° azimuth) orientations.
For the identity set, we selected a subset of head orientations (±90° azimuth x ±30° elevation in 30° increments = 21 orientations; Figure 1C, i), and co-varied facial morphology based on distinct trajectories within PCA-space (n = 65; Figure 1C, ii), including each of the first five PCs (which together account for 75% of the sample variance in facial morphology), with distinctiveness (Euclidean distance from the average face, ±4σ in 1σ increments = 8 levels, excluding the mean; Figure 1C, iii) for a total of 10,941 identity images.
For studies requiring more naturalistic stimuli, we also have the ability to generate a virtually limitless number of animations that promise great flexibility for studying dynamic facial behaviour. Here we have included a small selection of short animations (2 seconds or less per clip) as a proof of concept, which are rendered at 3840 x 2160 pixels and 60 frames per second, encoded with H.264 perceptually-lossless compression and saved in .mp4 format with a black background. For each action sequence, animations are rendered at 5 different head azimuth angles (-60, to 60° in 30° increments). All animations feature identical start and end frames, which allows the possibility of stitching multiple clips together using video editing software (such as the video editor included in Blender), to produce longer, seamless movies containing various permutations of action sequences. We provide a Python script to demonstrate automated compilation of animation clips using Blender’s video sequence editor. The animations were produced by manually coding video footage of real Rhesus macaques performing facial expressions and vocalizations.