Skip to main content
PLOS Biology logoLink to PLOS Biology
. 2024 Dec 3;22(12):e3002867. doi: 10.1371/journal.pbio.3002867

Hyperspectral imaging in animal coloration research: A user-friendly pipeline for image generation, analysis, and integration with 3D modeling

Benedict G Hogan 1,*, Mary Caswell Stoddard 1,*
Editor: Gail L Patricelli2
PMCID: PMC11614258  PMID: 39625994

Abstract

Hyperspectral imaging—a technique that combines the high spectral resolution of spectrophotometry with the high spatial resolution of photography—holds great promise for the study of animal coloration. However, applications of hyperspectral imaging to questions about the ecology and evolution of animal color remain relatively rare. The approach can be expensive and unwieldy, and we lack user-friendly pipelines for capturing and analyzing hyperspectral data in the context of animal color. Fortunately, costs are decreasing and hyperspectral imagers are improving, particularly in their sensitivity to wavelengths (including ultraviolet) visible to diverse animal species. To highlight the potential of hyperspectral imaging for animal coloration studies, we developed a pipeline for capturing, sampling, and analyzing hyperspectral data (here, in the 325 nm to 700 nm range) using avian museum specimens. Specifically, we used the pipeline to characterize the plumage colors of the King bird-of-paradise (Cicinnurus regius), Magnificent bird-of-paradise (C. magnificus), and their putative hybrid, the King of Holland’s bird-of-paradise (C. magnificus x C. regius). We also combined hyperspectral data with 3D digital models to supplement hyperspectral images of each specimen with 3D shape information. Using visual system-independent methods, we found that many plumage patches on the hybrid King of Holland’s bird-of-paradise are—to varying degrees—intermediate relative to those of the parent species. This was true of both pigmentary and structurally colored plumage patches. Using visual system-dependent methods, we showed that only some of the differences in plumage patches among the hybrid and its parent species would be perceivable by birds. Hyperspectral imaging is poised to become the gold standard for many animal coloration applications: comprehensive reflectance data—across the entire surface of an animal specimen—can be obtained in a matter of minutes. Our pipeline provides a practical and flexible roadmap for incorporating hyperspectral imaging into future studies of animal color.


Animal coloration research has flourished recently but the uptake of hyperspectral imaging methods has been slow. This paper develops a user-friendly pipeline for hyperspectral imaging and highlights its advantages in measuring animal coloration by comparing two birds-of-paradise and their rare hybrid offspring.

Introduction

Investigating animal coloration has proven useful for unraveling the evolutionary processes that generate phenotypic diversity [1]. In recent years, quantifying animal color has entailed two main methods: spectrophotometry and multispectral digital photography [26]. Spectrophotometry permits measurements of the reflectance across many wavelengths for a single point, while multispectral photography captures detailed spatial information in a limited number of color channels. Each approach has distinct advantages, but choosing one method over the other forces a trade-off between high spectral resolution (spectrophotometry) and high spatial resolution (photography, see Table 1).

Table 1. Table highlighting some differences and similarities among spectrophotometry, multispectral photography, and hyperspectral imaging, with respect to spectral resolution, spatial resolution, sample selection, animal vision models, and cost.

Not included in the table is temporal resolution because the amount of time required to sample an individual animal is highly variable both within and among techniques. For spectrophotometry, individual measurements are collected in under a second, but typically several measurements per patch and several patches per animal are collected, resulting in a temporal resolution spanning from several seconds to minutes. For multispectral photography, capture with limited spectral sensitivity (e.g., omitting ultraviolet light) can allow for video-speed capture, but increasing spectral sensitivity typically slows capture to the range of minutes per image (but see [7] for multispectral video including ultraviolet light). For hyperspectral imaging, a similar trade-off can occur: full-frame hyperspectral imagers with limited spatial and spectral sensitivity can capture images at video-speed. The hyperspectral imager used in the current paper has high spatial and spectral resolution, but each image takes several minutes to capture.

Spectrophotometry Multispectral photography Hyperspectral imaging
Spectral resolution High Low High
Spatial resolution Low High High
Sample selection Manual Comprehensive Comprehensive
May be applied to animal vision models Yes Yes Yes
Cost Low-medium Low Currently high

Spectrophotometry has been the go-to tool for animal color researchers since the 1980s and 1990s, when spectrophotometers became widespread, affordable, and capable of capturing wavelengths beyond the human visible range [4,6,8]. Using this approach, reflectance measurements of animal coloration are obtained with a spectrometer and a broadband light-source connected via optical fibers to a small probe. The probe, placed on a small region of an animal’s feathers, fur, or skin, captures detailed information about the light reflected over a broad range of wavelengths. The resulting reflectance spectrum has high spectral resolution (i.e., comprising hundreds of spectral measurements), but the approach suffers from a lack of spatial resolution: spectrophotometry typically provides inadequate 2D and 3D spatial information about animal color patterning. This is because only point samples can be obtained, so color sampling across the body of an animal (for example, a taxidermy specimen) is typically sparse. Consequently, researchers might fail to measure color patches (body regions with apparently uniform color; [9]) that are visible to some animals but not to humans [10] or fall short of capturing meaningful variation within a given color patch [11].

Over the last 15 years, multispectral digital photography has emerged as a complementary approach to spectrophotometry. Multispectral photography typically involves using a consumer camera that has been modified for broadband light sensitivity (to permit ultraviolet sensitivity, for example). External lens filters are sequentially attached to the camera, allowing for a series of images to be captured across multiple color channels encompassing (human) visible and ultraviolet wavelengths [4,5]. These images can then be stacked and processed (i.e., controlling for nonlinearity in raw pixel values, the spectral sensitivity of the particular model of camera, and lighting conditions) using specialized software. The result is a digital multispectral image comprising multiple channels: typically ultraviolet, blue, green, and red, or sometimes more (see [12]). Ultimately, multispectral photography generates images with high spatial resolution [4,5] because each pixel in the image represents a measurement of color. However, each of these pixels contains only a few spectral measurements (one for each channel), severely limiting spectral resolution. Overall, multispectral photography resolves some of the challenges of spectrophotometry—it captures detailed spatial information (including patterning) and permits dense sampling across an animal’s body—but at the expense of spectral resolution.

To quantify animal color accurately and comprehensively, it is desirable to capture both high spectral resolution and high spatial resolution [13]. Although reflectance spectra (high spectral resolution) and multispectral images (high spatial resolution) are both amenable to visual modeling and can be used to address a range of questions about how animals perceive color signals (reviewed in [14,15]), for many studies researchers must choose one approach in favor of the other. Reflectance spectra are essential when the goal is to identify metamers—two colors that look identical to a viewer but in fact have different spectral properties [16]. Moreover, knowing the shape and features of reflectance spectra can be especially valuable for addressing questions about the physical properties of an animal’s color, such as those related to the mechanisms underlying color production [1719] or to non-signaling functions of color like thermoregulation [20,21] and mechanical strength [22]. In these cases—in which a direct link to a particular visual system is often unimportant or unknown—analyses of reflectance spectra are usually the most appropriate (reviewed in [14]). On the other hand, for questions related to spatial information—such as those involving the distribution of patches and fine-scale patterning on an animal’s body (see [9,23])—multispectral images have the clear advantage [35]. Multispectral photography can also be combined with photogrammetry to produce multispectral 3D models, allowing for integrative studies of color and 3D body morphology [24].

Hyperspectral imaging is a compelling alternative to spectrophotometry and multispectral photography: it offers the best of both worlds by providing both high spectral resolution and high spatial resolution (Table 1). Hyperspectral imagers work by simultaneously capturing high-resolution color information at many locations, in essence taking a photograph with hundreds of channels of color information instead of relatively few as in traditional (RGB) or multispectral photography (see Fig 1 and Fig D in S1 Appendix). A hyperspectral imaging system separates the light spectrum into a series of narrow bands, each of which corresponds to a small wavelength range (Fig 1). The end result is a data array or stack (often termed a data-cube), recording radiance along two spatial dimensions (width, height) and one spectral dimension (wavelength). Each pixel in a hyperspectral image can correspond to a full reflectance spectrum (Fig 1C), and an image of the brightness of the whole specimen can be produced for any desired wavelength (by taking a slice of the data array, see Fig 1A and 1D). While hyperspectral imaging has become an important tool in agriculture, geology, medicine, archeology, art history, and human color science [16,25,26], applications of hyperspectral imaging to animal color quantification are in their relative infancy. To study animal color, early adopters of hyperspectral imaging used the technique to investigate the tuning of dichromatic visual pigment sensitivities to real forest and underwater scenes [27], surface color in butterfly wings [28], iridescent spiders [29], beetle wing cases [30], and frog skin [31], and camouflage in cuttlefish [32] and crabs on natural substrates [33]. Additionally, researchers have used hyperspectral imaging to simulate avian perception of warning colors on lepidopteran wings [34], to measure color and simulate warning coloration for newts [35], and to measure bioluminescence in sharks [36]. Kim and colleagues [37] combined hyperspectral imaging with 3D scanning to produce the first—to our knowledge—hyperspectral 3D model of an animal museum specimen, a colorful Papuan lorikeet (Charmosyna papou goliathina; now considered Stella's lorikeet, Charmosyna stellae).

Fig 1. Hyperspectral imaging data captures spectral and spatial information.

Fig 1

(a) Montage of selected slices across the spectral dimension of the hyperspectral data-cube, each indicating in grayscale the reflectance of one specimen (Magnificent bird-of-paradise dorsal view) at the wavelength indicated in the inset colored boxes. The blue, orange, green, and red dots indicate the plumage regions sampled (see also b and c). Note that intervals of 20 nm were chosen for plotting; the actual data were acquired in intervals of ~2.5 nm. (b) RGB image of the dorsal view of the specimen generated using 3 select slices (425 nm, 550 nm, 600 nm) with plumage regions sampled again indicated with colored dots. (c) Plot of the reflectance of each of the plumage samples shown in (a) and (b); each pixel in a hyperspectral image corresponds to a full reflectance spectrum. The dotted vertical line indicates the lower sensitivity bound of the hyperspectral image; reflectance values below this wavelength were extrapolated. (d) The birds-of-paradise measured in this study reflect relatively little ultraviolet light (300–400 nm). For comparison, shown here are hyperspectral images of the lateral view of a Gouldian finch (Chloebia gouldiae), which reflects substantial ultraviolet light in the breast patch. The data underlying this figure can be found at https://doi.org/10.5061/dryad.j0zpc86nf [99].

Despite these applications, widespread adoption of hyperspectral imaging in animal color research has been slow. There are three main obstacles. First, hyperspectral imaging systems are expensive and require substantial computational power and memory for data processing [37]. Second, hyperspectral imagers that are sensitive to the entire portion of the visible spectrum relevant to animals have not typically been commercially available [38]; for many animals, this range is 300 nm to 700 nm [39], with 300 nm to 400 nm comprising ultraviolet light. Finally, we lack a strong conceptual framework—and corresponding software tools—for analyzing hyperspectral data in a way that is convenient for animal color research. Fortunately, the first two of these challenges are quickly fading. Hyperspectral imaging systems are becoming increasingly affordable and computationally practical, and some imagers now have sensitivity across wavelengths (including ultraviolet) relevant to animal visual systems [34]. Thus, the stage is set for wider uptake of hyperspectral imaging, which could be a game-changing tool in animal color research. All of the advantages to multispectral photography [2,5] apply to hyperspectral images, but with the additional benefit of high-resolution spectral data. For example, a single data set of hyperspectral images might now be used in two ways: (i) to ask questions about an animal’s coloration and patterning (i.e., the images can be processed using models of animal vision and analyzed using existing tools for multispectral images); and (ii) to ask questions about the physical properties of the colors themselves (i.e., by comparing the shape and features of the reflectance spectra to those of known pigments and structural colors).

We currently lack a user-friendly, step-by-step pipeline for obtaining and analyzing hyperspectral images for animal color research. To address this gap, we developed such a pipeline and demonstrated how it can be used to visualize and compare animal colors in hyperspectral images. We used a commercially available Resonon Pika NUV imager (Resonon, Montana, United States of America), but the pipeline is general to hyperspectral animal coloration data captured by any hyperspectral imager. Our pipeline includes: imaging, sampling reflectance spectra from the images, and embedding the resulting colors into receiver-independent and receiver-dependent color spaces. To illustrate how hyperspectral imaging can provide new, detailed insights about animal color, we apply our pipeline to a study of the plumage of a rare hybrid bird-of-paradise, with only around 25 male museum specimens known in the world [40]. Birds, the most colorful land vertebrates, are especially amenable to hyperspectral imaging and analysis. Their colors are produced by diverse mechanisms (pigments and structural colors), they have tetrachromatic color vision, and bird specimens are readily available in museum collections [3]. Overall, hyperspectral imaging may represent a new gold standard for many avenues of animal coloration research.

Study system: Birds-of-paradise

The birds-of-paradise (family: Paradisaeidae), native to New Guinea and Australia, are a diverse and charismatic group known for their elaborate multimodal courtship displays [41]. The family comprises 44 species in 16 or 17 genera [40,42] (Fig 2). Different species exhibit a strong tendency to hybridize, with hybridization known or suspected among very different-looking species [43,44]. One rare bird-of-paradise hybrid—with just ~25 male museum specimens in the world [40]—is the King of Holland’s bird-of-paradise (Cicinnurus magnificus x C. regius) [4548]. Initially thought to be a new species (by Meyer in 1875 [47]), the King of Holland’s bird-of-paradise was later described as a hybrid of the Magnificent bird-of-paradise (C. magnificus, following the convention used in [42]; also known as Diphyllodes magnificus) and the King bird-of-paradise (C. regius). In 1927, Berlioz [45] described the King of Holland’s bird-of-paradise as the perfect combination of its parents’ phenotypes, writing (in French) that “an artist who tried to imagine a hybrid could not compose a more faithful reproduction.” More recently, Thörn and colleagues [49] confirmed genetically that a male King of Holland’s bird-of-paradise specimen was indeed an F1 hybrid. However, they also found that a male specimen originally identified as a King of Holland’s bird-of-paradise was genetically closer to a Magnificent bird-of-paradise, while a female specimen thought to be a King of Holland’s bird-of-paradise was more similar to a King bird-of-paradise.

Fig 2. The birds-of-paradise examined in this study.

Fig 2

Top row: Illustrations of the King bird-of-paradise, the hybrid King of Holland’s bird-of-paradise, and the Magnificent bird-of-paradise. The King and Magnificent illustrations are reproduced from Levaillant and Barraband (1806, [50]), which is in the public domain. The hybrid illustration is reproduced from Gould and colleagues (1875, [51]), which is in the public domain. Middle row: Icons and colors representing the parent species and hybrid throughout the paper. Bottom row: Phylogenetic reconstruction of the subfamily Paradisaeinae, based on our replication of the phylogeny by Irestedt and colleagues [52]. All individual specimens and all genera other than Cicinnurus and Paradisaea are collapsed for clarity. See Table B and Fig J in S1 Appendix for details of the phylogenetic reconstruction.

The male King, Magnificent, and King of Holland’s birds-of-paradise exhibit plumage colors produced by diverse plumage mechanisms. Avian plumage coloration is generated through two broad mechanisms, pigmentary and structural color, or a combination of the two [19]. Pigmentary color is generated by the deposition of colorful substances within the feather, whereas structural color arises from nano-scale physical structures in the feather barbs and barbules. Pigmentary mechanisms commonly generate a range of plumage colors, including gray and black (melanin), brown and rufous (pheomelanin), and yellow, orange, and red (carotenoids). Structural mechanisms are typically responsible for blue and iridescent colors, and green can be generated by a combination of yellow carotenoid pigmentation and structural blue. White plumage typically results from lack of pigmentation. Although there are few descriptions of color-production mechanisms in the King, Magnificent, and King of Holland’s birds-of-paradise (but see [53]), pigmentary [54] and structural colors have been described in a range of other bird-of-paradise species [5557]. Females of both the Magnificent and King birds-of-paradise are less colorful, with a brown head, shoulders, back and wings, and barred undersides [40]. The plumage of the female hybrid King of Holland’s bird-of-paradise is likely to be drab; at least one specimen labeled as a female hybrid was recently found to be a King bird-of-paradise, though the existence of female hybrids is evidenced by at least one apparent backcross between a fertile female King of Holland’s bird-of-paradise and a male King bird-of-paradise [49].

How hybridization affects coloration is not well understood, especially where coloration is non-uniform (i.e., consisting of distinct patches) across an animal’s body. In birds, it is common for even adjacent patches of coloration to be generated by different structural or pigmentary mechanisms [19]. In some species, a hybrid may have a mosaic of color patches that individually match one or the other parent species [58,59], or each patch may in itself appear to be intermediate to the parent species’ patches [60,61], or both [62,63]. In other cases, combinations of parental genotypes can produce apparently novel phenotypes, as in the crown patch of hybrid offspring of snow-capped (Lepidothrix nattereri) and opal-crowned manakins (L. iris [64]), and in hummingbird throat patches [65]. To what extent are the hybrid King of Holland’s bird-of-paradise plumage patch colors intermediate composites of its parent species? Are individual patches intermediate, or is the hybrid a mosaic of the patches of its parents? Are the colors intermediate in terms of physical properties (reflectance spectra)? Would they be perceived as intermediate by a relevant signal receiver, such as another bird-of-paradise? To determine—quantitatively, using objective color measurements—whether the hybrid plumage patches are intermediate, refined spectral and spatial color quantification is required. This is therefore an ideal system for the application and testing of hyperspectral imaging.

Methods

A pipeline for obtaining and analyzing hyperspectral color data

We developed a pipeline for color analysis that takes advantage of the spectrally and spatially rich data produced by hyperspectral images (Fig 1). We provide a subset of the images, code, and data used in this pipeline so that researchers can replicate our approach and modify it for their specific research purposes. The pipeline consists of three steps: (i) imaging; (ii) sampling; and (iii) low-dimensional embedding (see Fig 3). Imaging involves acquiring images from the hyperspectral camera (Fig 3A) using SpectrononPro software (Resonon, Montana, USA). The subsequent sampling and embedding steps are performed using custom python code (available at https://github.com/bghogan/HyperBirdOfParadise). The code uses functions from several packages, including colour-science (https://github.com/colour-science/colour), spectral python (https://www.spectralpython.net/), and UMAP (https://github.com/lmcinnes/umap). Sampling involves selecting relevant colors from the hyperspectral images for analysis (Fig 3B–3D). Low-dimensional embedding involves generating informative transformations of the spectral data for interpretation and analysis (Fig 3E and 3F). Following low-dimensional embedding, researchers can perform a range of analyses to quantify color diversity and color difference. We discuss two potential low-dimensional embeddings: a visual system-independent approach and a visual system-dependent approach, here using an avian tetrahedral color space [6668] to quantify aspects of avian color perception. As an optional fourth step of the pipeline, we describe ways in which researchers might supplement hyperspectral data with 3D models to understand how the 3D surface of the subject in the image might influence color. Here, we employ this pipeline to describe the plumage coloration of two parent species (the Magnificent bird-of-paradise and the King bird-of-paradise) and their hybrid offspring (the King of Holland’s bird-of-paradise). Specifically, we explore the extent to which the hybrid color phenotype is intermediate to the parental phenotypes.

Fig 3. Overview of the pipeline used to compare colors from the King, hybrid King of Holland’s, and Magnificent birds-of-paradise.

Fig 3

(a) Imaging. (i) A diagram of the hyperspectral imager in use. A bird specimen on a linear translation stage is moved across the single-line field-of-view of the hyperspectral imager (in green). (ii) Inset diagram indicating how light from the field-of-view of the camera is processed inside the spectrometer. Incident light (in green) is split into its constituent wavelengths by a diffraction grating and brightness recorded on a sensor. (iii) A diagram of a hyperspectral data-cube being generated. The width (W) and spectral resolution (Wavelength) of the resulting image are set by the resolution of the imager and software settings, but the height (H) of the image is set by the number of exposures selected by the user. (b) Sampling. Six specimens (3 shown) were imaged from a number of positions (4 shown). Sampling locations from each image for the “whole” patch are indicated by colored dots. These samples correspond to reflectance spectra shown in both (c) and (d). (c) Each sampled spectrum colored according to their appearance (to human eyes, brightened by 300% for plotting). (d) The same sampled spectra shown in (c), now colored according to the species from which the sample came, as shown in the key on the far left (Green: King bird-of-paradise, Blue: hybrid King of Holland’s bird-of-paradise, Orange: Magnificent bird-of-paradise). In both (c) and (d), the vertical dotted line indicates the lower bound of the hyperspectral imager’s sensitivity; reflectance values below this were extrapolated. (e) Embedding. These same spectral samples embedded in a visual system-independent color space. We generated PCA coordinates for the samples (20 dimensions, see main text and Methods), followed by dimensionality reduction using UMAP (see main text and Methods). Embedded points are colored according to their appearance (to human eyes, brightened by 300% for plotting). (f) Kernel density plot over the embedding shown in (e), with contour lines to indicate the (80% density) contribution of each species to the distribution of colors in the color space. Line color indicates species, as shown in the key to the left of (b). The data underlying this figure can be found at https://doi.org/10.5061/dryad.j0zpc86nf [99]. PCA, principal component analysis; UMAP, uniform manifold approximation and projection.

Step 1: Capturing hyperspectral images

The first step of the pipeline is capturing hyperspectral images. We collected hyperspectral images of King, Magnificent, and hybrid specimens (two male specimens each) in the following positions: ventral, dorsal, lateral (left flank), and lateral (right flank). In these images, the specimen’s body was parallel to the imaging plane. Additionally, we obtained a second image of the ventral position, this time with the specimen’s head angled approximately 30° toward the imager, for a total of five images of the body of each specimen. We also took hyperspectral images of the dorsal side of the tail of each specimen. Specimens were loaned by the American Museum of Natural History (New York, USA, Table A in S1 Appendix). We used a Resonon Pika NUV hyperspectral camera with a CoastalOpt UV-VIS-IR 60 mm lens (Jenoptik AG., Jena, Germany). Data collection and initial spatial calibration were conducted using SpectrononPro software (Resonon). We illuminated specimens in these images using an ultraviolet-rich and spectrally broadband light source (450W Xenon ozone free arc lamp: Newport Corp., Irvine, California, powered with a Newport OPS-A1000 power supply). This light source was housed in an enclosure (Newport 66921 arc lamp housing), the output of which was first filtered for infrared light by passing through a water filter filled with deionized water (Newport 6123NS liquid filter) and then directed vertically downward using an ultraviolet-reflective turning mirror (beam turning mirror: Newport 66246, beam turning mirror housing: Newport 66225). The light, enclosure, filter, and turning mirror were placed on a small rolling table so that the turning mirror could be placed as close to the lens of the hyperspectral imager as possible. This resulted in bright directional light, from approximately the same angle as the imager’s receptive field, cast over the entirety of the push-broom field of view of the hyperspectral imager (see Fig A in S1 Appendix). One limitation of this lighting set-up is that objects closer to the light source will appear artificially lighter. Therefore, improving the control and directionality of incident light is a goal for future work (see Discussion). Each image was calibrated spatially (i.e., we corrected for non-uniformity in lighting across the field of view of the imager) using SpectrononPro. This process involves collecting a “bright image” of a rectangular block of white PTFE that encompasses the entire field of view of the imager, which is used to account for differences in brightness across the field of view due to directional lighting. Each image also contained a Spectralon 99% reflectance standard (LabSphere, North Sutton, New Hampshire), which was used later to obtain spectrally calibrated reflectance images (see below, Step 2). All images were taken with the imager at the same height (90 cm) from the linear translation stage, so the scale is identical across images.

The Pika NUV camera works in a push-broom manner (Fig 3A), where a given exposure of the sensor generates one row of the data-cube at a time, in essence taking a line of 1,500 (Fig 3A, width) spectrophotometric readings simultaneously. By passing the subject across the field of view of the camera and capturing sequential exposures, the second spatial dimension of readings (Fig 3A, height) is generated. A linear translation stage is used to push the subject across the field of view at a controlled slow pace. We used SpectrononPro to operate the camera and stage, and to collect the data-cube. We collected exposures at 11 frames per second (each image took around 5.5 min to complete), and the average number of exposures was 3,600, resulting in images with a width of 1,500 pixels and an average height of 3,600 pixels, with a spectral dimension of 294 bands between 274 nm and 812 nm (each band is approximately 1.83 nm wide). Visual inspection of the data indicates that bands below 325 nm contain only noise, which means that in our configuration, this system is sensitive from 325 nm to 700 nm, encompassing almost the complete range of avian visual sensitivity (300 nm to 700 nm). These data-cubes were saved in a standard band-interleaved-by-line format, along with a header file which contains metadata. The data-cubes produced by SpectrononPro software averaged 3.2 gigabytes in size.

Hyperspectral data-cubes can be analyzed in a number of software packages/environments, including SpectrononPro, MATLAB (MathWorks, Natick, Massachusetts), Python (Python Software Foundation, Wilmington, DE; available at www.python.org), and ENVI (NV5 Geospatial Solutions, Broomfield, Colorado). In the current study, we used custom code to extract and further analyze sets of spectra from the hyperspectral images in Python (v3.7.4, see all subsequent steps of the pipeline), but one of the key advantages of hyperspectral imaging is that the data are amenable to both image-based and spectra-based color analyses. For example, it is possible to apply visual models to entire hyperspectral images to generate cone stimulation images that can then be analyzed using popular packages, including MICA/QCPA for ImageJ [2,5,69]. We include in our code an example of this. Alternatively, users may be interested only in one or a small number of wavelength bands of interest. These may be easily extracted and analyzed as individual images. Yet another approach might be to extract only a few spectra for further color analysis using the R package pavo [70]. We include in our code examples of this approach (and see section “Some King of Holland’s bird-of-paradise color patches that are spectrally intermediate would be indistinguishable from those of its parent species when modeled with avian color vision using only one spectrum per patch”).

Step 2: Extracting sample reflectance spectra from hyperspectral images

The second step of the pipeline is to define a spatial sampling scheme in which we select certain pixels (each corresponding to a reflectance spectrum) for further analysis. Sampling from the hyperspectral images allowed us to (i) control the computational complexity of the analysis; and (ii) reduce redundancy due to spatial autocorrelation (i.e., pixels tend to share the color properties of neighboring pixels). We first manually generated for each image a set of regions-of-interest indicating the locations of pixels within select body parts/patches: whole (the entire body), breast (the pectoral shield as per [71]), belly/vent (all of the ventral side of the bird, below the pectoral shield excluding the legs), shoulder, and back (including wings, see Figs K–O in S1 Appendix for details of the placement of these patches). To sample colors from each of these patches, we defined a uniform grid of sampling locations in each image, the spacing of which was chosen to approximately equalize the number of sampling locations in images of each patch within a specimen, and across all specimens (Fig 3B). This ensured that no species, specimen, image, or patch should be overrepresented in our analysis. Because the ventral view (with the bird’s long axis oriented with the image plane) did not well capture the iridescent breast patch coloration, for this patch we only sampled from the images obtained in the ventral angled position (with the bird rotated with its beak approximately 30° toward the camera, see Step 4 below, Discussion). Care should be taken in generating a sampling scheme; here we chose to equalize sampling across patch and specimen (so that each patch/specimen has the same number of samples), but this came at the cost of not having a uniform sampling density (samples per cm2, for instance), which may be more important for other research questions. In addition, we selected samples using a regular grid. This worked well for our purposes because the bird specimens did not contain patches with complex patterning (such as spots or stripes), but care should be taken when sampling in order to collect a fair representation of the colors present.

At each sampling location, we selected a square of 10 by 10 pixels and obtained the median of the spectra corresponding to these pixels as our sample for analysis (hereafter sample reflectance spectra, Fig 3C and 3D). Each of these sample reflectance spectra was spectrally calibrated (by division) with the median spectrum of the pixels corresponding to the Spectralon 99% reflectance standard in each image. We then interpolated spectra to 1 nm bands (from 325 nm to 700 nm) and smoothed the spectra by application of a Savitzky–Golay filter (window 45 nm, polynomial order 2). It is possible that interpolating spectra to a spectral resolution of greater than 1 nm would reduce computational complexity of the analyses, potentially at little cost. We note that the values of these smoothing parameters were chosen here to aid in stable extrapolation of reflectance values below 325 nm (where small amounts of noise around 325 nm produce large differences in the resulting extrapolation), but that the result is relatively extreme smoothing (see Figs 1, 3, and B and C in S1 Appendix). Future research should determine levels of smoothing appropriate for varied applications. We then extrapolated reflectance values in the wavelength range 300 to 325 nm by applying a cubic 1D monotonic smoothing spline fit to the range 325 nm to 700 nm (see Fig B in S1 Appendix for a figure contrasting “raw” sampled spectra with filtered and extrapolated versions, and see Fig C in S1 Appendix for a figure contrasting “raw” natural spectra with filtered and extrapolated versions). Because birds have limited sensitivity to light at wavelengths shorter than 325 nm, we anticipate that results for visual system-dependent analyses will be very robust to extrapolation. The end result is that for each patch of each specimen, we collected approximately 1,000 sample reflectance spectra for analysis in Step 3. At this point, a range of analyses on the sampled reflectance spectra can be undertaken (see, for example, [8]).

Step 3: Embedding sample reflectance spectra in low-dimensional spaces

The third step of the pipeline is to embed the sample reflectance spectra generated in Step 2 in a low-dimensional space (Fig 3E and 3F). The advantage to this is ease of visualization; it also serves as a gateway to other analyses [8,68]. Here, we demonstrate two different ways in which to embed the reflectance spectra: in a visual system-independent color space and in a visual system-dependent space. A visual system-independent approach takes full advantage of the spectral resolution provided by the hyperspectral images because it does not a priori select a subset of wavelengths of interest, instead using the variability in the data to inform the embedding (Fig 3E and 3F). It also makes no assumptions about a potential signal viewer. Because this approach requires embedding high-resolution reflectance spectra, it would not be possible to produce an identical color space using multispectral photography. A visual system-dependent approach allows for spectra to be represented in a way that is relevant to a signal receiver (reviewed in [14,15]). Here, we embedded spectra in a tetrahedral color space [17,6668]. The result of this embedding is similar but not identical to that obtained using multispectral photography (see details below).

Embedding in a visual system-independent color space

To embed the spectra in a low-dimensional, visual system-independent color space, we begin by applying principal component analysis (PCA, implemented using scikit-learn [72]) to our sampled spectra (one PCA per plumage patch). We subtracted the mean reflectance from each spectrum before applying PCA (see [8]). The calculated PCA coordinates (20 dimensions) are not easily interpretable since any 2D plot (for instance, axis 1 versus axis 2) may not fully capture the distances between the embedded colors. To reduce this dimensionality, we used uniform manifold approximation and projection (UMAP: see below and Fig E and F in S1 Appendix) in order to embed spectra in a lower dimensional space that attempts to maintain distances between PCA coordinates. Spectra that fall close together in this space will tend to be those with low distances between PCA coordinates, and those far apart will tend to have high distances. The result is a single, easily interpretable 2D plot of the distances among sampled spectra. Our goal here is to illustrate and explore the spectra rather than to conduct statistical analyses, but PCA coordinates are amenable to statistical approaches: for instance, one might look for statistical separation in the distributions of points of each species (e.g., using MANOVA).

Embedding in a visual system-dependent color space

It is often advantageous to estimate how an animal would perceive color signals. One way to accomplish this is to quantify cone catch values (i.e., an estimate of how an animal’s photoreceptors would be stimulated by a certain color/reflectance spectrum; for a review of these methods, see [14,15]). Here, to estimate how a bird-of-paradise might perceive the plumage colors of the King, Magnificent, and hybrid King of Holland’s birds-of-paradise, we embedded each of our sampled reflectance spectra in a tetrahedral model of avian vision [67,68] assuming ideal illuminant and background. Birds-of-paradise are presumed to have violet-sensitive (VS) visual systems [41], so we used visual pigment sensitivities for an average VS bird (extracted from the R package pavo [70]), as well as cone ratio values of (1,1,1:2) reported for Rock dove (Columba livia) [73] and a chromatic Weber fraction of 0.1.

For each sample reflectance spectrum, this analysis generates 4 cone catch values, one each for a bird-of-paradise’s violet-sensitive (VS), short (SWS), medium (MWS), and long (LWS) wavelength-sensitive cones. These values can be converted to 3D x, y, and z coordinates in tetrahedral color space for plotting and analysis [68]. The process here is identical to that used if a single reflectance spectrum is measured (as with a spectrophotometer); here, we have many more spectra because hyperspectral imaging allows for much more dense sample selection than would be practical with spectrophotometry. Representing reflectance data in a visual system-dependent color space is also possible with multispectral imaging, where specialized software has been developed to convert multispectral images into cone stimulation images [2,5]. However, the accuracy of data derived from hyperspectral images will be greater; multispectral imaging requires the user to make some assumptions when modeling the mapping between the sensitivity of the camera and the sensitivity of the visual system, but hyperspectral data provides (near) complete spectral information, so fewer assumptions are required.

Step 4: Combining hyperspectral data with a 3D digital model

The fourth step of the pipeline is optional, allowing users to generate a 3D digital model of the imaged specimen. We generated 3D models of each of our specimens using photogrammetric software (Metashape v2.0.2, Agisoft LLC, Saint Petersburg, Russia). We note that several free and open-source alternatives for photogrammetric software exist. 3D models are powerful because they combine color and morphometric data—and they are likely to become an increasingly common resource for animal color researchers [24,37]. Our specific aims for this step were twofold: first, to create a 3D model to accurately represent the specimen’s shape, and second, to estimate the angle (relative to the hyperspectral imager) of the surface of the bird specimen (hereafter surface normal) for any reflectance spectrum sampled from the hyperspectral images (Fig 3B). Angular information is useful because bird plumage is not Lambertian (i.e., reflectance is a function of viewing and lighting geometry) for both iridescent [74] and probably many non-iridescent colors [75]. Supplementing our 2D image with 3D information about the surface normal will allow us to inspect the dependence of measured color on 3D angle, at the level of the patch.

During imaging, we were cautious when manipulating the specimens due to their rarity and age. We custom designed and 3D printed small stands to hold the specimen, which were then affixed to a small rotating stage, allowing us to photograph the specimen from many angles (see Fig R in S1 Appendix). For each specimen, we collected two sets of images: one with the bird supine (lying on its back) and another with the bird prone (lying on its front). For each set of images, we collected approximately 120 photographs using a consumer Sony (Minato City, Tokyo, Japan) A7 DSLR camera under typical laboratory fluorescent lighting. We rotated the bird about 360 degrees in increments of 30 degrees, resulting in 12 photographs per rotation. Between rotations, we changed the elevation and angle of the camera to capture 10 such rings of 12 photographs (see Fig S in S1 Appendix). In some cases, we collected extra photographs (largest number 158) and these were also included in the analysis.

To generate 3D structure from multiple photographs, the position and angle of each of the photographs must be estimated from the images. Metashape calls this “alignment” and computes the alignment by correlating the position of features across images (see Fig S in S1 Appendix). To facilitate this process, we generated binary masks in the software that eliminated the stationary background and other unwanted features from the alignment process. With a set of aligned cameras, Metashape can generate a triangulated 3D model, which may then be textured by algorithmically extracting colors from the photographs that generated the model and placing these on the polygonal faces of the model (see Fig T in S1 Appendix and S1 Video). We followed this process for each specimen.

In order to relate information from the 3D models to the hyperspectral images, some registration between the 2 was required (see Fig 4). We imported RGB versions of our hyperspectral images into the photograph alignments generated above. We then marked on each hyperspectral image and each 3D model a set of points of known location (for example, the tip of the beak, the lores, and chin). This allowed us to align our hyperspectral image with the 3D model (treating the hyperspectral image as if it were another source photograph), making it possible to render the 3D model from the approximate point of view of the hyperspectral camera. We rendered images of the 3D model false-colored according to the angular surface normal in relation to the camera (see Fig 4C and 4D, and S1 Video).

Fig 4. Combining hyperspectral data with 3D digital models.

Fig 4

(a) Three renders (digital images) of one 3D model produced using our pipeline. The left image shows a color render of the 3D model. For the middle and right images, a simplified mesh is superimposed over the colored render of the model to illustrate the 3D information contained. (b) An RGB version of the ventral hyperspectral image of a Magnificent bird-of-paradise specimen. For each such hyperspectral image, we extracted 3D information from the 3D models to augment the color data. (c, d) Images illustrating the 3D information extracted from 3D models for each of our hyperspectral images. These images indicate the surface normal of the specimen relative to the camera in color from π to -π. The color indicates the horizontal component (roll, c) and vertical component (pitch, d) of the specimens’ surface normal relative to the camera, with values close to 0 signifying surfaces nearly parallel to the camera, and more extreme values indicating extreme angles relative to the camera. For roll (c), π indicates that the surface normal of the specimen is facing to the left of the camera and -π indicates a normal facing to the right. For pitch (d), π indicates a surface normal facing up relative to the camera, and -π indicates a surface normal facing downwards. We generated similar information for all hyperspectral images in our data set. (e) Scatterplot of samples of the breast patch across all specimens, where the x- and y-axes represent the horizontal and vertical components of the surface normal extracted from 3D the models for each sample. Points are colored according to their appearance to human eyes (brightened by 300% for plotting). Dotted lines across the y-axis show the bins of vertical angles plotted in (f). (f) A plot of the median (black lines) and median ± the median absolute deviation (shaded areas, with overlap in shaded areas indicated by darker shading) of the reflectance of breast patch samples across all specimens. The three lines represent samples binned by the vertical (pitch) component of the samples (see E). The data underlying this figure can be found at https://doi.org/10.5061/dryad.j0zpc86nf [99].

Dryad DOI

https://doi.org/10.5061/dryad.j0zpc86nf [99]

Results

Application of the pipeline to the birds-of-paradise

Here, in a single hyperspectral data-cube, we captured high-resolution spectral information: for every pixel in the data-cube, we recovered complete, high-resolution reflectance spectra from 325 nm to 700 nm. We simultaneously collected spatial information: at each band from 325 nm to 700 nm (sampling every ~2.5 nm), we obtained a complete photograph/image of each specimen. We have therefore illustrated the utility of hyperspectral imaging as a tool to simultaneously produce data with high spectral and high spatial resolution. In the sections below, we summarize our key findings about the plumage colors of the King, Magnificent, and hybrid King of Holland’s birds-of-paradise.

(i) Analyzing plumage colors with a visual system-independent approach

Fig 5 shows a visual system-independent embedding of colors sampled from patches of the King, Magnificent, and hybrid King of Holland’s birds-of-paradise. Shown are low-dimensional embeddings of the PCA coordinates of the sampled spectra. The top row of Fig 5A–5E illustrates the distribution of colors of each patch across King, Magnificent, and hybrid specimens. The contribution of each species to the color distribution for each patch can be seen in the middle row (Fig 5F–5J). Here, approximately 80% of a species’ contribution is enclosed by a density contour with color indicating King, Magnificent, or hybrid. The spectra corresponding to the most typical colors for the King, Magnificent, and hybrid King of Holland’s patches are plotted in the bottom row (Fig 5K–5N).

Fig 5. Visual system-independent embedding of sampled spectra for each patch.

Fig 5

Top row: Icons indicate the patch represented in each column of the figure. (a–e) Plot of low-dimensional embedding (UMAP embedding of PCA coordinates) of the sample reflectance spectra, colored according to their appearance to human eyes (brightened by 300% for plotting). Note that the embeddings were computed separately for each patch; each column shows a different color space, so they should not be directly compared. (f–j) Kernel density contour plots indicating the 80% density contribution of each species to the color spaces in a–e. Line and point colors correspond to the different species as indicated in the key at the right of the figure. Areas of overlap indicate that the colors from more than one species occupy the enclosed area of color space. Points indicate the area of highest kernel density for each of the species: these areas represent the most common spectra for each species for that patch. (k–n) Plot of the 50 closest (in embedded space) spectra to the areas of highest kernel density for each patch and species (points in f–j). Bold line indicates median colored by species, shaded areas indicate median ± the median absolute deviation, and shaded colors represent approximate color of the samples (to human eyes). In general, the plumage colors measured in this study reflected very low levels of ultraviolet light. The data underlying this figure can be found at https://doi.org/10.5061/dryad.j0zpc86nf [99]. PCA, principal component analysis; UMAP, uniform manifold approximation and projection.

The hybrid King of Holland’s bird-of-paradise has plumage colors that are generally intermediate between its parent species, but it more closely resembles one or the other parent in different patches

Collectively, the King, Magnificent, and hybrid King of Holland’s birds-of-paradise contain plumage colors that are red, orange, brown, yellow, green, and white (Fig 5A). In general, the colors measured in this study reflected very low levels of ultraviolet light (Fig 5K–5N). Examining the distribution of colors for the King, Magnificent, and hybrid King of Holland’s birds-of-paradise separately (Fig 5F) reveals that the hybrid’s colors are—in general—intermediate between those of the parent species. Closer examination of individual patches sheds light on this. The breast color of the King and Magnificent is green, with the hybrid intermediate (in terms of λmax, the wavelength of peak reflectance) between these greens (Fig 5B, 5G, and 5K). The difference is subtle—in fact the green colors are likely to be indistinguishable from one another to a bird eye (see next section)—but after inspecting the spectral shape of the typical breast colors for each of King, Magnificent, and hybrid (Fig 5K), it is clear that the hybrid green has λmax intermediate to those of the parent species. Interpreting the magnitudes of the peak reflectance for the breast patch across the specimens is more difficult. This is because some of the hyperspectral images may by chance contain more samples that are geometrically closer to the angle of peak reflectance for this iridescent patch (but see Discussion). The belly/vent color of the King is white, while the belly/vent color of the Magnificent is brown (Fig 5C and 5H). The hybrid’s belly/vent is brown, more closely resembling that of the Magnificent (Fig 5H). This is evidenced by the spectra for the most typical colors (Fig 5L), with the hybrid’s light brown more similar in shape to the dark brown of the Magnificent. The shoulder color of the King is red/orange, while the shoulder color of the Magnificent is yellow (Fig 5D). The hybrid’s shoulder color is orange and more closely resembles the red/orange color of the King (Fig 5D and 5I), but is spectrally intermediate to the parent species. The back of the King is red/orange, while the back of the Magnificent is yellow (Fig 5E and 5J). The hybrid’s back is intermediate, with an orange-red color (Fig 5J), but it more closely resembles the King bird-of-paradise, a result that is also clear when examining the spectra for typical colors (Fig 5N).

Our results show that the degree to which the hybrid’s color is intermediate between its parent species can differ among patches, and intermediate phenotypes can result in patches colored by both pigmentary and structural mechanisms. Our approach also highlights the remarkable color diversity within these generally “uniformly colored” patches—information that is overlooked using spectrophotometry. For example, by inspecting the contributions of the King, Magnificent, and hybrid King of Holland’s birds-of-paradise (Fig 5G–5J) to the color distributions (Fig 5B–5E) for each patch, we see that in all cases patches contain a variety of colors, such that only one or a small number of spectrophotometric readings would likely inadequately describe the variety of spectral shapes present.

An important disclaimer is that these inferences are based on spectral shape only. They emerge from the ways in which these spectra are embedded in a visual system-independent space with relation only to physical properties of the spectra—not to animal color vision. Therefore, many of the differences we have described in color may not be perceptually meaningful, though they may hint at important mechanistic or developmental color production processes in the hybrid and its parents. To consider perceptually relevant questions in the context of signaling, we turn to visual system-dependent approaches in the next section.

(ii) Analyzing plumage colors with a visual system-dependent approach

Fig 6 shows sampled spectra from King, Magnificent, and hybrid King of Holland’s patches embedded in tetrahedral avian color space (plotted in 2D for convenience), representative of a bird-of-paradise with a VS color vision system [41]. In the tetrahedral color space, an embedded point’s proximity to each of the four vertices corresponds to the degree of stimulation of each of the four color cones in typical avian eyes [68]. Thus, a color that we predict would appear green to a bird would fall near the MWS vertex of the tetrahedron. However, the space accounts only for the hue and saturation of colors, not perceived brightness (indeed, black and white fall in the same location, at the center of the tetrahedron). Colors located farther apart in this space are more different in hue and/or saturation according to models of avian vision, but the space is not perceptually uniform (i.e., the same distance between points in different parts of the space do not correspond with equal discriminability [15], see Fig Q in S1 Appendix for a plot using the receptor noise limited color space). As in Fig 5, we show the distribution of colors in the upper row of Fig 6 (Fig 6A–6E). Here, an expanded section of the tetrahedral color space is shown, with points plotted for each sample colored according to their appearance (to human eyes). Panels Fig 6F–6J highlight the contribution of the King, Magnificent, and hybrid King of Holland’s birds-of-paradise to the distribution of colors in the color space. The contour lines for each species enclose their most typical colors. Here, we have used an avian tetrahedral color space model, but—as with reflectance spectra obtained from spectrophotometry or images obtained from multispectral photography—the spectral data can be modeled for a great variety of animal visual systems for which the photoreceptor sensitivities are known [2,15,70].

Fig 6. Tetrahedral avian color space embedding of sampled spectra for each patch.

Fig 6

Top row: Icons indicate the patch represented in each column of the figure. Left top: 3D tetrahedron with icon indicating the view shown in all other panels; where we view the 3D tetrahedron down through the VS axis since none of the sampled colors reflect significant ultraviolet/violet light. Left bottom: 2D plot of embedded samples colored by species (according to key on far right) plotted in avian tetrahedral color space. The location of points indicates their relative stimulation of the photoreceptors in bird-of-paradise eyes (top: MWS, left: SWS, right: LWS). The black rectangle indicates the expanded section in each of the other panels. (a–e) Sampled spectra plotted for each patch colored according to their appearance (to human eyes, brightened by 300% for plotting) in the expanded section of the avian tetrahedral color space. (f–j) Kernel density contour plots of the same data, with contours (80% density) indicating the contribution of each species to the distribution of spectra in the color space. Line colors correspond to the different species as indicated in the key at the right of the figure. The data underlying this figure can be found at https://doi.org/10.5061/dryad.j0zpc86nf [99]. LWS, long wavelength-sensitive; MWS medium wavelength-sensitive; SWS, short wavelength-sensitive; VS, violet-sensitive.

The overall colors of the King, Magnificent, and hybrid are primarily constrained to the portions of tetrahedral color space between the achromatic point (center) and the MWS vertex, and between the achromatic point and the LWS vertex (Fig 6 and 6A). This is expected given that the predominant plumage colors are green, yellow, reddish-brown, and white. The breast colors of the King, Magnificent, and hybrid are largely overlapping in tetrahedral color space (Fig 6B and 6G) and the same is generally true for the belly/vent colors (Fig 6C and 6H; recall that hue and saturation but not brightness are encoded in this space, so buff white and brown can fall in similar locations of color space). For the shoulder and back patches, the King bird-of-paradise occupies areas of tetrahedral space that are more saturated, and slightly redder in hue than the hybrid, and the Magnificent bird-of-paradise occupies a more tightly concentrated area of yellow hue (Fig 6D, 6E, 6I, and 6J). In general, embedded points are more concentrated and overlapping than in the visual system-independent approach, making patterns of occupancy for the breast and belly/vent patches much harder to distinguish.

Some King of Hollands bird-of-paradise color patches that are spectrally intermediate would be indistinguishable from those of its parent species when modeled with avian color vision using only one spectrum per patch

In order to explore whether differences between King, Magnificent, and hybrid King of Holland’s plumage patches may be perceptually relevant to avian viewers, we applied the Vorobyev and Osorio [73] receptor noise model to generate chromatic contrasts (dS) and we generated luminance contrasts (dL, implemented in [70]). These contrasts compare two colors in terms of hue/saturation and brightness respectively, and—in general—contrast values above one are expected to be discriminable by a given visual system [76]. A caveat of this approach is that a number of important parameters (cone density ratios, Weber fractions) must be assumed, the choice of which strongly influences calculated contrasts [76]. Due to these assumptions, and because the models assume discrimination at the limit of physiological ability under ideal conditions, more conservative threshold contrast values of two to four may be more appropriate when determining whether two colors are discriminable [77,78]. For simplicity, we calculated contrasts using the median spectrum for each patch on each specimen, and we considered two colors to be discriminable if the contrast value was ≥ 3 (see Fig I in S1 Appendix). The breast patch is not likely to be discriminable (in terms of chromatic or luminance contrast) among the King, Magnificent, and hybrid birds-of-paradise. The belly/vent patch is likely to be discriminable between the hybrid and King, and between the Magnificent and King, on the basis of luminance contrasts. However, the belly/vent patch is likely to be indiscriminable between the hybrid and Magnificent. For the shoulder and back chromatic contrasts, the King and hybrid are likely discriminable from the Magnificent, but unlikely to be discriminable from one another. On the basis of luminance contrasts, the shoulder and back patches of the hybrid are likely discriminable from both the King and Magnificent (see Fig I in S1 Appendix).

While a visual system-independent approach uncovered some differences between colors, for instance in breast coloration, visual models indicate that not all of these differences are perceptually relevant (when assessed using a single spectrum for each patch, but see Fig Q in S1 Appendix). This analysis highlights the importance of tailoring the approach (visual system-independent versus visual system-dependent) to the scientific question at hand. If we were interested in the evolution of pigmentation in the context of signaling behavior, a visual system-independent approach may produce irrelevant or even misleading results. In contrast, if we were interested in the physical or optical properties of pigmentation irrespective of visual system, the situation would be reversed.

(iii) Analyzing the color of bird-of-paradise tail feathers

Many birds-of-paradise have unusual tail feathers. The King, Magnificent, and hybrid King of Holland’s birds-of-paradise each have a modified central pair of iridescent tail feather “wires” that appear to have barbs only on the outside edge [40]. These modified tail wires are variably curved; the tail of the King bird-of-paradise resembles a tightly wound green spatula or disk, while the tail of the Magnificent bird-of-paradise is sickle-shaped and teal (Fig 7A). The hybrid bird-of-paradise tail appears to be intermediate to these in both color and shape (Fig 7A). Using quantitative methods, we explored the extent to which hybrid tail color is intermediate to the parent species (Fig 7B–7D). We extracted samples from hyperspectral images of the dorsal side of the tails of each species (approximately 500 samples per tail feather per specimen). We then calculated PCA coordinates for all samples before using UMAP to embed the spectral distances in a visual system-independent color space (Fig 7B and 7C). The aquamarine color of the barbs of the tail feathers in the hybrid King of Holland’s bird-of-paradise appears to be intermediate to the emerald green of the King bird-of-paradise and the teal shade of the Magnificent bird-of-paradise. This observation was supported by sampling and embedding using the visual space-independent methods described above, and inspection of the sampled spectra (Fig 7B and 7C). The reflectance of Magnificent bird-of-paradise tails shows an ultraviolet-blue spectrum, while the King bird-of-paradise shows a saturated green, with the hybrid appearing to be a mix of the two (Fig 7D). We also explored the striking shape of the tail feathers and whether a simple model could replicate the curvature of the rachis of each species. We found that a simple model of bias in the growth of sections of the rachis during feather growth replicated the tail shapes, and that the hybrid has bias generally intermediate to the King and Magnificent birds-of-paradise (see Fig U in S1 Appendix).

Fig 7. Analyzing the shapes and colors of bird-of-paradise tails.

Fig 7

The modified tail feathers (tail wires) of the hybrid King of Holland’s bird-of-paradise appear to be morphologically intermediate to those of the King and Magnificent birds-of-paradise. (a) Overlaid photographs of the dorsal side of the central modified tail feathers from each of the species, as indicated by the location of the species icons. Photos: David Ocampo. (b) A plot of a UMAP embedding of PCA coordinates for spectra sampled from the pennaceous parts of each species’ tails colored according to their appearance (to human eyes, brightened by 300% for plotting). (c) Kernel density estimates (80%) showing the contribution of each species to the color space shown in panel (b). Colors indicate species as per the species icons (a). Points indicate manually selected points of interest for plotting in panel (d). (d) Plots of the reflectance of the closest 20 spectra to each point of interest shown in panel (c). Bold lines indicate the median, shaded area indicates the median absolute deviation, and color indicates species. The data underlying this figure can be found at https://doi.org/10.5061/dryad.j0zpc86nf [99]. PCA, principal component analysis; UMAP, uniform manifold approximation and projection.

Discussion

The hyperspectral imaging pipeline we presented here provides the tools necessary to harness the power of off-the-shelf hyperspectral imagers in animal color research. We have demonstrated how hyperspectral imagers can overcome the spatial limitations of spectrophotometers and the spectral limitations of multispectral photography to provide a versatile and powerful tool for quantifying animal coloration. We applied visual system-independent and visual system-dependent methods to quantify the diversity of plumage colors present in a rare hybrid and its parent species—and to explore the extent to which the hybrid’s plumage colors are intermediate composites of those of its parent species.

The plumage colors of the hybrid are generally intermediate to those of the King and Magnificent birds-of-paradise (Figs 5A, 5F, 6A and 6F). However, the degree of intermediacy appears to differ by patch. The back and shoulder of the hybrid are more similar to the King bird-of-paradise (Fig 5D, 5E, 5I and 5J), whereas the belly/vent is more similar to the Magnificent bird-of-paradise (Fig 5C and 5H). In both the breast and tail, we found that the hybrid was strikingly intermediate: this is most clear when inspecting the spectral shapes of the reflectance of tail feathers (Fig 7D). The hybrid’s plumage patches span the gamut of avian color production from carotenoid and melanin pigments to structural coloration, so differences in intermediacy among patches may reflect different genetic and developmental effects of hybridization across coloration mechanisms. By applying avian visual models, we show that some but not all of the differences in color among the hybrid, King, and Magnificent birds-of-paradise are likely to be perceptually discriminable to birds and therefore potentially behaviorally relevant (see Fig I in S1 Appendix, at least when modeled using only a single spectrum per patch, also see Fig Q in S1 Appendix). This observation highlights the importance of tailoring the approach to the scientific question (i.e., modeling with respect to visual signal receiver or not), which is possible even after data collection due to the versatility of hyperspectral images. Further, we found that in general, patches are not homogenous: summarizing plumage using single or average color spectra may obscure substantial differences within patches. Overall, we showed that hyperspectral imaging is a feasible, efficient, and effective method for collecting detailed color information from museum specimens. Animal color research need not be limited by the trade-off between spectral and spatial resolution, and hyperspectral imaging represents a promising avenue for future researchers.

Hyperspectral imaging: Future challenges and opportunities

Moving forward, several challenges and opportunities remain for developing methods for hyperspectral imaging of animal coloration. Angle-dependent (iridescent or any non-Lambertian) colors are difficult to quantify [79]. For these colors, the angle of observation, the angle of the specimen’s surface (surface normal), and the angle/s of incident illumination (together termed the viewing geometry) determine in large part the perceived (and measured) color [74]. In spectrophotometry, this problem is often approached either by integrating across all possible viewing geometries (using an integrating sphere) or by systematically taking many measurements using various viewing geometries [80]. Generally, in multispectral imaging (and in the current paper, with hyperspectral imaging) this problem is approached by diversifying the viewing geometries used in color measurement by taking multiple measurements of the same patch (for instance, by imaging the sample at two or more angles e.g., [41,81]). An extension of this approach is to more precisely understand the viewing geometries within each image. Any image of a 3D (non-flat) iridescent or angle-dependent colored sample will contain a variety of viewing geometries because the surface normal of the sample is not uniform. In this paper, we piloted a method to recover information about the viewing geometries contained within an image through registration with a separately generated 3D model. In this way, we supplemented our hyperspectral images with surface normal information to better explore the angle dependency of the breast patch in the three birds-of-paradise (see Fig 4). This approach could easily be adapted for use with multispectral images. A remaining goal is to better control, diversify, and record the direction of illumination for color measurements in hyperspectral images of animal colors. Kim and colleagues [37] introduced a related approach: the authors collected 3D surface information concurrent with hyperspectral imaging, allowing registration between the 3D structure of a bird specimen while color measurements were taken. However, this approach required specialized and highly customized equipment [37]. Further development of these, or similar, approaches will allow animal color researchers to move away from methods that implicitly assume that all animal colors are Lambertian (perfectly diffusely reflective) and to thereby assess animal colors in a more realistic way.

Animal coloration is often patchy (i.e., a given animal’s color is distributed across the body in patches), which complicates analysis of overall color difference and similarity between individuals or species. In general, researchers define color patches a priori, allowing them to compare homologous locations—an approach we took here. It is becoming possible, however, to generate patches for animal colors automatically or semi-automatically. For example, powerful clustering algorithms are being developed for quantifying animal coloration patterns [2]; other algorithms are being used to segregate a sample from the background [82,83] and to find physically homologous body locations in images of animals [84,85]. It seems likely that soon some combination of these efforts will result in a robust means by which to identify discrete patches in an objective and automated way, free of human biases.

At present, hyperspectral imaging typically has relatively poor temporal resolution (Table 1). Hyperspectral imaging of live animals moving in the wild (at least with high spectral and spatial resolution, and broad wavelength sensitivity) is not yet practical because images take minutes to generate. To some extent, this problem is also true of multispectral photography, although a study recently introduced the use of a beam splitting mirror to photograph a scene using ultraviolet- and visible light-sensitive cameras simultaneously (typically at framerates up to 30 per second), allowing the collection of multispectral video [7]. Both hyperspectral imaging and multispectral photography may benefit from better integration of imaging with 3D modeling and animation, where dynamic animal motion may be introduced and investigated in silico ([24,86]). For now, hyperspectral imaging for animal coloration as implemented here may be largely limited to museum specimens. However, museum specimens do not always perfectly represent the colors of live animals. For example, the physical posture for the specimen chosen during preparation can strongly influence the colors visible on the animal surface. Further, as specimens age, their color can change or fade, especially if the specimen is exposed to sunlight [3]. Lastly, complete metadata is often unavailable for historical specimens, sometimes limiting their use for study. Overall, hyperspectral imaging is currently primed for imaging animal specimens, but in situ animal color measurements will have to wait for advances in hardware (such as video-speed capture) to allow rapid collection of full-frame hyperspectral data with high spatial and spectral resolution and broad wavelength sensitivity.

Although our study focused on plumage colors, hyperspectral imaging holds great promise for the study of animal color in diverse invertebrate and vertebrate taxa, including butterflies [28,34], beetles [30], spiders [29], and cephalopods [32], to name a few. Hyperspectral imaging is likely to yield new insights about the evolution of color and pattern in bird eggs, bird nests, fruit, and flowers—and could reveal rich details about animal color in the contexts of urbanization and climate change. Similar insights might be gleaned from spectrophotometry and multispectral imaging, too—but only hyperspectral imaging provides high-resolution spectral and spatial data simultaneously. Hyperspectral imaging systems remain expensive. Resonon is currently selling its ultraviolet-sensitive imaging system for about $55,000 USD, and the light source and lens used here increase the total cost to around $75,000 USD. However, we expect that costs will continue to decrease in the future. As hyperspectral imaging becomes increasingly mainstream in animal coloration research, we hope that researchers will collaborate to establish shared tools, software, and “best practices,” all of which could help to fuel large-scale efforts to digitize and characterize diverse animal specimens. Future software should ideally be freely available, highly computationally efficient (and amenable to cloud computing), and inter-operable with existing tools for multispectral image analysis (such as MICA/QCPA [2,5]) and spectral analysis (such as the R package pavo [70]).

Insights into the plumage color of a hybrid bird-of-paradise

We found evidence of intermediacy (in terms of spectral properties) across patches of color generated by both pigmentary (back and shoulder, Fig 5M and 5N) and structural (breast and tail, Figs 5K and 7D) mechanisms in the hybrid. In birds, yellow and red colors are typically generated by carotenoids, which are metabolized from dietary sources for deposition in the integument [87,88]. Research has shown that intermediate orange pigmentation in the hybrid offspring of red and yellow birds can correlate with greater concentrations of incompletely oxidized carotenoid products [8992]. In addition, the genetic basis of carotenoid coloration in birds is becoming clearer [87,88,9395]. Future work might explore intermediacy in carotenoid coloration in the hybrid bird-of-paradise through characterization of the carotenoid composition of their feathers as well as investigation of hybrid genetics at relevant loci.

We also described intermediacy in the color of the hybrid’s structurally colored breast and tail patches (Figs 5K and 7D), a finding that contrasts with those of Barrera-Guzmán and colleagues [64] and Eliason and colleagues [65], who found —in other avian taxa— non-intermediate coloration in patches of structural coloration in hybrid offspring. Here, we found that the structural colors from both parent species instead gave rise to intermediate hybrid breast and tail colors. Some nanostructures in the Magnificent bird-of-paradise have been investigated; for example, melanosomes in the breast patch are known to be solid (as opposed to air-filled, though they appear to be porous [96], as cited in [97]), round and rod-like, about 130 nm in diameter, and arranged in double layers apparently unique to Paradisaeidae [97]. Less is known about the melanosomes or their arrangement in the King bird-of-paradise, though there are hints that its melanosomes are also round and solid [97]. Whether and which nano- or micro-structural parameters of the hybrid breast and tail patches are intermediate to that of the parents could be investigated directly using microscopy.

Mysteries about the hybrid King of Holland’s bird-of-paradise remain. For example, the Magnificent bird-of-paradise is strongly fluorescent under a blacklight (pers. comm. Glenn Seeholzer). We confirmed this: fluorescence is especially prominent in the highly reflective yellow shoulder/cape patch (see Fig V in S1 Appendix). Neither the King nor the hybrid bird-of-paradise shows significant fluorescence: whatever the mechanism, we do not see evidence of intermediacy in the hybrid. We note that strong fluorescence can complicate the measurement of animal color; the light produced by fluorescence can introduce errors in inferred reflectance (pers. comm. Jolyon Troscianko). Exactly what mechanism generates fluorescence in the Magnificent bird-of-paradise, and why it appears absent in the hybrid, could be investigated by sectioning and imaging feather samples—and by testing for any influence of specimen preparation methods [98]. Further characterization of the fluorescence, including specifics of the excitation and emission wavelengths, might also be studied using a hyperspectral imager in concert with spectral filters and/or a monochromator. Another question relates to bare-part coloration across these birds. One disadvantage of avian museum specimens is that bare-skin colored parts do not retain their colors from life. Both the King and Magnificent birds-of-paradise have dark blue legs in life [40]. Additionally, the inside of the mouth and the tongue of the Magnificent bird-of-paradise are a striking pale green, while the inside of the King’s mouth is pale aqua-green [40]. Since the hybrid is only known from specimens, how and whether these bare-part colors compare may never be known unless live hybrids can be found.

In summary, we have shown that hyperspectral imaging is becoming a practical and advantageous alternative to spectrophotometry and multispectral photography for measuring animal colors. Hyperspectral images have high spectral and spatial resolution, allowing for detailed visual system-independent and visual system-dependent analyses of animal colors. Using the example of the birds-of-paradise, we provide a guide for capturing and analyzing hyperspectral data, which we hope other researchers will adopt and update. We also provide example code showing how hyperspectral data can be easily incorporated into existing spectra- and image-based color analysis toolkits (pavo: [70], MICA/QCPA: [2,5]). Overall, hyperspectral imaging may help galvanize high-throughput projects that allow for a more complete description of animal color, thereby providing new insights into the mechanistic, evolutionary, and developmental processes that drive phenotypic diversity.

Supporting information

S1 Appendix. Supplementary text and figures (Tables A and B and Figs A–V).

Raw data and code for all supplementary figures are available (see the Data availability section) (.PDF).

(PDF)

pbio.3002867.s001.pdf (10.9MB, pdf)
S1 Video. Video illustrating 3D models of all specimens (2 King bird-of-paradise, 2 Magnificent bird-of-paradise, and 2 hybrid King of Holland’s bird-of-paradise).

In the first half of the video, the models are colored according to the RGB images used to generate the models (see main text and S1 Appendix). In the second half of the video, models are colored according to the surface normals of the visible polygons in the video. Polygons facing left relative to the camera are teal, while those facing right are magenta. Polygons facing up relative to the camera are darker, while those facing down are lighter. Models are to scale. The sphere in the bottom left has diameter 2 cm, and the cube in the bottom right has edge length 2 cm. (.MP4).

(MP4)

Download video file (8.8MB, mp4)

Acknowledgments

We thank colleagues at Resonon for detailed technical discussions about hyperspectral imaging and for sharing equipment. We are grateful to the American Museum of Natural History (New York City) for loaning us the birds-of-paradise specimens; we especially thank Paul Sweet, Joel Cracraft, Brian Smith, Tom Trombone, and Bentley Bird. We appreciated discussions with Glenn Seeholzer, who observed fluorescence in the plumage of the Magnificent bird-of-paradise. We thank David Ocampo for photographing the specimens, Derya Akkaynak, Rick Prum, Roger Hanlon, Innes Cuthill, and Henry Knowles for early discussions about hyperspectral imaging, and Audrey Miller, Klara Nordén, Jarome Ali, Sarah Solie, Harold Eyster, Rosalyn Price-Waldman, and Lazarena Lazarova for constructive feedback.

Abbreviations

LWS

long wavelength-sensitive

MWS

medium wavelength-sensitive

PCA

principal component analysis

SWS

short wavelength-sensitive

VS

violet-sensitive

UMAP

uniform manifold approximation and projection

Data Availability

All hyperspectral images and all code for this analysis are available on Dryad (DOI: 10.5061/dryad.j0zpc86nf) and GitHub (https://github.com/bghogan/HyperBirdOfParadise), respectively. An archived copy of the code is stored on Zenodo (DOI: 10.5281/zenodo.13749562).

Funding Statement

This work was supported by a Packard Fellowship for Science and Engineering (to MCS), an Alfred P. Sloan Research Fellowship (to MCS), a Polymaths Award from Schmidt Sciences (to MCS), National Science Foundation grant 2029538 (to MCS), the High Meadows Environmental Institute, and general funds from Princeton University (to MCS). The funders had no role in study design, data collection and analysis, decision to publish, or preparation of the manuscript.

References

  • 1.Cuthill IC, Allen WL, Arbuckle K, Caspers B, Chaplin G, Hauber ME, et al. The biology of color. Science. 2017. Aug 4;357(6350):eaan0221. doi: 10.1126/science.aan0221 [DOI] [PubMed] [Google Scholar]
  • 2.van den Berg CP, Troscianko J, Endler JA, Marshall NJ, Cheney KL. Quantitative colour pattern analysis (QCPA): A comprehensive framework for the analysis of colour patterns in nature. Methods Ecol Evol. 2020. Dec 2;11(2):316–32. [Google Scholar]
  • 3.Burns KJ, McGraw KJ, Shultz AJ, Stoddard MC, Thomas DB. Advanced Methods for Studying Pigments and Coloration Using Avian Specimens. In: The Extended Specimen: Emerging Frontiers in Collections-based Ornithological Research. CRC Press; 2017. [Google Scholar]
  • 4.Stevens M, Párraga CA, Cuthill IC, Partridge JC, Troscianko TS. Using digital photography to study animal coloration. Biol J Linn Soc. 2007. Jan 31;90(2):211–37. [Google Scholar]
  • 5.Troscianko J, Stevens M. Image calibration and analysis toolbox–a free software suite for objectively measuring reflectance, colour and pattern. Methods Ecol Evol. 2015. Jul 8;6(11):1320–31. doi: 10.1111/2041-210X.12439 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 6.Bennett ATD, Théry M. Avian color vision and coloration: multidisciplinary evolutionary biology. Am Nat. 2007;169(S1):S1–S6. [Google Scholar]
  • 7.Vasas V, Lowell MC, Villa J, Jamison QD, Siegle AG, Katta PKR, et al. Recording animal-view videos of the natural world using a novel camera system and software package. PLOS Biol. 2024. Jan 23;22(1):e3002444. doi: 10.1371/journal.pbio.3002444 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 8.Cuthill IC, Bennett ATD, Partridge JC, Maier EJ. Plumage reflectance and the objective assessment of avian sexual dichromatism. Am Nat. 1999. Feb;153(2):183–200. doi: 10.1086/303160 [DOI] [PubMed] [Google Scholar]
  • 9.Mason NA, Bowie RCK. Plumage patterns: Ecological functions, evolutionary origins, and advances in quantification. The Auk. 2020. Sep 16;137(4):ukaa060. [Google Scholar]
  • 10.Eaton MD. Human vision fails to distinguish widespread sexual dichromatism among sexually “monochromatic” birds. Proc Natl Acad Sci U S A. 2005. Aug 2;102(31):10942–6. doi: 10.1073/pnas.0501891102 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 11.Andersson S, Prager M. Quantifying colors. In: Bird coloration. Cambridge, Mass: Harvard University Press; 2006. p. 41–89. [Google Scholar]
  • 12.Tedore C, Nilsson DE. Avian UV vision enhances leaf surface contrasts in forest environments. Nat Commun. 2019. Jan 22;10(1):238. doi: 10.1038/s41467-018-08142-5 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 13.Rosenthal GG. Spatiotemporal dimensions of visual signals in animal communication. Annu Rev Ecol Evol Syst. 2007. Dec;38(1):155–78. [Google Scholar]
  • 14.Kemp DJ, Herberstein ME, Fleishman LJ, Endler JA, Bennett ATD, Dyer AG, et al. An integrative framework for the appraisal of coloration in nature. Am Nat. 2015. Jun 1;185(6):705–24. doi: 10.1086/681021 [DOI] [PubMed] [Google Scholar]
  • 15.Renoult JP, Kelber A, Schaefer HM. Colour spaces in ecology and evolutionary biology. Biol Rev. 2017. Oct 15;92(1):292–315. doi: 10.1111/brv.12230 [DOI] [PubMed] [Google Scholar]
  • 16.Foster DH, Amano K. Hyperspectral imaging in color vision research: tutorial. JOSA A. 2019. Apr;36(4):606–27. [DOI] [PubMed] [Google Scholar]
  • 17.Burkhardt D. UV vision: a bird’s eye view of feathers. J Comp Physiol A. 1989. Nov 1;164(6):787–96. [Google Scholar]
  • 18.Hanley D, Grim T, Cassey P, Hauber ME. Not so colourful after all: eggshell pigments constrain avian eggshell colour space. Biol Lett. 2015. May 1;11(5):20150087. doi: 10.1098/rsbl.2015.0087 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 19.Hill GE, McGraw KJ. Bird coloration: Mechanisms and measurements. Cambridge, Mass: Harvard University Press; 2006. [Google Scholar]
  • 20.Medina I, Newton E, Kearney MR, Mulder RA, Porter WP, Stuart-Fox D. Reflection of near-infrared light confers thermal protection in birds. Nat Commun. 2018. Sep 6;9(1):3610. doi: 10.1038/s41467-018-05898-8 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 21.Wisocki PA, Kennelly P, Rojas Rivera I, Cassey P, Burkey ML, Hanley D. The global distribution of avian eggshell colours suggest a thermoregulatory benefit of darker pigmentation. Nat Ecol Evol. 2020. Jan;4(1):148–55. doi: 10.1038/s41559-019-1003-2 [DOI] [PubMed] [Google Scholar]
  • 22.Pannkuk EL, Siefferman LM, Butts JA. Colour phases of the eastern screech owl: a comparison of biomechanical variables of body contour feathers. Funct Ecol. 2010. Feb 26;24(2):347–53. [Google Scholar]
  • 23.Stoddard MC, Osorio D. Animal coloration patterns: Linking spatial vision to quantitative analysis. Am Nat. 2019. Feb;193(2):164–86. doi: 10.1086/701300 [DOI] [PubMed] [Google Scholar]
  • 24.Miller AE, Hogan BG, Stoddard MC. Color in motion: Generating 3-dimensional multispectral models to study dynamic visual signals in animals. Front Ecol Evol [Internet]. 2022. [cited 2022 Dec 16];10. Available from: https://www.frontiersin.org/articles/10.3389/fevo.2022.983369. [Google Scholar]
  • 25.Grahn HF, Geladi P, editors. Techniques and applications of hyperspectral image analysis [Internet]. Chichester, UK: John Wiley & Sons, Ltd; 2007. [cited 2022 Feb 21]. Available from: http://doi.wiley.com/10.1002/9780470010884. [Google Scholar]
  • 26.Khan MJ, Khan HS, Yousaf A, Khurshid K, Abbas A. Modern trends in hyperspectral image analysis: a review. IEEE Access. 2018. Mar 12;6:14118–29. [Google Scholar]
  • 27.Chiao CC, Vorobyev M, Cronin TW, Osorio D. Spectral tuning of dichromats to natural scenes. Vision Res. 2000. Jan 1;40(23):3257–71. doi: 10.1016/s0042-6989(00)00156-5 [DOI] [PubMed] [Google Scholar]
  • 28.Medina JM, Nascimento SMC, Vukusic P. Hyperspectral optical imaging of two different species of lepidoptera. Nanoscale Res Lett. 2011. May 4;6(1):369. doi: 10.1186/1556-276X-6-369 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 29.Hsiung BK, Siddique RH, Stavenga DG, Otto JC, Allen MC, Liu Y, et al. Rainbow peacock spiders inspire miniature super-iridescent optics. Nat Commun. 2017. Dec 22;8(1):2278. doi: 10.1038/s41467-017-02451-x [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 30.Kjernsmo K, Whitney HM, Scott-Samuel NE, Hall JR, Knowles H, Talas L, et al. Iridescence as camouflage. Curr Biol. 2020. Feb 3;30(3):551–555.e3. doi: 10.1016/j.cub.2019.12.013 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 31.Pinto F, Mielewczik M, Liebisch F, Walter A, Greven H, Rascher U. Non-invasive measurement of frog skin reflectivity in high spatial resolution using a dual hyperspectral approach. PLoS ONE. 2013. Sep 18;8(9):e73234. doi: 10.1371/journal.pone.0073234 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 32.Chiao CC, Wickiser JK, Allen JJ, Genter B, Hanlon RT. Hyperspectral imaging of cuttlefish camouflage indicates good color match in the eyes of fish predators. Proc Natl Acad Sci U S A. 2011. May 31;108(22):9148–53. doi: 10.1073/pnas.1019090108 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 33.Russell BJ, Dierssen HM. Use of hyperspectral imagery to assess cryptic color matching in Sargassum associated crabs. PLoS ONE. 2015. Sep 9;10(9):e0136260. doi: 10.1371/journal.pone.0136260 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 34.Penacchio O, Halpin CG, Cuthill IC, Lovell PG, Wheelwright M, Skelhorn J, et al. A computational neuroscience framework for quantifying warning signals. Methods Ecol Evol. 2024;15(1):103–116. [Google Scholar]
  • 35.Moniz HA, Kruleski SM, Risbud AD, Louden SJH, Hanlon RT, Stokes AN, et al. Aposematic coloration of Pacific newts (Taricha) provides a qualitatively but not quantitatively honest signal to predators. Biol J Linn Soc. 2023. May 1;139(1):1–17. [Google Scholar]
  • 36.Gruber DF, Loew ER, Deheyn DD, Akkaynak D, Gaffney JP, Smith WL, et al. Biofluorescence in catsharks (Scyliorhinidae): fundamental description and relevance for elasmobranch visual ecology. Sci Rep. 2016. Apr 25;6(1):24751. doi: 10.1038/srep24751 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 37.Kim MH, Harvey TA, Kittle DS, Rushmeier H, Dorsey J, Prum RO, et al. 3D imaging spectroscopy for measuring hyperspectral patterns on solid objects. ACM Trans Graph. 2012. Jul 1;31(4):38:1–38:11. [Google Scholar]
  • 38.Nevala NE, Baden T. A low-cost hyperspectral scanner for natural imaging and the study of animal colour vision above and under water. Sci Rep. 2019. Jul 25;9(1):10799. doi: 10.1038/s41598-019-47220-6 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 39.Jacobs GH. Photopigments and the dimensionality of animal color vision. Neurosci Biobehav Rev. 2018. Mar 1;86:108–30. doi: 10.1016/j.neubiorev.2017.12.006 [DOI] [PubMed] [Google Scholar]
  • 40.Beehler BM, Frith CB. The birds of paradise: Paradisaeidae. New York: Oxford University Press; 1998. (Bird families of the world; 6.). [Google Scholar]
  • 41.Ligon RA, Diaz CD, Morano JL, Troscianko J, Stevens M, Moskeland A, et al. Evolution of correlated complexity in the radically different courtship signals of birds-of-paradise. PLoS Biol. 2018. Nov 20;16(11):e2006962. doi: 10.1371/journal.pbio.2006962 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 42.Beehler BM, Pratt TK. Birds of New Guinea: Distribution, Taxonomy, and Systematics. Princeton University Press; 2016. [Google Scholar]
  • 43.Koch A. Discovery of a rare hybrid specimen known as Maria’s bird of paradise at the Staatliches Naturhistorisches Museum in Braunschweig. Zoosystematics Evol. 2018. May 31;94(2):315. [Google Scholar]
  • 44.Martin PR. The paradox of the Birds-of-Paradise: persistent hybridization as a signature of historical reinforcement. Ideas Ecol Evol [Internet]. 2015. Aug 4 [cited 2021 Apr 5];8(1). Available from: https://ojs.library.queensu.ca/index.php/IEE/article/view/5485. [Google Scholar]
  • 45.Berlioz J. Remarques sur l’hybridation naturelle chez les oiseaux. Bull Soc Zool Fr. 1927;LII(1):393–403. [Google Scholar]
  • 46.Iredale T. Birds of paradise and bower birds. Melbourne: Georgian House; 1950. xii, p. 239. [Google Scholar]
  • 47.Meyer AB. A new bird of paradise from the island of Waigeou, near New Guinea. Nature. 1875. Jan;11(272):208–208. [Google Scholar]
  • 48.Stresemann E. Welche Paradiesvogelarten der literatur sing hybriden ursprungs. Novit Zool. 1930;XXXVI:6–17. [Google Scholar]
  • 49.Thörn F, Soares AER, Müller IA, Päckert M, Frahnert S, van Grouw H, et al. Contemporary intergeneric hybridization and backcrossing among birds-of-paradise. Evol Lett. 2024. Oct 1;8(5):680–94. doi: 10.1093/evlett/qrae023 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 50.Levaillant F, Barraband J. Histoire naturelle des oiseaux de paradis et des rolliers [Internet]. Paris: Denné le jeune; 1806. Available from: https://www.biodiversitylibrary.org/bibliography/61652 [Google Scholar]
  • 51.Gould J, Sharpe RB, Hart WM, Walter T, Mintern Bros. The birds of New Guinea and the adjacent Papuan islands: including many new species recently discovered in Australia [Internet]. London: Henry Sotheran & Co.; 1875. p. 242. Available from: https://www.biodiversitylibrary.org/bibliography/133931. [Google Scholar]
  • 52.Irestedt M, Jønsson KA, Fjeldså J, Christidis L, Ericson PG. An unexpectedly long history of sexual selection in birds-of-paradise. BMC Evol Biol. 2009. Sep 16;9(1):235. doi: 10.1186/1471-2148-9-235 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 53.Gadow HF. On the colour of feathers as affected by their structure. Proc Zool Soc Lond. 1882;50(3):409–422. [Google Scholar]
  • 54.Graves GR. Sequence of plumage evolution in the standardwing bird of paradise. Wilson Bull. 1995. Jun;107(2):371–3. [Google Scholar]
  • 55.McCoy DE, Feo T, Harvey TA, Prum RO. Structural absorption by barbule microstructures of super black bird of paradise feathers. Nat Commun. 2018. Jan 9;9(1):1. doi: 10.1038/s41467-017-02088-w [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 56.Stavenga DG, Leertouwer HL, Marshall NJ, Osorio D. Dramatic colour changes in a bird of paradise caused by uniquely structured breast feather barbules. Proc R Soc B Biol Sci. 2010. Dec 15;278(1715):2098–104. doi: 10.1098/rspb.2010.2293 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 57.Wilts BD, Michielsen K, Raedt HD, Stavenga DG. Sparkling feather reflections of a bird-of-paradise explained by finite-difference time-domain modeling. Proc Natl Acad Sci U S A. 2014. Mar 25;111(12):4363–8. doi: 10.1073/pnas.1323611111 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 58.Brelsford A, Toews DPL, Irwin DE. Admixture mapping in a hybrid zone reveals loci associated with avian feather coloration. Proc R Soc B Biol Sci. 2017. Nov 15;284(1866):20171106. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 59.Vallender R, Gagnon JP, Lovette I. An intergeneric wood-warbler hybrid (Mniotilta varia × Dendroica coronata) and use of multilocus DNA analyses to diagnose avian hybrid origins. Wilson J Ornithol. 2009. Jun 1;121(2):298–305. [Google Scholar]
  • 60.Baldassarre DT, White TA, Karubian J, Webster MS. Genomic and morphological analysis of a semipermeable avian hybrid zone suggests asymmetrical introgression of a sexual signal. Evolution. 2014. Jun 2;68(9):2644–57. doi: 10.1111/evo.12457 [DOI] [PubMed] [Google Scholar]
  • 61.Brumfield RT, Jernigan RW, McDonald DB, Braun MJ. Evolutionary implications of divergent clines in an avian (manacus: Aves) hybrid zone. Evolution. 2001. Oct;55(10):2070–87. doi: 10.1111/j.0014-3820.2001.tb01322.x [DOI] [PubMed] [Google Scholar]
  • 62.Céspedes-Arias LN, Cuervo AM, Bonaccorso E, Castro-Farias M, Mendoza-Santacruz A, Pérez-Emán JL, et al. Extensive hybridization between two Andean warbler species with shallow divergence in mtDNA. Ornithology. 2021. Jan 4;138(1):ukaa065. [Google Scholar]
  • 63.Graves GR. Hybrid wood warblers, Dendroica striata x Dendroica castanea (Aves: Fringillidae: tribe Parulini) and the diagnostic predictability of avian hybrid phenotypes. Proc Biol Soc Wash. 1996. Jun 25;(109 (2)):373–90. [Google Scholar]
  • 64.Barrera-Guzmán AO, Aleixo A, Shawkey MD, Weir JT. Hybrid speciation leads to novel male secondary sexual ornamentation of an Amazonian bird. Proc Natl Acad Sci U S A. 2018. Jan 9;115(2):E218–25. doi: 10.1073/pnas.1717319115 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 65.Eliason CM, Cooper JC, Hackett SJ, Zahnle E, Pequeño Saco TZ, Maddox JD, et al. Interspecific hybridization explains rapid gorget colour divergence in Heliodoxa hummingbirds (Aves: Trochilidae). R Soc Open Sci. 2023. Mar 1;10(3):221603. doi: 10.1098/rsos.221603 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 66.Goldsmith TH. Optimization, constraint, and history in the evolution of eyes. Q Rev Biol. 1990. Sep;65(3):281–322. doi: 10.1086/416840 [DOI] [PubMed] [Google Scholar]
  • 67.Endler JA, Mielke PW. Comparing entire colour patterns as birds see them. Biol J Linn Soc. 2005. Dec 1;86(4):405–31. [Google Scholar]
  • 68.Stoddard MC, Prum RO. Evolution of avian plumage color in a tetrahedral color space: a phylogenetic analysis of new world buntings. Am Nat. 2008. Jun;171(6):755–76. doi: 10.1086/587526 [DOI] [PubMed] [Google Scholar]
  • 69.Schneider CA, Rasband WS, Eliceiri KW. NIH Image to ImageJ: 25 years of image analysis. Nat Methods. 2012. Jun 28;9(7):671–5. doi: 10.1038/nmeth.2089 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 70.Maia R, Eliason CM, Bitton P, Doucet SM, Shawkey MD, Tatem A. pavo: an R package for the analysis, visualization and organization of spectral data. Methods Ecol Evol. 2013. May 22;4(10):906–13. [Google Scholar]
  • 71.Currie RP. A new bird of paradise. Proc U S Natl Mus. 1900;22(1204):497–499. [Google Scholar]
  • 72.Pedregosa F, Varoquaux G, Gramfort A, Michel V, Thirion B, Grisel O, et al. Scikit-learn: Machine learning in Python. J Mach Learn Res. 2011;12:2825–2830. [Google Scholar]
  • 73.Vorobyev M, Osorio D. Receptor noise as a determinant of colour thresholds. Proc R Soc B. 1998. Mar 7;265(1394):351–8. doi: 10.1098/rspb.1998.0302 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 74.Osorio D, Ham AD. Spectral reflectance and directional properties of structural coloration in bird plumage. J Exp Biol. 2002. Jul 15;205(14):2017–27. doi: 10.1242/jeb.205.14.2017 [DOI] [PubMed] [Google Scholar]
  • 75.Santos SICO, De Neve L, Lumeij JT, Förschler MI. Strong effects of various incidence and observation angles on spectrometric assessment of plumage colouration in birds. Behav Ecol Sociobiol. 2007. Jul 1;61(9):1499–506. [Google Scholar]
  • 76.Olsson P, Lind O, Kelber A, Simmons L. Chromatic and achromatic vision: parameter choice and limitations for reliable model predictions. Behav Ecol. 2018. Mar 14;29(2):273–82. [Google Scholar]
  • 77.Siddiqi A, Cronin TW, Loew ER, Vorobyev M, Summers K. Interspecific and intraspecific views of color signals in the strawberry poison frog Dendrobates pumilio. J Exp Biol. 2004. Jun 15;207(14):2471–85. [DOI] [PubMed] [Google Scholar]
  • 78.Maia R, White TE. Comparing colors using visual models. Behav Ecol. 2018. May 9;29(3):649–59. [Google Scholar]
  • 79.Ospina-Rozo L, Roberts A, Stuart-Fox D. A generalized approach to characterize optical properties of natural objects. Biol J Linn Soc. 2022. Nov 1;137(3):534–55. [Google Scholar]
  • 80.Meadows MG, Morehouse NI, Rutowski RL, Douglas JM, McGraw KJ. Quantifying iridescent coloration in animals: a method for improving repeatability. Behav Ecol Sociobiol. 2011. Jun 1;65(6):1317–27. [Google Scholar]
  • 81.Hogan BG, Stoddard MC. Synchronization of speed, sound and iridescent color in a hummingbird aerial courtship dive. Nat Commun. 2018. Dec 18;9(1):5260. doi: 10.1038/s41467-018-07562-7 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 82.He Y, Varley ZK, Nouri LO, Moody CJA, Jardine MD, Maddock S, et al. Deep learning image segmentation reveals patterns of UV reflectance evolution in passerine birds. Nat Commun. 2022. Aug 29;13(1):5068. doi: 10.1038/s41467-022-32586-5 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 83.Kirillov A, Mintun E, Ravi N, Mao H, Rolland C, Gustafson L, et al. Segment anything [Internet]. arXiv; 2023. [cited 2023 Sep 27]. Available from: http://arxiv.org/abs/2304.02643. [Google Scholar]
  • 84.Valvo JJ, Aponte JD, Daniel MJ, Dwinell K, Rodd H, Houle D, et al. Using Delaunay triangulation to sample whole-specimen color from digital images. Ecol Evol. 2021. Aug 20;11(18):12468–84. doi: 10.1002/ece3.7992 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 85.Van Belleghem SM, Papa R, Ortiz-Zuazaga H, Hendrickx F, Jiggins CD, Owen McMillan W, et al. patternize: An R package for quantifying colour pattern variation. Methods Ecol Evol. 2017. Jul 11;9(2):390–8. doi: 10.1111/2041-210X.12853 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 86.Irschick DJ, Christiansen F, Hammerschlag N, Martin J, Madsen PT, Wyneken J, et al. 3D visualization processes for recreating and studying organismal form. iScience. 2022. Aug 4;25(9):104867. doi: 10.1016/j.isci.2022.104867 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 87.Funk ER, Taylor SA. High-throughput sequencing is revealing genetic associations with avian plumage color. The Auk. 2019. Oct 21;136(4):ukz048. [Google Scholar]
  • 88.Lopes RJ, Johnson JD, Toomey MB, Ferreira MS, Araujo PM, Melo-Ferreira J, et al. Genetic basis for red coloration in birds. Curr Biol. 2016. Jun 6;26(11):1427–34. doi: 10.1016/j.cub.2016.03.076 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 89.Hudon J, Wiebe KL, Pini E, Stradi R. Plumage pigment differences underlying the yellow-red differentiation in the northern flicker (Colaptes auratus). Comp Biochem Physiol B Biochem Mol Biol. 2015. May 1;183:1–10. [DOI] [PubMed] [Google Scholar]
  • 90.Hudon J, Wiebe KL, Stradi R. Disruptions of feather carotenoid pigmentation in a subset of hybrid northern flickers (Colaptes auratus) may be linked to genetic incompatibilities. Comp Biochem Physiol B Biochem Mol Biol. 2021. Jan;251:110510. [DOI] [PubMed] [Google Scholar]
  • 91.Aguillon SM, Walsh J, Lovette IJ. Extensive hybridization reveals multiple coloration genes underlying a complex plumage phenotype. Proc R Soc B Biol Sci. 2021. Jan 27;288(1943):20201805. doi: 10.1098/rspb.2020.1805 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 92.Wiebe KL, Bortolotti GR. Variation in colour within a population of northern flickers: a new perspective on an old hybrid zone. Can J Zool. 2001. Jun 1;79(6):1046–52. [Google Scholar]
  • 93.Khalil S, Welklin JF, McGraw KJ, Boersma J, Schwabl H, Webster MS, et al. Testosterone regulates CYP2J19-linked carotenoid signal expression in male red-backed fairywrens (Malurus melanocephalus). Proc R Soc B Biol Sci. 2020. Sep 30;287(1935):20201687. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 94.Mundy NI, Stapley J, Bennison C, Tucker R, Twyman H, Kim KW, et al. Red carotenoid coloration in the zebra finch is controlled by a cytochrome p450 gene cluster. Curr Biol. 2016. Jun 6;26(11):1435–40. doi: 10.1016/j.cub.2016.04.047 [DOI] [PubMed] [Google Scholar]
  • 95.Price-Waldman R, Stoddard MC. Avian coloration genetics: Recent advances and emerging questions. J Hered. 2021. Jul 1;112(5):395–416. doi: 10.1093/jhered/esab015 [DOI] [PubMed] [Google Scholar]
  • 96.Dorst J, Gastaldi G, Hagege R, Jacquemart J. Différents aspects des barbules de quelques Paradisaeidés sur coupes en microscopie électronique. Comptes Rendus L’Academie Sci Paris. 1974;278:285–290. [Google Scholar]
  • 97.Nordén KK, Eliason CM, Stoddard MC. Evolution of brilliant iridescent feather nanostructures. eLife. 2021. Dec 21;10:e71179. doi: 10.7554/eLife.71179 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 98.Travouillon KJ, Cooper C, Bouzin JT, Umbrello LS, Lewis SW. All-a-glow: spectral characteristics confirm widespread fluorescence for mammals. R Soc Open Sci. 2023. Oct 4;10(10):230325. doi: 10.1098/rsos.230325 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 99.Hogan BG, Stoddard MC. Hyperspectral images of King, Magnificent, and hybrid King of Holland’s Bird-of-Paradise (Cicinnurus regius, C. magnificus, and C. magnificus x C. regius). doi: 10.5061/dryad.j0zpc86nf [DOI] [Google Scholar]

Decision Letter 0

Richard Hodge

21 Dec 2023

Dear Dr Hogan,

Thank you for submitting your manuscript entitled "Hyperspectral imaging is a powerful tool for animal coloration research: A case study of a rare hybrid bird-of-paradise" for consideration as a Methods and Resources Article by PLOS Biology. Please accept my apologies for the delay in getting back to you as we consulted with an academic editor about your submission.

Your manuscript has now been evaluated by the PLOS Biology editorial staff, as well as by an academic editor with relevant expertise, and I am writing to let you know that we would like to send your submission out for external peer review.

However, before we can send your manuscript to reviewers, we need you to complete your submission by providing the metadata that is required for full assessment. To this end, please login to Editorial Manager where you will find the paper in the 'Submissions Needing Revisions' folder on your homepage. Please click 'Revise Submission' from the Action Links and complete all additional questions in the submission questionnaire.

Once your full submission is complete, your paper will undergo a series of checks in preparation for peer review. After your manuscript has passed the checks it will be sent out for review. To provide the metadata for your submission, please Login to Editorial Manager (https://www.editorialmanager.com/pbiology) within two working days, i.e. by Dec 23 2023 11:59PM.

If your manuscript has been previously peer-reviewed at another journal, PLOS Biology is willing to work with those reviews in order to avoid re-starting the process. Submission of the previous reviews is entirely optional and our ability to use them effectively will depend on the willingness of the previous journal to confirm the content of the reports and share the reviewer identities. Please note that we reserve the right to invite additional reviewers if we consider that additional/independent reviewers are needed, although we aim to avoid this as far as possible. In our experience, working with previous reviews does save time.

If you would like us to consider previous reviewer reports, please edit your cover letter to let us know and include the name of the journal where the work was previously considered and the manuscript ID it was given. In addition, please upload a response to the reviews as a 'Prior Peer Review' file type, which should include the reports in full and a point-by-point reply detailing how you have or plan to address the reviewers' concerns.

During the process of completing your manuscript submission, you will be invited to opt-in to posting your pre-review manuscript as a bioRxiv preprint. Visit http://journals.plos.org/plosbiology/s/preprints for full details. If you consent to posting your current manuscript as a preprint, please upload a single Preprint PDF.

Feel free to email us at plosbiology@plos.org if you have any queries relating to your submission.

Kind regards,

Richard

Richard Hodge, PhD

Senior Editor, PLOS Biology

rhodge@plos.org

PLOS

Empowering researchers to transform science

Carlyle House, Carlyle Road, Cambridge, CB4 3DN, United Kingdom

ORCiD I plosbio.org I @PLOSBiology I Blog

California (U.S.) corporation #C2354500, based in San Francisco

Decision Letter 1

Richard Hodge

21 Feb 2024

Dear Dr Hogan,

Thank you for your continued patience while your manuscript "Hyperspectral imaging is a powerful tool for animal coloration research: A case study of a rare hybrid bird-of-paradise" was peer-reviewed at PLOS Biology as a Methods and Resources Article. Please accept my sincere apologies for the delays that you have experienced during the peer review process. Your manuscript has now been evaluated by the PLOS Biology editors, an Academic Editor with relevant expertise, and by three independent reviewers. Please note that Reviewer #2 has provided the review as a PDF document which is attached to this decision letter.

In light of the reviews, which you will find at the end of this email, we would like to invite you to revise the work to thoroughly address the reviewers' reports.

As you will see, the reviewers are generally very positive about your hyperspectral imaging pipeline and think the method will be useful for the field. However, Reviewers #2 and #3 raise some overlapping concerns, including the overall framing of the paper and noting that a greater emphasis should be placed on the methodological advance. In addition, Reviewer #3 notes that the manuscript lacks direct benchmarking to other existing methods in order to demonstrate a proven advantage. After discussions with the Academic Editor, we agree that this would be a valuable addition and we ask that direct benchmarking experiments are included in the revision. Finally, Reviewer #2 asks that additional reporting details should be included for the lighting and elaborating on other applications offered by the approach.

Given the extent of revision needed, we cannot make a decision about publication until we have seen the revised manuscript and your response to the reviewers' comments. Your revised manuscript is likely to be sent for further evaluation by all or a subset of the reviewers.

We expect to receive your revised manuscript within 3 months. Please email us (plosbiology@plos.org) if you have any questions or concerns, or would like to request an extension.

At this stage, your manuscript remains formally under active consideration at our journal; please notify us by email if you do not intend to submit a revision so that we may withdraw it.

**IMPORTANT - SUBMITTING YOUR REVISION**

Your revisions should address the specific points made by each reviewer. Please submit the following files along with your revised manuscript:

1. A 'Response to Reviewers' file - this should detail your responses to the editorial requests, present a point-by-point response to all of the reviewers' comments, and indicate the changes made to the manuscript.

*NOTE: In your point-by-point response to the reviewers, please provide the full context of each review. Do not selectively quote paragraphs or sentences to reply to. The entire set of reviewer comments should be present in full and each specific point should be responded to individually, point by point.

You should also cite any additional relevant literature that has been published since the original submission and mention any additional citations in your response.

2. In addition to a clean copy of the manuscript, please also upload a 'track-changes' version of your manuscript that specifies the edits made. This should be uploaded as a "Revised Article with Changes Highlighted" file type.

*Re-submission Checklist*

When you are ready to resubmit your revised manuscript, please refer to this re-submission checklist: https://plos.io/Biology_Checklist

To submit a revised version of your manuscript, please go to https://www.editorialmanager.com/pbiology/ and log in as an Author. Click the link labelled 'Submissions Needing Revision' where you will find your submission record.

Please make sure to read the following important policies and guidelines while preparing your revision:

*Published Peer Review*

Please note while forming your response, if your article is accepted, you may have the opportunity to make the peer review history publicly available. The record will include editor decision letters (with reviews) and your responses to reviewer comments. If eligible, we will contact you to opt in or out. Please see here for more details:

https://blogs.plos.org/plos/2019/05/plos-journals-now-open-for-published-peer-review/

*PLOS Data Policy*

Please note that as a condition of publication PLOS' data policy (http://journals.plos.org/plosbiology/s/data-availability) requires that you make available all data used to draw the conclusions arrived at in your manuscript. If you have not already done so, you must include any data used in your manuscript either in appropriate repositories, within the body of the manuscript, or as supporting information (N.B. this includes any numerical values that were used to generate graphs, histograms etc.). For an example see here: http://www.plosbiology.org/article/info%3Adoi%2F10.1371%2Fjournal.pbio.1001908#s5

*Blot and Gel Data Policy*

We require the original, uncropped and minimally adjusted images supporting all blot and gel results reported in an article's figures or Supporting Information files. We will require these files before a manuscript can be accepted so please prepare them now, if you have not already uploaded them. Please carefully read our guidelines for how to prepare and upload this data: https://journals.plos.org/plosbiology/s/figures#loc-blot-and-gel-reporting-requirements

*Protocols deposition*

To enhance the reproducibility of your results, we recommend that if applicable you deposit your laboratory protocols in protocols.io, where a protocol can be assigned its own identifier (DOI) such that it can be cited independently in the future. Additionally, PLOS ONE offers an option for publishing peer-reviewed Lab Protocol articles, which describe protocols hosted on protocols.io. Read more information on sharing protocols at https://plos.org/protocols?utm_medium=editorial-email&utm_source=authorletters&utm_campaign=protocols

Thank you again for your submission to our journal. We hope that our editorial process has been constructive thus far, and we welcome your feedback at any time. Please don't hesitate to contact us if you have any questions or comments.

Sincerely,

Richard

Richard Hodge, PhD

Senior Editor, PLOS Biology

rhodge@plos.org

------------------------------------

REVIEWS:

Reviewer #1: GENERAL

A splendid paper outlining a potentially transformative new approach to avian color quantification. I want to highlight, in particular, the beautiful and informative figures. The authors strike a nice balance among justifying this new approach relative to two others, describing the pipeline and potential applications, and illustrating its use with an interesting example. I have mainly minor comments throughout, see below for details.

No methods section? Can methods and results be separated? Not sure about journal format, but L233-455 read like methods and L456-678 read like results.

Figure 2: Put hybrid image between parentals, logical and aligns better with subsequent figures.

RESULTS

L242: Double-check your figure panel letters here.

L261: Commentary about missing those shortest 25 nm here or elsewhere? These species don't have much UV reflectance, but how could this affect species with more prominent UV patches? Might be interesting to do an analysis validating the extrapolation technique.

L457: This section is somewhat redundant with some previous sections. I think you've already clearly made the case for hyperspectral imaging and how/why it's different/better than the other two methods (e.g., Table 1).

Figure 5: This is the only figure with embedded descriptive text, could be removed and covered in the caption.

L495: I would rethink the use of the word "favors" here. "Resembles" might do.

L552: I think "further" should be "farther" here, as this refers to an actual and not metaphorical distance. Look for other instances of this throughout the ms if there are any.

DISCUSSION

L742: I think this section could be expanded upon to discuss the potential limitations of basing color measurements/comparisons off specimens alone (e.g., effects of age, preservation method/quality, metadata availability).

L753: Watch for capitalization consistency in species names.

Reviewer #2 (Daniel Hanley, signs review): Please see the PDF document attached to this e-mail for the full review.

Reviewer #3 (Jolyon Troscianko, signs review): This paper showcases a pipeline for hyperspectral imaging of museum specimens, and uses it to investigate the provenance of a putative hybrid bird of paradise.

Overall I really like the methods, and wholeheartedly welcome the release of open-source tools for the community. The manuscript is nicely written, though I feel it could be shortened in parts. Hyperspectral imaging techniques will be of increasing interest to a wide audience, and birds-of-paradise are a charismatic group to investigate.

I make a large number of minor points below, but none should be difficult to deal with, it mostly reflects the fact that (I think) I understand the methods quite well.

I have just two main concerns (again, easy to address, but might entail some work). One trying to work out the rationale for this paper, the other technical.

Rationale: The core hypothesis being tested is not entirely convincing as-is. e.g. deciding whether a species of apparently intermediate coloration/physiology between two other species really represents a hybrid would ideally need a full phylogenetic comparison (measure lots of species with this technique, including other known hybrids), or at least some meaningful out-group, control or reference. As such, there's no clear alternate hypothesis. For example, the statement that the (alleged) hybrid is 'strikingly intermediate' has no baseline for comparison. I don't know how much the authors want to sell this finding as important though, as I feel the real value of the paper lies in its methodology (which is great), and this hypothesis is just used as a demo. In any case it's shortcomings are worth discussing. However, if the authors do want to sell the methodology there are some important things I'd like to see to convince readers that this method is worth using themselves. e.g. How does it compare to other methods in practice (show evidence)? Would alternative conclusions be reached if this study if other (cheaper/easier) methods were used (i.e. direct comparison of the output of methods, or modelling of circumstances where high spectral resolution is required?)

Processing methods: (L342) I'd like more information on the rationale for the sampling and smoothing applied here. And importantly I'd like to see some raw data before any smoothing is applied. 45 nm is a surprisingly spectrally wide window, and 1nm a surprisingly high spectral resolution. This stage will be computationally inefficient as it's artificially increasing the spectral resolution from that of the native hardware. This will also cause all following calculations to be equally inefficient for no apparent gain in accuracy. Up-sampling might give false impressions of spectral resolution. Why not just downsample the entire cube to e.g. a 10nm resolution that wouldn't need smoothing, and would give pretty similar effects for less computational overheads? Or even more elegant, just work with the spectral range you're given by the camera and adjust the bin-width in the calculations as necessary (as these will presumably be non-linear).

Related question: L336 It strikes me that the processing pipeline would be far more efficient if each pixel were first converted to avian cone-catch values (or non-system-specific bins), and then the rest of the analysis done simply on conventional images. This doesn't strike me as particularly computationally tricky or slow. Even a single processor should take less than a minute to churn through a 5MPx200 channel cube with efficient code (python might not be great for this heavy lifting though). I haven't seen the code, but would like to have a look,

Minor points:

It's worth pointing out that high spectral resolution is not always required. Natural reflectance spectra and solar illumination rarely have sharp changes across a 5-10nm range. This means we can infer the way colours will look to a given animal under given (natural) lighting with a fairly high degree of confidence. The reflectance profiles shown in fig.1 for example, would behave very predictably at lower spectral resolution sampling. Where this fails is when either the light source is non broadband (important for my current work in artificial light pollution), or when reflectance profiles are particularly rare (e.g. sometimes achieved in structural coloration).

L20 I'd query whether hyperspectral data are computationally costly. It depends on the spectral, spatial, and temporal resolution, but the hardware has typically been the limiting factor here. I also query the affordability below.

Table 1. I'd like a further row of "temporal resolution" added. This highlights the issue with using multispectral and hyperspectral cameras and moving targets. e.g. some systems can work instantaneously, or capture video, but others require slower scanning/repeat photos taking. The camera used here requires scanning, so cannot take instantaneous images. It's perhaps worth referencing Vasas et al's recent dichroic camera PLoS Biol paper here as an example of low-spectral, but high spatial and temporal resolution (you cite the pre-print, but can cite the full thing now).

L77 multispectral imaging doesn't always use one camera spectral channel for one receptor channel (though often does). Worth just rephrasing slightly. e.g. people often model achromatic and dichromatic vision using RGB cameras.

L91 Multispectral and hyperspectral data are both able to generate reflectance spectra. It's simply a matter of resolution and range. Indeed, conventional spectrometry can't handle the full UV-vis-IR-thermalIR ranges required to understand all the effects discussed here.

L104 'brightness' is not the right word here. The system will be measuring radiance (specifically energy of light in a known solid angle or a known spectral range).

L141 Could you say how much the system used here cost? This would be valuable for readers. Include the broadband stable lighting setup used (these can be super expensive too), plus the ~$10,000 UV-VIS lens. It'll all add up quite a bit. Presumably $20-30k? Which dents the angle that this is affordable.

L254 Personally I'd like to see other colour spaces added. e.g. the RNL Chromaticity colour space is much easier to use, interpret, and analyse with reference to discriminablility (though I'd caution that almost all colour spaces are pretty flawed, so it's not worth getting too fussy).

L247 Are you getting at measuring gloss and iridescence here? i.e. goniometry? If so, maybe be explicit. You bring this up later at least, but generally there's scope of more concise writing.

L264 This spatial correction is interesting, presumably normalising each pixel to account for non-diffuse lighting. But this highlights real problems for scanning/imaging anything that's not flat. Things nearer the lights than the 2D white calibration tile will appear artificially lighter. So for 3D subjects illuminated from non-diffuse, close light sources care needs to be taken for accurate reflectance estimates.

L262 birds have very low sensitivity to UV below 325nm, and the sun doesn't give out much energy at these wavelengths either. I'm just saying this in case anyone worries that part of the range is 'missing' - it's not likely to be an issue as far as I can tell.

L288 Really happy to see use of Python (great to keep code open source). But is it all in python? As an interpreted language it's not always very quick for image processing.

L355 I disagree here - cameras are frequently used to compare spectral reflectance. e.g. NDVI methods, and enormous numbers of methods and studies. I get what the authors are hinting at, but the only real difference with the system used here is its spectral resolution, not the ability of the system to measure spectral reflectance across a given spectral range. Indeed, even common light meters (not cameras) often work by comparing a 'full-spectrum' to an infrared measurement to determine brightness in the visible range.

L376 Why not just focus on the PCA-based method then? I like the PCA method, and it's used a fair bit in the field (as you say, it's easy to understand and apply stats etc...). Could make the paper more concise, and help the readers.

L408 But you are only measuring each point once or twice without actively controlling the angles, so unlike goniometry you can't quite work out iridescence with this method (without more work).

L535 How sure can we be that these differences aren't due to noise of some sort? Also, all spectral recordings will differ, but how can we say different any two are without some point of reference or threshold?

L590 If you're converting to RNL values why use the tetrahedral colour space at all? Just convert to RNL XYZ chromaticity, then you have your hue axes just like the tetrahedral space, but now Euclidean distance is equal to RNL delta-S values (i.e. perceptually uniform-ish). Hue and discriminability can be compared in one go.

L595 What visual system (spectral sensitivity curves, Weber fraction and cone ratios do you use here?)

L628 I like this method for modelling feather growth. Though maybe you could have just fitted a bezier spline? It might be worth mentioning why developmentally-inspired methods are used rather than just absolute modelling here, but this approach is not used for spectral reflectance curves (even though the discussion goes over many developmentally relevant colour creation mechanisms)?Playing devil's advocate, this is a lot of work and effort to say something that is obvious to any viewer of the feathers (the intermediate one is intermediate). I really like the methods, but their use doesn't seem enormously well justified here - persuade me I'm wrong.

L703 As above, I'm not convinced there are meaningful differences shown between spectra of the same patch. What exactly would 'meaningful' be given everything in the world has some degree of difference?

L798 Fluorescence is a pig! Relating to my above point, artificial light sources aren't able to simulate how something looks if strong fluorescence is involved as it violates the illuminant/reflectance equations. Few people realise this, so it's perhaps worth pointing out to readers. Your use of xenon lighting was fairly good in this regard, and moderately likely to replicate natural viewing (although the UV component will be far higher than sunlight). Most other light sources will create more errors in inferred reflectance.

The discussion gets very speculative regarding bird of paradise crosses, and to be honest I don't find it particularly useful if the thrust of this paper is methods. The bird of paradise based work here doesn't shed light on many of the issues touched on in the discussion.

Jolyon Troscianko

Attachment

Submitted filename: Hogan and Stoddard Review.pdf

pbio.3002867.s003.pdf (334.4KB, pdf)

Decision Letter 2

Richard Hodge

13 Aug 2024

Dear Dr Hogan,

Thank you for your patience while we considered your revised manuscript "Hyperspectral imaging is a powerful tool for animal coloration research: A case study of a rare hybrid bird-of-paradise" for publication as a Methods and Resources Article at PLOS Biology. This revised version of your manuscript has been evaluated by the PLOS Biology editors, the Academic Editor and the original reviewers. Please note that Reviewer #2 has provided their review as a PDF file which is attached to this letter.

Based on the reviews, I am pleased to say that we are likely to accept this manuscript for publication, provided you satisfactorily address the remaining minor points raised by the reviewers. Please also make sure to address the following data and other policy-related requests that I have provided below (A-F):

(A) We routinely suggest changes to titles to ensure maximum accessibility for a broad, non-specialist readership, and to ensure they reflect the contents of the paper. In this case, we would suggest a minor edit to the title, as follows. Please ensure you change both the manuscript file and the online submission system, as they need to match for final acceptance:

“A user-friendly pipeline for the generation and analysis of hyperspectral images for animal coloration research”

(B) You may be aware of the PLOS Data Policy, which requires that all data be made available without restriction: http://journals.plos.org/plosbiology/s/data-availability. For more information, please also see this editorial: http://dx.doi.org/10.1371/journal.pbio.1001797

Note that we do not require all raw data. Rather, we ask that all individual quantitative observations that underlie the data summarized in the figures and results of your paper be made available in one of the following forms:

-Supplementary files (e.g., excel). Please ensure that all data files are uploaded as 'Supporting Information' and are invariably referred to (in the manuscript, figure legends, and the Description field when uploading your files) using the following format verbatim: S1 Data, S2 Data, etc. Multiple panels of a single or even several figures can be included as multiple sheets in one excel file that is saved using exactly the following convention: S1_Data.xlsx (using an underscore).

-Deposition in a publicly available repository. Please also provide the accession code or a reviewer link so that we may view your data before publication.

Regardless of the method selected, please ensure that you provide the individual numerical values that underlie the summary data displayed in the following figure panels as they are essential for readers to assess your analysis and to reproduce it:

Figure 1C, 4F, 5A-N, 6A-J, 7B-D, S3, S4, S5A-N, S6A-N, S7, S8, S9, S16, S17, S21

NOTE: the numerical data provided should include all replicates AND the way in which the plotted mean and errors were derived (it should not present only the mean/average values).

(C) I tried to search for the hyperspectral image data deposited in the Dryad database using the DOI provided in the Data Availability Statement, but could not find it. I would be grateful if you could please make sure that the DOI is correct or to make the data publicly available at this stage.

(D) Please also ensure that each of the relevant figure legends in your manuscript include information on *WHERE THE UNDERLYING DATA CAN BE FOUND*, and ensure your supplemental data file/s has a legend.

(E) Please ensure that your Data Statement in the submission system accurately describes where your data can be found and is in final format, as it will be published as written there.

(F) Please note that we cannot accept sole deposition of code in GitHub, as this could be changed after publication. However, you can archive this version of your publicly available GitHub code to Zenodo. Once you do this, it will generate a DOI number, which you will need to provide in the Data Accessibility Statement (you are welcome to also provide the GitHub access information). See the process for doing this here: https://docs.github.com/en/repositories/archiving-a-github-repository/referencing-and-citing-content

------------------------------------------------------------------------

As you address these items, please take this last chance to review your reference list to ensure that it is complete and correct. If you have cited papers that have been retracted, please include the rationale for doing so in the manuscript text, or remove these references and replace them with relevant current references. Any changes to the reference list should be mentioned in the cover letter that accompanies your revised manuscript.

We expect to receive your revised manuscript within three weeks.

To submit your revision, please go to https://www.editorialmanager.com/pbiology/ and log in as an Author. Click the link labelled 'Submissions Needing Revision' to find your submission record. Your revised submission must include the following:

- a cover letter that should detail your responses to any editorial requests, if applicable, and whether changes have been made to the reference list

- a Response to Reviewers file that provides a detailed response to the reviewers' comments (if applicable, if not applicable please do not delete your existing 'Response to Reviewers' file.)

- a track-changes file indicating any changes that you have made to the manuscript.

NOTE: If Supporting Information files are included with your article, note that these are not copyedited and will be published as they are submitted. Please ensure that these files are legible and of high quality (at least 300 dpi) in an easily accessible file format. For this reason, please be aware that any references listed in an SI file will not be indexed. For more information, see our Supporting Information guidelines:

https://journals.plos.org/plosbiology/s/supporting-information

*Published Peer Review History*

Please note that you may have the opportunity to make the peer review history publicly available. The record will include editor decision letters (with reviews) and your responses to reviewer comments. If eligible, we will contact you to opt in or out. Please see here for more details:

https://blogs.plos.org/plos/2019/05/plos-journals-now-open-for-published-peer-review/

*Press*

Should you, your institution's press office or the journal office choose to press release your paper, please ensure you have opted out of Early Article Posting on the submission form. We ask that you notify us as soon as possible if you or your institution is planning to press release the article.

*Protocols deposition*

To enhance the reproducibility of your results, we recommend that if applicable you deposit your laboratory protocols in protocols.io, where a protocol can be assigned its own identifier (DOI) such that it can be cited independently in the future. Additionally, PLOS ONE offers an option for publishing peer-reviewed Lab Protocol articles, which describe protocols hosted on protocols.io. Read more information on sharing protocols at https://plos.org/protocols?utm_medium=editorial-email&utm_source=authorletters&utm_campaign=protocols

Please do not hesitate to contact me should you have any questions.

Best wishes,

Richard

Richard Hodge, PhD

Senior Editor, PLOS Biology

rhodge@plos.org

------------------------------------------------------------------------

Reviewer remarks:

Reviewer #1: Well I feel like a dolt for giving such a mild-mannered and brief review compared to the two incredibly detailed critiques you received from the other two reviewers! Thankfully your ms made it onto the desks of two of the foremost experts in this field (plus me!). Having read through their reviews, your responses, (and the few minor responses warranted to my comments), and your revised ms, it's clear that this version has been greatly improved. In retrospect, I agree with them that - for this type of paper - the emphasis should have been more on the technique (both in terms of greater technical detail and greater discussion of applicability) than the BoP example. Your revision accomplishes this change of focus nicely. I commend the authors on a thoughtful and nuanced revision.

Reviewer #2 (Daniel Hanley, signs review): Please see the attached PDF

Reviewer #3 (Jolyon Troscianko, signs review): Overall the authors have done a very good job at addressing my concerns, and those of the other reviewers, making a number of significant changes to the manuscript and analyses. Although the authors haven't provided new evidence that this method is superior to others (e.g. benchmarking), their discussion provides readers with enough information to make an informed judgement, and the method is a valuable addition to the field.

Minor points:

Re. the addition of temporal resolution to table 1, I would say that spectrometers should be high (they measure all the dimensions they can measure near-instantaneously). Multispectral should be high for RGB, and medium for wider spectral bands, and hyperspectral should be medium or low. Obviously, hyperspectral methods must trade off the dimensions against each-other, so will typically be slower. Otherwise, you could specify that the table only refers to techniques in the vis-UV range.

Regarding the spectral filtering, FigS2 is useful, and the raw spectra don't look bad. The opsin curves applied to these raw spectra will be smooth, and will also have the effect of smoothing out the noise substantially. In any case, the effects are now easy for the reader to see.

Regarding costs, thanks for adding these - pretty pricey! I would disagree that costs are coming down though. Off-the-shelf hyperspectral imagers simply don't have a large enough consumer base, and UV optics and light sources haven't become cheaper in over the past 15 years or so I've been paying attention. Given the renewed focus on methods, I would still recommend you add an estimate of the full cost, presumably nearer $70k?

Attachment

Submitted filename: PBiology-D-23-03297_r2 comments.pdf

pbio.3002867.s005.pdf (210KB, pdf)

Decision Letter 3

Richard Hodge

27 Sep 2024

Dear Ben,

On behalf of my colleagues and the Academic Editor, Gail Patricelli, I am pleased to say that we can accept your manuscript for publication, provided you address any remaining formatting and reporting issues. These will be detailed in an email you should receive within 2-3 business days from our colleagues in the journal operations team; no action is required from you until then. Please note that we will not be able to formally accept your manuscript and schedule it for publication until you have completed any requested changes.

In addition, please ensure that the data deposited in the Dryad database is now made publicly available at this stage.

Please take a minute to log into Editorial Manager at http://www.editorialmanager.com/pbiology/, click the "Update My Information" link at the top of the page, and update your user information to ensure an efficient production process.

PRESS

We frequently collaborate with press offices. If your institution or institutions have a press office, please notify them about your upcoming paper at this point, to enable them to help maximise its impact. If the press office is planning to promote your findings, we would be grateful if they could coordinate with biologypress@plos.org. If you have previously opted in to the early version process, we ask that you notify us immediately of any press plans so that we may opt out on your behalf.

We also ask that you take this opportunity to read our Embargo Policy regarding the discussion, promotion and media coverage of work that is yet to be published by PLOS. As your manuscript is not yet published, it is bound by the conditions of our Embargo Policy. Please be aware that this policy is in place both to ensure that any press coverage of your article is fully substantiated and to provide a direct link between such coverage and the published work. For full details of our Embargo Policy, please visit http://www.plos.org/about/media-inquiries/embargo-policy/.

Thank you again for choosing PLOS Biology for publication and supporting Open Access publishing. We look forward to publishing your study. 

Best wishes, 

Richard

Richard Hodge, PhD

Senior Editor, PLOS Biology

rhodge@plos.org

PLOS

Empowering researchers to transform science

Carlyle House, Carlyle Road, Cambridge, CB4 3DN, United Kingdom

ORCiD I plosbio.org I @PLOSBiology I Blog

California (U.S.) corporation #C2354500, based in San Francisco

Associated Data

    This section collects any data citations, data availability statements, or supplementary materials included in this article.

    Supplementary Materials

    S1 Appendix. Supplementary text and figures (Tables A and B and Figs A–V).

    Raw data and code for all supplementary figures are available (see the Data availability section) (.PDF).

    (PDF)

    pbio.3002867.s001.pdf (10.9MB, pdf)
    S1 Video. Video illustrating 3D models of all specimens (2 King bird-of-paradise, 2 Magnificent bird-of-paradise, and 2 hybrid King of Holland’s bird-of-paradise).

    In the first half of the video, the models are colored according to the RGB images used to generate the models (see main text and S1 Appendix). In the second half of the video, models are colored according to the surface normals of the visible polygons in the video. Polygons facing left relative to the camera are teal, while those facing right are magenta. Polygons facing up relative to the camera are darker, while those facing down are lighter. Models are to scale. The sphere in the bottom left has diameter 2 cm, and the cube in the bottom right has edge length 2 cm. (.MP4).

    (MP4)

    Download video file (8.8MB, mp4)
    Attachment

    Submitted filename: Hogan and Stoddard Review.pdf

    pbio.3002867.s003.pdf (334.4KB, pdf)
    Attachment

    Submitted filename: Hogan_Stoddard_ResponseToReviewers.pdf

    pbio.3002867.s004.pdf (291.3KB, pdf)
    Attachment

    Submitted filename: PBiology-D-23-03297_r2 comments.pdf

    pbio.3002867.s005.pdf (210KB, pdf)
    Attachment

    Submitted filename: Hogan_Stoddard_ResponseToReviewers2.pdf

    pbio.3002867.s006.pdf (138.7KB, pdf)

    Data Availability Statement

    All hyperspectral images and all code for this analysis are available on Dryad (DOI: 10.5061/dryad.j0zpc86nf) and GitHub (https://github.com/bghogan/HyperBirdOfParadise), respectively. An archived copy of the code is stored on Zenodo (DOI: 10.5281/zenodo.13749562).


    Articles from PLOS Biology are provided here courtesy of PLOS

    RESOURCES