Computer vision cracks the leaf code
Edited by Andrew H. Knoll, Harvard University, Cambridge, MA, and approved February 1, 2016 (received for review December 14, 2015)
Significance
The botanical value of angiosperm leaf shape and venation (“leaf architecture”) is well known, but the astounding complexity and variation of leaves have thwarted efforts to access this underused resource. This challenge is central for paleobotany because most angiosperm fossils are isolated, unidentified leaves. We here demonstrate that a computer vision algorithm trained on several thousand images of diverse cleared leaves successfully learns leaf-architectural features, then categorizes novel specimens into natural botanical groups above the species level. The system also produces heat maps to display the locations of numerous novel, informative leaf characters in a visually intuitive way. With assistance from computer vision, the systematic and paleobotanical value of leaves is ready to increase significantly.
Abstract
Understanding the extremely variable, complex shape and venation characters of angiosperm leaves is one of the most challenging problems in botany. Machine learning offers opportunities to analyze large numbers of specimens, to discover novel leaf features of angiosperm clades that may have phylogenetic significance, and to use those characters to classify unknowns. Previous computer vision approaches have primarily focused on leaf identification at the species level. It remains an open question whether learning and classification are possible among major evolutionary groups such as families and orders, which usually contain hundreds to thousands of species each and exhibit many times the foliar variation of individual species. Here, we tested whether a computer vision algorithm could use a database of 7,597 leaf images from 2,001 genera to learn features of botanical families and orders, then classify novel images. The images are of cleared leaves, specimens that are chemically bleached, then stained to reveal venation. Machine learning was used to learn a codebook of visual elements representing leaf shape and venation patterns. The resulting automated system learned to classify images into families and orders with a success rate many times greater than chance. Of direct botanical interest, the responses of diagnostic features can be visualized on leaf images as heat maps, which are likely to prompt recognition and evolutionary interpretation of a wealth of novel morphological characters. With assistance from computer vision, leaves are poised to make numerous new contributions to systematic and paleobotanical studies.
Sign up for PNAS alerts.
Get alerts for new articles, or get an alert when an article is cited.
Leaves are the most abundant and frequently fossilized plant organs. However, understanding the evolutionary signals of angiosperm leaf architecture (shape and venation) remains a significant, largely unmet challenge in botany and paleobotany (1–3). Leaves show tremendous variation among the hundreds of thousands of angiosperm species, and a single leaf can contain many thousands of vein junctions. Numerous living angiosperm clades supported by DNA data do not have consistently recognized leaf or even reproductive characters (4); the clades’ fossil records are often lacking or poorly known (5) but are probably “hiding in plain sight” among the millions of fossil leaves curated at museums. Leaf fossils are most commonly preserved in isolation, lacking standard botanical cues such as arrangement, organization, stipules, trichomes, and color (6). Moreover, improved ability to identify living foliage would advance field botany because most plants flower and fruit infrequently (7, 8).
The only major survey of leaf architectural variation is more than 40 y old (1), long preceding the reorganization of angiosperm phylogeny based on molecular data (9, 10). Traditional leaf architectural methods (11, 12) are extremely labor-intensive, inhibiting detailed analyses of leaf variation across angiosperms; accordingly, recent studies have been limited in taxonomic and character coverage (3, 13, 14). Modern machine learning and computer vision methods offer a transformative opportunity, enabling the quantitative assessment of leaf features at a scale never before remotely possible. Many significant computer vision contributions have focused on leaf identification by analyzing leaf-shape variation among species (15–19), and there is community interest in approaching this problem through crowd-sourcing of images and machine-identification contests (see www.imageclef.org). Nevertheless, very few studies have made use of leaf venation (20, 21), and none has attempted automated learning and classification above the species level that may reveal characters with evolutionary significance. There is a developing literature on extraction and quantitative analyses of whole-leaf venation networks (22–25). However, those techniques mostly have not addressed identification problems, and they are methodologically limited to nearly undamaged leaves, which are rare in nature and especially in the fossil record.
Despite the numerous difficulties involved in analyzing leaf architecture, experienced botanists regularly identify leaves to species and higher taxonomic categories quickly and correctly, often without the use of verbally defined characters (“leaf gestalt”). Here, we attempt to automate leaf gestalt, using computer vision and machine learning on a database of several thousand cleared-leaf images to test whether an algorithm can generalize across diverse genera and species to learn, recognize, and reveal features of major angiosperm clades. This challenge is wholly different from that of identifying species, as typically practiced in the computer vision field (see above discussion), because the leaf variation within a species is nearly insignificant compared with that among species within a family or order, as seen in example leaf pairs from various families (Fig. 1). Nevertheless, generations of botanists and paleobotanists have emphasized the recognizability of leaf traits at the family and higher levels (1, 7, 26). We test this fundamental supposition by asking whether a machine vision algorithm can perform a comparable task, learning from examples of large evolutionary categories (families or orders) to correctly discriminate other members of those lineages. We then visualize the machine outputs to display which regions of individual leaves are botanically informative for correct assignment above the species level.
Fig. 1.
To train the system (see Materials and Methods for details), we vetted and manually prepared 7,597 images of vouchered, cleared angiosperm leaves from 2,001 phylogenetically diverse genera and ca. 6,949 species from around the world, and we updated nomenclature to the Angiosperm Phylogeny Group (APG) III system (10). Cleared leaves (Fig. 1) are mounted specimens whose laminar tissues have been chemically altered and then stained to allow close study of venation patterns in addition to the leaf outlines. Importantly, we did not, except in severe cases (Materials and Methods), exclude leaves with common imperfections such as bubbles or crystallization in the mounting medium, leaf damage (insect, fungal, mechanical), or imaging defects (poor focus, background artifacts). Four sets of images were processed for the analyses: those from the (i) 19 families and (ii) 14 orders with minimum sample sizes of 100 images each and those from the (iii) 29 families and (iv) 19 orders with minimum sample sizes of 50 images each. In all analyses, random halves of the image sets were used to train the system and the other half to test, and this procedure was repeated 10 times. Accuracy measures correspond to averages and SDs across these 10 repetitions.
The computer vision algorithm (see Materials and Methods for details; Fig. S1) used a sparse coding approach, which learns a codebook (Fig. 2) from training images and uses it to represent a leaf image as a linear combination of a small number of elements (27). Similar methods have been applied to a variety of problems (28–30). Of interest here regarding “leaf gestalt,” seminal work in vision science (27) showed that training a sparse coding model on images of natural scenes resulted in receptive fields that resemble those of the mammalian visual cortex. Our approach used Scale Invariant Feature Transform (SIFT) visual descriptors (31) extracted from individual locations on the training leaf images to learn a codebook (Fig. 2) and then to model each location as a linear combination of the codebook elements. Support Vector Machine (SVM) (32), a machine-learning classifier, was trained on the pooled maximum coefficients for each codebook element and the associated taxonomic labels (families or orders). Training resulted in a classification function that was applied to predict familial and ordinal identifications of novel images. Heat maps of visualized diagnostic responses were generated directly on individual leaf images (Fig. 1).
Fig. 2.
Fig. S1.
To our knowledge, no prior approach has adopted this powerful combination of tools for leaf identification, used cleared leaves or analyzed leaf venation for thousands of species, attempted to learn and discriminate traits of evolutionary lineages above the species level, or directly visualized informative new characters.
Results
Classification accuracy was very high (Fig. 3 and Fig. S2), especially considering the well-known challenges of identifying leaves to higher taxa visually and the deliberate inclusion of many imperfect specimens and low-quality images. Accuracy for the 19 families with ≥100 images was 72.14 ± 0.75%, ca. 13 times better than random chance (5.61 ± 0.54%; accuracy is defined as the grand mean and SD of the 10 trial averages of the diagonal values of the confusion matrices, as in Fig. 3). Thus, the algorithm successfully generalized across a few thousand highly variable genera and species to recognize major evolutionary groups of plants (see also Materials and Methods, Success in Generalization of Learning across Taxonomic Categories and Collections). Significantly, the system scored high accuracies (>70%) for several families that are not usually considered recognizable from isolated blades, including Combretaceae, Phyllanthaceae, and Rubiaceae. Of great biological interest, our heat maps directly illustrate leaf features that are diagnostic for classification (Fig. 1; extended heat-map outputs are hosted on Figshare, https://doi.org/10.6084/m9.figshare.1521157).
Fig. 3.
Fig. S2.
Accuracy at the ordinal level was also many times greater than chance. This result partly reflects the recognition of the constituent families rather than generalization across families to recognize orders; however, even after testing for this effect, there was still a significant signal for generalization (Materials and Methods). On first principles, generalization across genera to recognize families should be more successful, as seen here, than generalization across families to distinguish orders. The genera within a family are usually much more closely related to each other, and thus likely to be phenotypically similar, than are the families within an order. Hypothetically, an increased number of well-sampled families in targeted orders would allow improved recognition of orders and generalization across families (see also Fig. S3). Artifacts specific to particular image sets contributed noise but were largely overcome (Materials and Methods), boding well for applying the system to a range of new and challenging datasets.
Fig. S3.
For a benchmark comparison, we classified the images from the 19 families in Fig. 3A via holistic analysis of leaf shape alone. We compared the well-known shape-context algorithm (15) (Materials and Methods) with our sparse coding algorithm on extracted leaf outlines. For family classification at the 100-image minimum, shape context performed above random chance (12.44 ± 0.74% vs. 5.61 ± 0.54% chance result) but below sparse coding of the same outlines (27.94 ± 0.85%), and both scored far below sparse coding of images with venation (ca. 72%; Fig. 3A). These results help to demonstrate the diagnostic significance of both leaf venation and shape. The importance of leaf venation is long established (26). However, the results from leaf outlines refute the widely held view, seen in many of the hundreds of papers cited in the extended bibliography of ref. 33, that leaf shape is principally selected by climate and that phylogeny plays a negligible role.
Discussion
All of our results validate the hypotheses that higher plant taxa contain numerous diagnostic leaf characters and that computers can acquire, generalize, and use this knowledge. In addition to the potential for computer-assisted recognition of living and fossil leaves, our algorithm represents an appealing, novel approach to leaf architecture. Conventional leaf architecture (12) currently employs ca. 50 characters to describe a range of characters that are usually visible over large areas of a specimen, from patterns of shape and major venation to categories of fine venation and marginal configuration. Scoring these characters requires immense manual effort, and their systematic significance is poorly understood. In sharp contrast, the learned codebook elements derived here encode small leaf features, mostly corresponding to previously undescribed orientations and junctions of venation and leaf margins (Fig. 2); the codebooks and diagnostic significance values (Fig. 1) are automatically generated.
Our heat maps (Fig. 1 and extended outputs at https://doi.org/10.6084/m9.figshare.1521157) provide a trove of novel information for developing botanical characters that probably have phylogenetic significance as well as for visual training (improved “gestalt”). The dark red patches on the heat maps indicate leaf locations with the most significant diagnostic responses to codebook elements, providing numerous leads for the conventional study of previously overlooked patterns. Diagnostic features appear to occur most often at tertiary and higher-order vein junctions, as well as along the margins and associated fine venation of both toothed and untoothed leaves. Close examination of the heat maps reveals variation among families. For example, responses at leaf teeth are strongest along basal tooth flanks in Fagaceae and Salicaceae, versus tooth apices in Rosaceae; in Ericaceae, responses are located at teeth and on minute marginal variations that do not form teeth. Responses at margins tend to be concentrated in the medial third of the lamina in Annonaceae and Combretaceae versus the distal half in Fabaceae, Phyllanthaceae, and Rubiaceae. Tertiary-vein junctions are significant across groups, with apparently higher sensitivity at junctions with primary veins in Fagaceae, perhaps reflecting the tendency toward zigzag midveins in that family; with secondary and intersecondary veins in Apocynaceae and Betulaceae; and with other tertiaries and finer veins in several families such as Anacardiaceae. Correspondingly, many Anacardiaceae are known for having unusual tertiary venation (13, 34).
Many of the analyzed families have numerous species with visually distinctive, generally holistic traits that are widely applied in field identification (1, 7, 8). Examples include acute basal secondary veins and low vein density in Lauraceae; palmate venation in Malvaceae; the prevalence of multiple, strongly curving primary veins in Melastomataceae; the distinctive perimarginal veins of Myrtaceae; and the characteristic (1) teeth of Betulaceae, Fagaceae, Rosaceae, and Salicaceae. However, the small image regions used here mostly bypassed these patterns, and using larger image regions (by down-sampling the images) significantly reduced accuracy (Table 1). Nevertheless, the novel, fine-scale features will probably improve understanding of conventional characters. For example, the leaf margins of Melastomataceae show diagnostic responses and generally follow the trends of the well-known, strongly curving primary veins. The perimarginal vein configurations of Myrtaceae appear quite responsive, especially the little-studied (35) ultimate venation. Leaf teeth are closer in size to the patches and are often “hot,” but the leaf teeth’s responsive areas are difficult to relate to the standard tooth-type classifications (1), which are now ripe for revision.
Table 1.
Scenario | 64 | 128 | 256 | 512 | 768 | 1,024 | 1,536 | 2,048 |
---|---|---|---|---|---|---|---|---|
Families, ≥100 images, % | 15.82 ± 3.17 | 40.50 ± 1.17 | 58.07 ± 0.93 | 67.57 ± 0.86 | 71.54 ± 0.67 | 72.14 ± 0.75 | 73.43 ± 0.92 | 72.08 ± 7.27 |
Orders, ≥100 images, % | 10.56 ± 1.85 | 29.37 ± 1.67 | 44.51 ± 0.87 | 52.95 ± 0.78 | 57.21 ± 0.97 | 57.26 ± 0.81 | 58.54 ± 0.23 | 60.40 ± 1.70 |
Analyses in this study were all done at 1,024-pixel resolution (Fig. 3). The resolution is pixels along longest dimension. Uncertainties are SDs.
In sum, along with the demonstration that computers can recognize major clades of angiosperms from leaf images and the promising outlook for computer-assisted leaf classification, our results have opened a tap of novel, valuable botanical characters. With an increased image library and testing on well-preserved fossils, analyses from heat-map data are likely to achieve the longstanding goals of quantifying leaf-phylogenetic signal across angiosperms and systematic placements of large numbers of fossil leaves.
Materials and Methods
Cleared Leaf Dataset.
We assembled a dataset of 7,597 vetted images of cleared leaves (or leaflets for compound leaves), representing nearly as many species, from 2,001 genera and ca. 6,949 species collected around the world (Dataset S1). The images mostly represent woody angiosperms and come from three herbarium-vouchered sources: (i) the Jack A. Wolfe USGS National Cleared Leaf Collection, housed at the Smithsonian Institution, 5,063 images (clearedleavesdb.org); (ii) our scans of the E. P. Klucking Leaf Venation Patterns book series (36), 2,305 images; and (iii) the Daniel Axelrod Cleared Leaf Collection, University of California, Berkeley, 229 images (ucmp.berkeley.edu/science/clearedleaf.php). Images used were those with white space present around the leaf, more than ca. 85% of leaf visible, midvein not folded or discontinuously broken, and at least some tertiary venation visible. Importantly, we did not screen other common sources of noise such as tears, folds in the lamina, insect or fungal damage, and bubbles or crystallization in the mounting medium. Many images from the Klucking series were of notably low quality and bore a moiré pattern from printing.
The selection of taxonomic groups was entirely opportunistic, based on the cleared-leaf resources available. However, the resulting collection is highly diverse phylogenetically because Wolfe, Klucking, and Axelrod each attempted to maximize their coverage of genera within families. We considered families with at least 100 images and randomly split the images from each family so that one half was used to train the classifier and the other to test the system (random halving was used as a conservative splitting procedure; we note that increasing the relative proportion of training images further increased accuracies in all scenarios). This cutoff yielded 19 families selected with at least 100 images each (Fig. 3A): Anacardiaceae, Annonaceae, Apocynaceae, Betulaceae, Celastraceae, Combretaceae, Ericaceae, Fabaceae, Fagaceae, Lauraceae, Malvaceae, Melastomataceae, Myrtaceae, Passifloraceae, Phyllanthaceae, Rosaceae, Rubiaceae, Salicaceae, and Sapindaceae. We also attempted a lower cutoff of 50 images per family, adding 10 additional families (Fig. S2A): Bignoniaceae, Euphorbiaceae, Meliaceae, Menispermaceae, Moraceae, Oleaceae, Proteaceae, Rhamnaceae, Rutaceae, and Sapotaceae. We repeated the random-halving procedure 10 times and here report the mean accuracy and SD over all 10 splits (Fig. 3 and Fig. S2). We conducted a similar protocol for classification at the ordinal level, yielding 14 APG III orders selected at the 100 image minimum (Fig. 3B): Celastrales, Ericales, Fabales, Fagales, Gentianales, Lamiales, Laurales, Magnoliales, Malpighiales, Malvales, Myrtales, Ranunculales, Rosales, and Sapindales. We also attempted a lower cutoff of 50 images per order, adding five more orders (Fig. S2B): Cornales, Oxalidales, Proteales, Santalales, and Saxifragales.
We prepared each image manually to focus processing power on the leaf blades and to minimize uninformative variation. We rotated images to have vertical leaf orientation, cropped them to an approximately uniform white-space border, adjusted them to gray-scale with roughly consistent contrast, removed petioles, and erased most background objects. Sets of all prepared images are available on request from P.W. All images were resized so that their maximum width or height dimension was 1,024 pixels, maintaining the original aspect ratio. This resolution offered the optimal balance between performance and computing cost (Table 1). We updated generic names using multiple standard sources and revised familial and ordinal assignments using APG III (10).
The resulting collection (Dataset S1) contains images of 182 angiosperm families, 19 of them with ≥100 images each and totaling 5,314 images (Fig. 3A). We considered each family as an independent category in the computer vision system, regardless of its relationships with the other groups, thus testing the traditional idea (1, 7, 36) that leaves within angiosperm families have visually similar features despite the enormous morphological variation among constituent genera and species.
Computer Vision Algorithm.
We analyzed leaf architecture using a sparse coding learning approach (code in Dataset S2). This method aims to represent image spaces with linear combinations of a small number of elements (27) from a visual codebook that is learned from training images. In our approach (Fig. S1), adapted from ref. 28, SIFT visual descriptors (31) were first extracted from all database images. These 128-dimensional descriptors were computed from small (16 × 16 pixel) image patches that are densely sampled (with a stride of 6 pixels corresponding to a 10-pixel overlap between sampled patches). This process resulted in about 2,000 to ∼30,000 SIFT descriptors generated per image, depending on image size. We then executed 10 random splits of the image set into training and testing halves as described above. For each split, the algorithm randomly sampled 200,000 SIFT descriptors from the training set to learn an overcomplete codebook of d =1,024 elements (Fig. 2) using the SPArse Modeling Software (SPAMS) toolbox (spams-devel.gforge.inria.fr).
Next, we applied a point-wise linear transform to both the training and test images: for a given location within an image, the corresponding SIFT descriptor was modeled as a linear combination of the learned codebook elements. After this step, each image location became associated with a set of d coefficients, each corresponding to a different codebook element. Because of the sparsity constraint, at any location only a few of these coefficients are nonzero. We subsequently implemented a multiscale spatial pyramid (28), whereby a maximum was computed for each coefficient over all locations, within each cell of a pyramid of grids placed across the whole image at three different scales (1 × 1, 2 × 2, and 4 × 4 grids, totaling 21 cells). For each input image, this maximum pooling operation resulted in d max coefficient values for each of the 21 cells. Concatenating these 21 values for each of the d = 1,024 coefficients yielded a 21,504-dimensional feature vector for each input image. SVM (32) was then trained using the feature vectors from a training set of leaf images and their associated taxonomic labels (families or orders). The resulting learned classification function was then applied to predict family or order identifications for a testing set of leaves from their feature vectors. Here, we used the LIBSVM software (https://csie.ntu.edu.tw/∼cjlin/libsvm), using a multiclass setting with a radial basis function kernel. Chance level was assessed using random permutations of the taxonomic labels of the training images.
Success in Generalization of Learning Across Taxonomic Categories and Collections.
To assess the robustness of the machine-vision system, we evaluated the extent to which the algorithm succeeds in generalizing recognition across taxonomic categories versus basing its decisions primarily on detecting the presence or absence of constituent subcategories that it has learned [i.e., (i) families within orders and (ii) genera within families]. We addressed this issue by redoing the analyses with and without systematic separation of the respective subcategories among each training vs. testing half, with 10 repetitions as in the main experiments.
To test the first case, generalization across families to detect orders, we conducted an experiment in which we selected all orders that each had at least two families with at least 50 images apiece, namely Ericales, Fagales, Gentianales, Lamiales, Malpighiales, Myrtales, Rosales, and Sapindales. We eliminated any remaining families with fewer than 50 images each and then trained and tested the system in two scenarios. First, half the leaf images from each order was randomly selected for inclusion in the training pool, regardless of family membership, whereas the other half was used in the testing pool, thus allowing family overlaps (accuracy of order recognition: 67.49 ± 0.88%; chance level 12.38 ± 0.77%). Second, half the families from each order was randomly selected for training and the remaining families for testing, thus excluding overlaps (accuracy of order recognition: 26.52 ± 2.33%; chance level: 12.31 ± 1.72%). The fact that the system accuracy was much larger for the first than the second scenario suggests comparatively low success in generalization across families, but the success rate was still more than double that of chance. Thus, ordinal classification in the machine-vision system has a substantial contribution from generalization at the family level, but even with these small numbers of families per order, there is still a significant signal for generalization across families.
To test the second case, generalization across genera to recognize families, we conducted a similar experiment in which we selected the families that each contain at least two genera that (i) each have at least 25 images and (ii) together total at least 100 images. The resulting six families were Combretaceae, Ericaceae, Lauraceae, Myrtaceae, Phyllanthaceae, and Salicaceae. After eliminating remaining genera with fewer than 25 images each, we trained and tested the system in two scenarios. First, half the images from each family were randomly sampled for training, regardless of genus, whereas the other half was used for testing, thus allowing generic overlaps (accuracy of family recognition: 88.01 ± 0.93%; chance level: 16.37 ± 3.09%). Second, the training and testing samples were obtained by randomly selecting half the genera for the training pool and those remaining for the testing pool; in other words, no overlap of genera was allowed between the training and testing halves (accuracy of family recognition: 70.59 ± 3.73%; chance level: 17.37 ± 1.83%). The fact that accuracy was larger for the first scenario but remained very significantly above chance for the second suggests that the system exhibits robust generalization across genera to recognize families, and is not, for example, dependent on well-sampled genera such as Acer (maples, Sapindaceae) and Quercus (oaks, Fagaceae). The higher possibility of identification errors for genera than for families on the original samples biases against this positive result, which confirms conventional wisdom regarding the general identifiability of families as well as the potential for machine vision to recognize finer taxonomic levels.
We also explored the extent to which the machine-vision system may be exploiting possible biases associated with individual collections, such as the moiré printing pattern in the Klucking images. We conducted an experiment in which we selected all orders that had at least 50 images each in both the Klucking and the Wolfe collections, namely Laurales, Magnoliales, Malpighiales, Malvales, and Myrtales. We trained and tested the system in three scenarios, following the same methods that we used in the main experiments. First, we trained on the Klucking and tested on the Wolfe image collection (accuracy of order recognition: 37.44 ± 0.66%; chance level: 19.40 ± 3.53%). Second, we trained on the Wolfe and tested on the Klucking collection (accuracy of order recognition: 41.12 ± 0.94%; chance level: 20.04 ± 5.00%). Third, we combined the two sources and randomly sampled half the images from each order for training and half for testing (accuracy of order recognition: 64.68 ± 2.65%; chance level: 18.96 ± 3.78%). The higher accuracy in scenario 3 suggests that biases exist (e.g., image background) that can be picked up by the system. However, the significant generalization of the system in scenarios 1 and 2, especially when taking into account the overall much lower image quality of the Klucking collection, suggests that these biases only account for a small fraction of the overall system accuracy.
Benchmark Using Shape Context.
The shape-context algorithm has been successfully used for species-level leaf classification (15). We tested both the shape-context and sparse-coding algorithms on leaf outlines extracted from the same image subset used for the 19 family, 100-image-minimum scenario shown in Fig. 3A. Leaves were segmented automatically by combining the Berkeley contours detection system (37) with second-order Gaussian derivative filters, and the corresponding closed boundary was then filled to obtain foreground masks. Shape-context features were then extracted from the leaf silhouette, following the approach of Ling and Jacobs (15). Leaf contours were sampled uniformly at 128 landmark points, and shape-context descriptors were extracted at every sampled point, using source code available online (www.dabi.temple.edu/∼hbling/code_data.htm). This procedure led to a 96 × 128 dimensional visual representation (the shape context) for each leaf image. Matching between two leaves’ shape contexts was obtained through dynamic programming, where matching cost is used to measure similarity. The pairwise similarities were then used to derive a Kernel matrix that was fed to an SVM, which was trained and tested using the same pipeline as for the sparse model described in Materials and Methods, Computer Vision Algorithm.
Data Availability
Data deposition: Data reported in this paper have been deposited in the Figshare repository (https://doi.org/10.6084/m9.figshare.1521157).
Acknowledgments
We thank A. Young and J. Kissell for image preparations, the anonymous reviewers for helpful comments, Y. Guo for software, A. Rozo for book scanning, and D. Erwin for assistance with the Axelrod collection. We acknowledge financial support from the David and Lucile Packard Foundation (P.W.); National Science Foundation Early Career Award IIS-1252951, Defense Advanced Research Projects Agency Young Investigator Award N66001-14-1-4037, Office of Naval Research Grant N000141110743, and the Brown Center for Computation and Visualization (T.S.); and National Natural Science Foundation of China Grant 61300111 and Key Program Grant 61133003 (to S.Z.).
Supporting Information
Supporting Information (PDF)
Supporting Information
- Download
- 414.14 KB
Dataset_S02 (PDF)
Supporting Information
- Download
- 159.90 KB
pnas.1524473113.sd01.xlsx
- Download
- 436.96 KB
References
1
LJ Hickey, JA Wolfe, The bases of angiosperm phylogeny: Vegetative morphology. Ann Mo Bot Gard 62, 538–589 (1975).
2
LJ Hickey, JA Doyle, Early Cretaceous fossil evidence for angiosperm evolution. Bot Rev 43, 2–104 (1977).
3
JA Doyle, Systematic value and evolution of leaf architecture across the angiosperms in light of molecular phylogenetic analyses. Cour Forschungsinst Senckenb 258, 21–37 (2007).
4
DE Soltis, PS Soltis, PK Endress, MW Chase Phylogeny and Evolution of Angiosperms (Sinauer, Sunderland, MA, 2005).
5
EM Friis, PR Crane, KR Pedersen Early Flowers and Angiosperm Evolution (Cambridge Univ Press, Cambridge, UK, 2011).
6
P Wilf, Fossil angiosperm leaves: Paleobotany’s difficult children prove themselves. Paleontol Soc Pap 14, 319–333 (2008).
7
AH Gentry A Field Guide to the Families and Genera of Woody Plants of Northwest South America (Colombia, Ecuador, Peru) (Conservation International, Washington, DC, 1993).
8
R Keller Identification of Tropical Woody Plants in the Absence of Flowers: A Field Guide (Birkhäuser Verlag, Basel, 2004).
9
MW Chase, et al., Phylogenetics of seed plants: An analysis of nucleotide sequences from the plastid gene rbcL. Ann Mo Bot Gard 80, 528–580 (1993).
10
; Angiosperm Phylogeny Group, An update of the Angiosperm Phylogeny Group classification for the orders and families of flowering plants: APG III. Bot J Linn Soc 161, 105–121 (2009).
11
LJ Hickey, A revised classification of the architecture of dicotyledonous leaves. Anatomy of the Dicotyledons, eds CR Metcalfe, L Chalk (Clarendon, Oxford), pp. 25–39 (1979).
12
B Ellis, et al. Manual of Leaf Architecture (Cornell Univ Press, Ithaca, NY, 2009).
13
M Martínez-Millán, SRS Cevallos-Ferriz, Arquitectura foliar de Anacardiaceae. Rev Mex Biodivers 76, 137–190 (2005).
14
MR Carvalho, FA Herrera, CA Jaramillo, SL Wing, R Callejas, Paleocene Malvaceae from northern South America and their biogeographical implications. Am J Bot 98, 1337–1355 (2011).
15
H Ling, DW Jacobs, Shape classification using the inner-distance. IEEE Trans Pattern Anal Mach Intell 29, 286–299 (2007).
16
PN Belhumeur, et al., Searching the world’s herbaria: A system for visual identification of plant species. Computer Vision – ECCV 2008, Lecture Notes in Computer Science, eds Forsyth D, Torr P, Zisserman A (Springer, Berlin, Heidelberg) Vol 5305, pp 116–129. (2008).
17
R Hu, W Jia, H Ling, D Huang, Multiscale distance matrix for fast plant leaf recognition. IEEE Trans Image Process 21, 4667–4672 (2012).
18
N Kumar, et al., Leafsnap: A computer vision system for automatic plant species identification. Computer Vision – ECCV 2012, Part II, Lecture Notes in Computer Science, eds Fitzgibbon A, Lazebnik S, Perona P, Sato Y, Schmid C (Springer, Berlin, Heidelberg) Vol 7573, pp 502–516. (2012).
19
H Laga, S Kurtek, A Srivastava, M Golzarian, SJ Miklavcic, A Riemannian elastic metric for shape-based plant leaf classification, 2012 International Conference on Digital Image Computing Techniques and Applications (DICTA), Fremantle, Australia, December 3–5, 2012, pp 1–7, 10.1109/DICTA.2012.6411702.
20
Y Nam, E Hwang, D Kim, A similarity-based leaf image retrieval scheme: Joining shape and venation features. Comput Vis Image Underst 110, 245–259 (2008).
21
MG Larese, et al., Automatic classification of legumes using leaf vein image features. Pattern Recognit 47, 158–168 (2014).
22
S Bohn, B Andreotti, S Douady, J Munzinger, Y Couder, Constitutive property of the local organization of leaf venation networks. Phys Rev E Stat Nonlin Soft Matter Phys 65, 061914 (2002).
23
Y Li, Z Chi, DD Feng, Leaf vein extraction using independent component analysis. IEEE Int Conf Syst Man Cybern 5, 3890–3894 (2006).
24
CA Price, S Wing, JS Weitz, Scaling and structure of dicotyledonous leaf venation networks. Ecol Lett 15, 87–95 (2012).
25
E Katifori, MO Magnasco, Quantifying loopy network architectures. PLoS One 7, e37994 (2012).
26
C von Ettingshausen Die Blatt-Skelete der Dikotyledonen (K. K. Hof-und Staatsdruckerei, Vienna, 1861).
27
BA Olshausen, DJ Field, Emergence of simple-cell receptive field properties by learning a sparse code for natural images. Nature 381, 607–609 (1996).
28
J Yang, K Yu, Y Gong, T Huang, Linear spatial pyramid matching using sparse coding for image classification, IEEE Conf Comput Vision Pattern Recognit, Miami, June 20–26, 2009, pp 1794–1801, 10.1109/CVPR.2009.5206757.
29
JM Hughes, DJ Graham, DN Rockmore, Quantification of artistic style through sparse coding analysis in the drawings of Pieter Bruegel the Elder. Proc Natl Acad Sci USA 107, 1279–1283 (2010).
30
GSVS Sivaram, S Krishna Nemala, M Elhilali, TD Tran, H Hermansky, Sparse coding for speech recognition, IEEE Int Conf Acoustics Speech Signal Process, Dallas, March 14–19, 2010, pp 4346–4349, 10.1109/ICASSP.2010.5495649.
31
DG Lowe, Distinctive image features from scale-invariant keypoints. Int J Comput Vis 60, 91–110 (2004).
32
VN Vapnik The Nature of Statistical Learning Theory (Springer, New York, New York, 2000).
33
SA Little, SW Kembel, P Wilf, Paleotemperature proxies from leaf fossils reinterpreted in light of evolutionary history. PLoS One 5, e15161 (2010).
34
JD Mitchell, DC Daly, A revision of Spondias L. (Anacardiaceae) in the Neotropics. PhytoKeys 55, 1–92 (2015).
35
CC González, Arquitectura foliar de las especies de Myrtaceae nativas de la flora Argentina II: Grupos “Myrteola” y “Pimenta”. Bol Soc Argent Bot 46, 65–84 (2011).
36
EP Klucking Leaf Venation Patterns (J. Cramer, Berlin) Vols 1-8 (1986–1997).
37
P Arbeláez, M Maire, C Fowlkes, J Malik, Contour detection and hierarchical image segmentation. IEEE Trans Pattern Anal Mach Intell 33, 898–916 (2011).
38
C Vondrick, A Khosla, T Malisiewicz, A Torralba, HOGgles: Visualizing object detection features. Proc IEEE Conf Comput Vision, Sydney, Australia, December 1–8, 2013, pp 1-8, 10.1109/ICCV.2013.8.
Information & Authors
Information
Published in
Classifications
Data Availability
Data deposition: Data reported in this paper have been deposited in the Figshare repository (https://doi.org/10.6084/m9.figshare.1521157).
Submission history
Published online: March 7, 2016
Published in issue: March 22, 2016
Keywords
Acknowledgments
We thank A. Young and J. Kissell for image preparations, the anonymous reviewers for helpful comments, Y. Guo for software, A. Rozo for book scanning, and D. Erwin for assistance with the Axelrod collection. We acknowledge financial support from the David and Lucile Packard Foundation (P.W.); National Science Foundation Early Career Award IIS-1252951, Defense Advanced Research Projects Agency Young Investigator Award N66001-14-1-4037, Office of Naval Research Grant N000141110743, and the Brown Center for Computation and Visualization (T.S.); and National Natural Science Foundation of China Grant 61300111 and Key Program Grant 61133003 (to S.Z.).
Notes
This article is a PNAS Direct Submission.
Authors
Competing Interests
The authors declare no conflict of interest.
Metrics & Citations
Metrics
Citation statements
Altmetrics
Citations
Cite this article
113 (12) 3305-3310,
Export the article citation data by selecting a format from the list below and clicking Export.
Cited by
Loading...
View Options
View options
PDF format
Download this article as a PDF file
DOWNLOAD PDFLogin options
Check if you have access through your login credentials or your institution to get full access on this article.
Personal login Institutional LoginRecommend to a librarian
Recommend PNAS to a LibrarianPurchase options
Purchase this article to access the full text.