Classification of CITES-listed and other neotropical Meliaceae wood images using convolutional neural networks
© The Author(s) 2018
Received: 23 August 2017
Accepted: 14 March 2018
Published: 23 March 2018
The current state-of-the-art for field wood identification to combat illegal logging relies on experienced practitioners using hand lenses, specialized identification keys, atlases of woods, and field manuals. Accumulation of this expertise is time-consuming and access to training is relatively rare compared to the international demand for field wood identification. A reliable, consistent and cost effective field screening method is necessary for effective global scale enforcement of international treaties such as the Convention on the International Trade in Endagered Species (CITES) or national laws (e.g. the US Lacey Act) governing timber trade and imports.
We present highly effective computer vision classification models, based on deep convolutional neural networks, trained via transfer learning, to identify the woods of 10 neotropical species in the family Meliaceae, including CITES-listed Swietenia macrophylla, Swietenia mahagoni, Cedrela fissilis, and Cedrela odorata. We build and evaluate models to classify the 10 woods at the species and genus levels, with image-level model accuracy ranging from 87.4 to 97.5%, with the strongest performance by the genus-level model. Misclassified images are attributed to classes consistent with traditional wood anatomical results, and our species-level accuracy greatly exceeds the resolution of traditional wood identification.
The end-to-end trained image classifiers that we present discriminate the woods based on digital images of the transverse surface of solid wood blocks, which are surfaces and images that can be prepared and captured in the field. Hence this work represents a strong proof-of-concept for using computer vision and convolutional neural networks to develop practical models for field screening timber and wood products to combat illegal logging.
In the last decade, international interest in combating illegal logging has been on the rise (e.g. the US Lacey Act 2008; the Australian Illegal Logging Prohibition Act 2012; the European Union Timber Regulation 2013; Japan’s Act on Promotion of Distribution and Use of Legally Logged Wood Products 2016) as has interest in forensic methods to support them [1–3]. Although emphasis on laboratory-based forensic science is common, especially among laboratory scientists, one of the primary roadblocks to meaningful enforcement of these laws is the availability of efficient field-deployable tools for screening timber outside the laboratory . Conceptually separating laboratory-based forensic analysis of specimens submitted as evidence and field-screening of wood and wood products at ports and border crossings is central to defining the context of the problem to be solved and the degree of specificity necessary to solve it in a way that is meaningful in the real world. Because field law enforcement agents are, in most jurisdictions, required to establish some form of probable cause to detain or seize a shipment of wood, tools intended for field deployment should be designed to meet this need efficiently . The threshold of evidence for probable cause or its international analogs is much lower than forensic-level thresholds, so tools for field screening to establish probable cause can provide results with coarser resolution and lesser certainty than laboratory forensic methods. A typical field screening evaluates the veracity of a claim on a import-export form or shipping manifest. For example, a shipping manifest may claim that the wood is Khaya but a field agent determines that the wood is anatomically inconsistent with Khaya and in fact is a better match for Swietenia and so the shipment could be detained while a specimen is submitted for full laboratory forensic analysis.
This kind of field screening of wood has historically been done, if done at all, by human beings with hand lenses and keys, atlases of woods, or field manuals (e.g. [5–10] and others). Such keys are based on the fact that wood structure observed macroscopically shows abundant, characteristic variation typically permitting identification at the suprageneric or generic level, with greater specificity possible by highly trained experts or by accessing microscopic characters in the laboratory. Humans with hand lenses are still the state-of-the-art in the field in most countries,1 but the time and cost embodied in establishing and maintaining this human-based biological domain knowledge, and the variability of skill and accuracy among those applying such knowledge, means this approach is difficult to scale up to keep pace with increased international interest in and demand for field screening of timber and other wood products.
Computer vision has the potential to provide a practical and cost effective way to replace human-based biological domain knowledge for field screening of wood in trade. One of the primary advantages of this potential is the ability to generate reproducible identifications not dependent on individual human training , as long as sufficient images of the woods in question are available for training classifiers and can be captured in the field. In computer vison terms, the problem of image-based wood identification is one of texture-based image classification [12, 13]. Convolutional neural networks have achieved state-of-the-art [14–17] results for image classification in the past few years. While in general convolutional neural networks require large datasets (historically not readily available in the context of wood identification), transfer learning  (“Methods” section) provides a pathway to train competitive image classification models using moderate amounts of data by leveraging pre-trained networks, e.g. ones that have been trained on the ImageNet dataset . Convolutional neural networks trained on the ImageNet dataset have been shown to be powerful off-the-shelf feature extractors  and transfer learning effectively leverages these general purpose feature extractors, with parameter fine tuning, and permits the use of smaller application-specific datasets for training powerful classifiers. Successfully developing a field-deployable computer vision model for commercial wood species that are threatened or endangered [e.g. species proteted by the Convention on the Trade in Endangered Species (CITES)] is a step toward generating a scalable tool for law enforcement to use to combat global illegal logging.
The botanical issue of species delimitation is not a matter purely of taxonomy when it comes to illegal logging and species conservation through vehicles such as CITES. Any law or treaty that identifies and protects organisms at the species level necessarily depends on the taxonomic circumscription of those species as a foundational predicate for defining the protected organisms themselves. The complex interplay of laws for conservation, taxonomy, species circumscription, and the viability of field-level screening and forensic-level identification of those organisms or their derived products has prompted practical changes to species protection levels in CITES (e.g. the promotion of Swietenia macrophylla to be at the same protection level as Swietenia mahagoni and Swietenia humilis in 20032). Prior to this elevation, unscrupulous traders had the ability to claim a shipment was the less-protected species and forensics could not prove otherwise.
In this study we report on highly effective computer-vision classification models, based on deep convolutional neural networks trained via transfer learning, to identify 10 neotropical species in the family Meliaceae, including CITES-listed species Swietenia macrophylla, Swietenia mahagoni, Cedrela fissilis, and Cedrela odorata . We selected taxa that have real-world relevance in international timber trade and/or represent an interesting range of overlapping (inter- and intra-class variability) wood anatomical patterns, structural variability, and distinctness of anatomical pattern at multiple scales (Fig. 1). These models discriminate the various woods based on digital images of the transverse surface of solid wood blocks, using images roughly at a hand lens magnification, thus also suitable for human-mediated provisional identification. The transverse surface of wood specimens at a port, border crossing, or other point of control can be prepared for imaging with a modicum of training and a sharp utility knife. We demonstrate proof-of-concept for image-based wood identification using convolutional neural networks and suggest avenues of future inquiry, to develop and eventually deploy computer vision in the field.
Convolutional neural networks
Convolutional neural networks (CNNs)  are state-of-the-art classifiers [14–17] that have powered many recent advances in image classification. CNNs have a multilayer architecture of convolutional operations interspersed with non-linear activation functions and pooling operations which enable them to learn rich non-linear representations for image classification. The parameters of CNNs can be learnt automatically in an end-to-end fashion given sufficient data. While automated representation learning from data is an attractive feature, training CNNs from scratch typically requires large datasets which may not be available. A practical way to build CNN based image classifiers using moderately sized datasets is through transfer learning where features learnt using large datasets in a related domain are leveraged for the task at hand.
Transfer learning  is a machine learning technique for building powerful classifiers when large datasets are unavailable. In transfer learning, knowledge gained by training accurate classifiers (pre-trained models) using large datasets in one domain is reused/leveraged to build powerful classifiers in a related domain where access to large datasets is unavailable. In the context of image classification using CNNs, the layers closer to the input layer learn generic features such as edges and blobs. Transfer learning effectively exploits this observation and enables building powerful CNN based image classifiers using moderately sized datasets. Specifically, the lower layers (close to the input) are retained along with their learned parameters; whilst the top layers are removed/customized for the problem at hand and initialized with random parameters. All the parameters of this customized network are learnt using the available dataset and this process is called finetuning. The VGG16  model pre-trained on the ImageNet dataset  is well studied for image classification via transfer learning and we employ it to build classifiers for wood identification.
CNN architecture for wood identification
Specimen preparation and imaging
Training and testing splits of the image dataset by class at the species level
Patch dataset creation
Summary of patch datasets for species/genus level models
Model prediction accuracies
Patch level (%)
Image level (%)
Global average pooling
Species level (10 class)
Genus level (from 10-class species level)
Genus level (6 class)
Global max pooling
Species level (10 class)
Genus level (from 10-class species level)
Genus level (6 class)
Results and discussion
Wood anatomy typically varies characteristically at the generic rather than the specific level even when analyzed with light microscopy —species-level distinctions are typically based on external morphological, reproductive and vegetative characteristics that are not reflected in the wood anatomy, at least as analyzed by human experts. Given this traditional limitation of wood identification, it is necessary to distinguish between species-level and genus-level accuracy and hence we trained and evaluated 10-class species-level and 6-class genus-level models.
The overall accuracy of the predictions of our models is shown in Table 3. In order to calculate the genus-level accuracy from the 10-class species-level model (shown on the second row of Table 3 (“Genus level (from 10-class species level)”), we consider predictions of the wrong species but the correct genus as correct predictions and report those metrics. The image-level confusion matrices for the species-level and genus-level models are shown in Figs. 4 and 5 respectively.
10-Class species-level model
Slightly less than \(6\%\) of the images of Cabralea were misclassified as Guarea, and within Guarea, approximately \(7\%\) of the images of Guarea grandifolia were misclassified as Guarea glabra, but no images of either genus were classified as any genus outside these two. As shown in the confusion cladogram of Fig. 1, these results are in keeping with expectations based on traditional wood identification, and represent sensible errors.
The predictions made by the model for Carapa images are perfect, but the class also draws misclassified images from four species of three genera, which is again consistent with the known high variability of Carapa, as a taxon, as shown in Fig. 1, where the horizontal bar indicating variability is second only to that for Cedrela.
Within Cedrela, the genus identified as the most variable in Fig. 1, all the misclassified images (more than \(20\%\)) of Cedrela fissilis are predicted as Cedrela odorata and all the misclassified images (also more than \(20\%\)) of Cedrela odorata images are predicted as Cedrela fissilis. For Cedrela the model correctly determines the genus, but these CITES-listed species cannot be as reliably separated from each other as other species in our dataset. The absence of non-CITES-listed Cedrela in our dataset precludes the possibility of testing the ability of our model to discriminate between CITES-listed and non-CITES-listed species in this genus.
The model showed comparatively poor performance in classifying images of both species of Khaya, both in terms of the relatively low proportion of images correctly classified, and in that all misclassified images were assigned to species in other genera. Nearly all those images were attributed to Carapa guianensis, which is the closest nested relationship shown in the confusion cladogram (in Fig. 1), the remaining were classified as Swietenia, the next most closely related group in the cladogram.
Within Swietenia, the model’s classification of S. mahagoni images was perfect, but slightly less than \(4\%\) of S. macrophylla images were classified as Carapa guianensis and nearly \(5\%\) were incorrectly classified as S. mahagoni. Interestingly, no images of Swietenia were classified as Khaya or Cedrela.
When these species-level model results are reconsidered at the genus level, all the predictive errors within Cedrela and Guarea disappear, and less than \(2\%\) of Swietenia and less than \(1\%\) of Cedrela images are misclassified outside their genera. Because all the misclassified images of Khaya were attributed to species in different genera, consolidating the species-level results at the genus level does not alter the model’s relative performance in this genus.
6-Class genus-level model
Field screening of wood for most law enforcement purposes need not be accurate at the species level. Hence we also created an explicit genus level model in order to determine if clubbing species of the same genus into a single generic class would increase genus-level performance.
Table 3 presents summary data showing the improved performance of the explicit 6-class genus-level model compared to the genus-level results from the 10-class species-level model. The 6-class genus-level model (Fig. 5) shows major improvement for Cabralea, Cedrela, and Guarea, all of which are classified perfectly, and for Khaya which has only \(1\%\) of its images misclassified (as Cedrela). Interestingly, Carapa, despite being monotypic in the 10-class species-level model (and thus functionally a genus-level class in that model), loses specificity in the 6-class genus-level model, with approximately \(4\%\) of its images classified as Khaya, and another half-percent each as Cedrela and Swietenia. Roughly \(2\%\) of the Swietenia images are classified as Carapa, and roughly the same amount are classified as Khaya. This is interesting because in the 10-class species-level model, the only misclassification of a Swietenia image outside the genus was as Carapa. These results suggest that future work may benefit from targeted clubbing of some classes, especially if the real-world utility of species-level identification during field screening is minimal or non-existent.
In addition to achieving a useful level of resolution for field identification of wood specimens in trade, clubbing the individual species within each genus into one class has several potentially favorable side-effects. If one has access to expert-level biological domain knowledge about class variability in the dataset, targeted decisions on label space granularities can result in classes that are more favorable for training supervised machine learning algorithms . Lack of access to sufficient reference images at the species level is likely to be endemic and a limiting factor for image-based wood identification, but classes clubbed to the genus level are more likely to contain sufficient images. In addition to the biological and machine learning considerations and constraints, access to law enforcement expertise could further inform class definition taxonomies to ensure that the ultimate field-level tool is most relevant in the locales it is deployed.
The global context of trade in illegally logged wood necessarily invokes the need for large-scale or scalable solutions. Enforcement of existing law and support for additional protection requires a scientific and forensic basis for evaluating claims about wood and wood products, whether that claim is a species, a genus, a region of origin, or age. One part of a global solution is laboratory-based forensic methods that support successful prosecutions, but it is first necessary for law enforcement to identify, detain, and sample problematic shipments at points of control using effective field screening tools.
We presented a deep convolution neural network, trained using transfer learning, capable of separating anatomically similar commercial and endangered woods of the Meliaceae family at both the genus and species level, with image-level accuracy greater than 90%. This accuracy is far in excess of the minimum necessary to establish probable cause or other appropriate legal predicate for seizing or halting the transport of a shipment of wood. Our models operate on macroscopic images of the transverse surface of wood blocks—such a surface can be prepared and an image taken in situ by trained field agents. Convolutional neural networks trained end-to-end, either using transfer learning or trained from scratch (given sufficient datasets), clearly have the potential to provide a scalable way to accommodate model building in the various controlled contexts. Although we used the well-studied VGG16 pre-trained network to build our models, we are currently exploring other model architectures (e.g. [16, 17]). These alternate architectures, and their variants, have fewer parameters than the VGG networks and maybe well-suited for a system that can be deployed using mobile phones . We are also exploring scaling the models to hundreds of woods with human expert-informed label space taxonomies, and are studying methods to visualize [34, 35] and interpret the representation learned by the deep neural networks and compare it against traditional human-designed identification keys.
We believe that deep convolutional neural networks along with expert-informed label space taxonomies for controlling context show promise in developing an effective field screening tool for wood identification. For computer vision solutions to contribute most robustly in this area, either the context must be tightly controlled so that the number of classes remains low (e.g. a regional port with a limited number of local taxa) or the models must scale-up beyond the proof-of-concept we present here, by discriminating \(10^2\)–\(10^3\) classes of wood successfully, and such models must be tested and vetted in field application. The cooperation of machine learning experts, law enforcement officers, and forensic wood anatomists shows great potential to develop informed label space granularities that ensure the most relevant field-deployable models for field screening wood identification. Models developed, tested, and vetted cooperatively in this way can provide reliable, scalable field-screening of wood in trade to protect threatened and endangered species (e.g. CITES-listed species) and combat illegal logging.
Note that in Brazil, the state of São Paulo’s Instituto Florestal developed an “online” identification system where field agents transmitted macroscopic photos to experts in the laboratory who provided near real-time identifications to inform detention decisions.
The Madison (MADw) and Samuel J. Record (SJRw) collections were used.
RS prepared the samples for imaging. AC imaged the specimens. ACW initiated the project and curated the collected dataset. PR developed the machine learning models. ACW and PR designed the experiments, analyzed the results and wrote the paper. All authors read and approved the final manuscript.
We would like to thank Sarah Friedrich of the University of Wisconsin, Department of Botany for her help with the figures in this paper. We wish to acknowledge the assistance of Joe Destree, Dave Dostal, Dick Jordan, and Dave Eustice in developing hardware and tools for specimen preparation, and the assistance of Bruna Ferreira, Andre Lima, Tiago Ribeiro, and Tristan Paulus in developing the method for and doing much early work in specimen preparation. We further wish to acknowledge a number of useful general discussions about computer vision and machine learning for wood identification with John Hermanson.
The authors declare that they have no competing interests.
Availability of data and materials
The datasets generated and/or analysed during the current study are not yet publicly available as they are part of a larger dataset that will be curated and made public en masse, but are available in the interim from the corresponding author on reasonable request.
Consent for publication
Ethics approval and consent to participate
This work was supported in part by a grant from the US Department of State via Interagency Agreement number 19318814Y0010.
Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.
Open AccessThis article is distributed under the terms of the Creative Commons Attribution 4.0 International License (http://creativecommons.org/licenses/by/4.0/), which permits unrestricted use, distribution, and reproduction in any medium, provided you give appropriate credit to the original author(s) and the source, provide a link to the Creative Commons license, and indicate if changes were made. The Creative Commons Public Domain Dedication waiver (http://creativecommons.org/publicdomain/zero/1.0/) applies to the data made available in this article, unless otherwise stated.
- Wiedenhoeft A, Baas P, editors. Wood science for promoting legal timber harvest. IAWA J. 2011;32(2):121–296.Google Scholar
- Dormontt EE, Boner M, Braun B, Breulmann G, Degen B, Espinoza E, Gardner S, Guillery P, Hermanson JC, Koch G, Lee SL, Kanashiro M, Rimbawanto A, Thomas D, Wiedenhoeft AC, Yin Y, Zahnen J, Lowe AJ. Forensic timber identification: it’s time to integrate disciplines to combat illegal logging. Biol Conserv. 2015;191:790–8.View ArticleGoogle Scholar
- Lowe AJ, Dormontt EE, Bowie MJ, Degen B, Gardner S, Thomas D, Clarke C, Rimbawanto A, Wiedenhoeft A, Yin Y, Sasaki N. Opportunities for improved transparency in the timber trade through scientific verification. BioSci. 2016;66(11):990–8.View ArticleGoogle Scholar
- United Nations Office on Drugs and Crime: Best practice guide for forensic timber identification. 2016.Google Scholar
- Chalk I. Identification of hardwoods: a lens key. Forest Products Research Bulletin No. 25, USA. 1952.Google Scholar
- Ilic J. The CSIRO macro key for hardwood identification. Highett, Victoria, Australia: CSIRO. 1990.Google Scholar
- Miller R, Wiedenhoeft A. CITES identification guide—tropical woods: guide to the identification of tropical woods controlled under the convention on international trade in endangered species of wild fauna and flora. An Initiative of Environment Canada. 2002.Google Scholar
- Coradin VTR, Camargos JAA, Marques LF, Silva-Junior ER. Madeiras Similares Ao Mogno (Swietenia Macrophylla King): Chave Ilustrada Para Identificação Anatõmica em Campo. Brasília: Serviço Florestal Brasileiro/LPF; 2009.Google Scholar
- Wiedenhoeft A. Identification of Central American woods. Madison: Forest Products Society; 2011.Google Scholar
- Yin Y, Jiang X, Yuan L. Identification manual of endangered and precious timber species common in trades. Biological division. Beijing: Science Press; 2016.Google Scholar
- Hermanson JC, Wiedenhoeft AC. A brief review of machine vision in the context of automated wood identification systems. IAWA J. 2011;32(2):233–50.View ArticleGoogle Scholar
- Cimpoi M, Maji S, Vedaldi A. Deep filter banks for texture recognition and segmentation. In: Proceedings of the IEEE conference on computer vision and pattern recognition; 2015. p. 3828–3836.Google Scholar
- Filho PLP, Oliveira LS, Nisgoski S, Britto AS. Forest species recognition using macroscopic images. Mach Vis Appl. 2014;25(4):1019–31.View ArticleGoogle Scholar
- Krizhevsky A, Sutskever I, Hinton GE. Imagenet classification with deep convolutional neural networks. In: Pereira F, Burges CJC, Bottou L, Weinberger KQ, editors. Advances in neural information processing systems; 2012. p. 1097–105.Google Scholar
- Simonyan K, Zisserman A. Very deep convolutional networks for large-scale image recognition. 2014. CoRR arXiv:1409.1556.
- Szegedy C, Vanhoucke V, Ioffe S, Shlens J, Wojna Z. Rethinking the inception architecture for computer vision. 2015. CoRR arXiv:1512.00567.
- He K, Zhang X, Ren S, Sun J. Deep residual learning for image recognition. 2015. CoRR arXiv:1512.03385.
- Pan SJ, Yang Q. A survey on transfer learning. IEEE Trans Knowl Data Eng. 2010;22(10):1345–59.View ArticleGoogle Scholar
- Russakovsky O, Deng J, Su H, Krause J, Satheesh S, Ma S, Huang Z, Karpathy A, Khosla A, Bernstein M, Berg AC, Fei-Fei L. Imagenet large scale visual recognition challenge. Int J Comput Vis. 2015;115(3):211–52.View ArticleGoogle Scholar
- Razavian AS, Azizpour H, Sullivan J, Carlsson S. CNN features off-the-shelf: an astounding baseline for recognition. 2014. CoRR arXiv:1403.6382.
- Helgason T, Russell SJ, Monro AK, Vogel JC. What is mahogany? The importance of a taxonomic framework for conservation. Botan J Linnaeus Soc. 1996;122:47–59.Google Scholar
- Pennington TD, Styles BT, Taylor DAH. Meliaceae. Flora Neotropica. 1981;28:1–470.Google Scholar
- LeCun Y, Boser B, Denker JS, Henderson D, Howard RE, Hubbard W, Jackel LD. Backpropagation applied to handwritten zip code recognition. Neural Comput. 1989;1(4):541–51.View ArticleGoogle Scholar
- Nair V, Hinton GE. Rectified linear units improve restricted boltzmann machines. In: Proceedings of the 27th international conference on machine learning. ICML’10; 2010. p. 807–14.Google Scholar
- Ioffe S. Szegedy C. Batch normalization: accelerating deep network training by reducing internal covariate shift. 2015. CoRR arXiv:1502.03167.
- Hinton GE, Srivastava N, Krizhevsky A, Sutskever I, Salakhutdinov R. Improving neural networks by preventing co-adaptation of feature detectors. 2012. CoRR arXiv:1207.0580.
- Buda M, Maki A, Mazurowski MA. A systematic study of the class imbalance problem in convolutional neural networks. 2017. CoRR arXiv:1710.05381.
- Kingma DP, Ba J. Adam: a method for stochastic optimization. 2014. CoRR arXiv:1412.6980.
- Chollet F, et al. Keras. GitHub. 2015. https://github.com/fchollet/keras
- Abadi M, Agarwal A, Barham P, Brevdo E, Chen Z, Citro C, Corrado GS, Davis A, Dean J, Devin M, Ghemawat S, Goodfellow IJ, Harp A, Irving G, Isard M, Jia Y, Józefowicz R, Kaiser L, Kudlur M, Levenberg J, Mané D, Monga R, Moore S, Murray DG, Olah C, Schuster M, Shlens J, Steiner B, Sutskever I, Talwar K, Tucker PA, Vanhoucke V, Vasudevan V, Viégas FB, Vinyals O, Warden P, Wattenberg M, Wicke M, Yu Y, Zheng X. Tensorflow: large-scale machine learning on heterogeneous distributed systems. 2016. CoRR arXiv:1603.04467.
- Gasson P. How precise can wood identification be? Wood anatomy’s role in support of the legal timber trade, especially CITES. IAWA J. 2011;32(2):137–54.View ArticleGoogle Scholar
- Esteva A, Kuprel B, Novoa RA, Ko J, Swetter SM, Blau HM, Thrun S. Dermatologist-level classification of skin cancer with deep neural networks. Nature. 2017;542(7639):115–8.View ArticlePubMedGoogle Scholar
- Tang XJ, Tay YH, Siam NA, Lim SC. Rapid and robust automated macroscopic wood identification system using smartphone with macro-lens. 2017. CoRR arXiv:1709.08154.
- Zhou B, Khosla A, Lapedriza A, Oliva A, Torralba A. Learning deep features for discriminative localization. CVPR. 2016.Google Scholar
- Selvaraju RR, Das A, Vedantam R, Cogswell M, Parikh D, Batra D. Grad-cam: why did you say that? Visual explanations from deep networks via gradient-based localization. 2016. CoRR arXiv:1610.02391.