Hostname: page-component-745bb68f8f-kw2vx Total loading time: 0 Render date: 2025-02-11T11:13:14.404Z Has data issue: false hasContentIssue false

The Cyborg Astrobiologist: matching of prior textures by image compression for geological mapping and novelty detection

Published online by Cambridge University Press:  19 February 2014

P.C. McGuire*
Affiliation:
Planetary Sciences and Remote Sensing Group, Institute of Geological Sciences, Freie Universität Berlin, Berlin, Germany Formerly at: Department of the Geophysical Sciences, University of Chicago, Chicago, IL, USA
A. Bonnici
Affiliation:
Department of Systems and Control Engineering, University of Malta, Malta
K.R. Bruner
Affiliation:
Department of Geology and Geography, West Virginia University, Morgantown, WV, USA
C. Gross
Affiliation:
Planetary Sciences and Remote Sensing Group, Institute of Geological Sciences, Freie Universität Berlin, Berlin, Germany
J. Ormö
Affiliation:
Centro de Astrobiología, CSIC-INTA, Torrejón de Ardoz, Madrid, Spain
R.A. Smosna
Affiliation:
Department of Geology and Geography, West Virginia University, Morgantown, WV, USA
S. Walter
Affiliation:
Planetary Sciences and Remote Sensing Group, Institute of Geological Sciences, Freie Universität Berlin, Berlin, Germany
L. Wendt
Affiliation:
Planetary Sciences and Remote Sensing Group, Institute of Geological Sciences, Freie Universität Berlin, Berlin, Germany
Rights & Permissions [Opens in a new window]

Abstract

We describe an image-comparison technique of Heidemann and Ritter (2008a, b), which uses image compression, and is capable of: (i) detecting novel textures in a series of images, as well as of: (ii) alerting the user to the similarity of a new image to a previously observed texture. This image-comparison technique has been implemented and tested using our Astrobiology Phone-cam system, which employs Bluetooth communication to send images to a local laptop server in the field for the image-compression analysis. We tested the system in a field site displaying a heterogeneous suite of sandstones, limestones, mudstones and coal beds. Some of the rocks are partly covered with lichen. The image-matching procedure of this system performed very well with data obtained through our field test, grouping all images of yellow lichens together and grouping all images of a coal bed together, and giving 91% accuracy for similarity detection. Such similarity detection could be employed to make maps of different geological units. The novelty-detection performance of our system was also rather good (64% accuracy). Such novelty detection may become valuable in searching for new geological units, which could be of astrobiological interest. The current system is not directly intended for mapping and novelty detection of a second field site based on image-compression analysis of an image database from a first field site, although our current system could be further developed towards this end. Furthermore, the image-comparison technique is an unsupervised technique that is not capable of directly classifying an image as containing a particular geological feature; labelling of such geological features is done post facto by human geologists associated with this study, for the purpose of analysing the system's performance. By providing more advanced capabilities for similarity detection and novelty detection, this image-compression technique could be useful in giving more scientific autonomy to robotic planetary rovers, and in assisting human astronauts in their geological exploration and assessment.

Type
Research Article
Copyright
Copyright © Cambridge University Press 2014 

Introduction

In prior work, we have developed computer algorithms for real-time novelty detection and rarity mapping for astrobiological and geological exploration (McGuire et al. Reference McGuire, Ormö, Diaz, Rodriguez, Gomez, Ritter, Oesker and Ontrup2004, Reference McGuire, Diaz, Ormö, Gomez, Rodriguez, Sebastian, Ritter, Haschke, Oesker and Ontrup2005a, Reference McGuire, Gomez, Rodriguez, Sebastian, Ormö, Diaz, Ritter, Oesker, Haschke and Ontrupb, Reference McGuire, Gross, Wendt, Bonnici, Souza-Egipsy, Ormö, Diaz-Martinez, Foing, Bose, Walter, Oesker, Ontrup, Haschke and Ritter2010; Bartolo et al. Reference Bartolo, McGuire, Camilleri, Spiteri, Borg, Farrugia, Ormö, Gomez-Elvira, Rodriguez-Manfredi, Diaz-Martinez, Ritter, Haschke, Oesker and Ontrup2007; Gross et al. Reference Gross, Wendt, McGuire, Bonnici, Foing, Souza-Egipsy, Bose, Walter, Ormö and Diaz-Martinez2009, Reference Gross, Wendt, McGuire, Bonnici, Foing, Souza-Egipsy, Bose, Walter, Ormö, Díaz-Martínez, Oesker, Ontrup, Haschke and Ritter2010; Wendt et al. Reference Wendt, Gross, McGuire, Bonnici, Foing, Souza-Egipsy, Bose, Walter, Ormö, Díaz-Martínez, Oesker, Ontrup, Haschke and Ritter2009). These algorithms were tested at astrobiological field sites using mobile computing platforms originally (McGuire et al. Reference McGuire, Ormö, Diaz, Rodriguez, Gomez, Ritter, Oesker and Ontrup2004, Reference McGuire, Diaz, Ormö, Gomez, Rodriguez, Sebastian, Ritter, Haschke, Oesker and Ontrup2005a, Reference McGuire, Gomez, Rodriguez, Sebastian, Ormö, Diaz, Ritter, Oesker, Haschke and Ontrupb, Reference McGuire, Gross, Wendt, Bonnici, Souza-Egipsy, Ormö, Diaz-Martinez, Foing, Bose, Walter, Oesker, Ontrup, Haschke and Ritter2010) with a wearable computer connected to a digital video camera, but more recently (Bartolo et al. Reference Bartolo, McGuire, Camilleri, Spiteri, Borg, Farrugia, Ormö, Gomez-Elvira, Rodriguez-Manfredi, Diaz-Martinez, Ritter, Haschke, Oesker and Ontrup2007; Gross et al. Reference Gross, Wendt, McGuire, Bonnici, Foing, Souza-Egipsy, Bose, Walter, Ormö and Diaz-Martinez2009, Reference Gross, Wendt, McGuire, Bonnici, Foing, Souza-Egipsy, Bose, Walter, Ormö, Díaz-Martínez, Oesker, Ontrup, Haschke and Ritter2010; Wendt et al. Reference Wendt, Gross, McGuire, Bonnici, Foing, Souza-Egipsy, Bose, Walter, Ormö, Díaz-Martínez, Oesker, Ontrup, Haschke and Ritter2009; McGuire et al. Reference McGuire, Gross, Wendt, Bonnici, Souza-Egipsy, Ormö, Diaz-Martinez, Foing, Bose, Walter, Oesker, Ontrup, Haschke and Ritter2010) with a phone camera connected wirelessly to a local or remote server computer. The image features used in the novelty detection and rarity mapping in prior work were based only upon RGB or HSI colour.

Nonetheless, even with image features limited to colour, the mobile exploration systems worked very robustly. Based on an analysis of the procedures governing the team's geologists when analysing an encountered outcrop, we developed a concept of ‘novelty detection’ for guiding the Cyborg Astrobiologist system in a first step towards mimicking the geologist's approach. Very simplified, the geologist's decisions when approaching an outcrop are based upon observations that may be useful to ‘tell the tale’ about the outcrop, i.e. relations between individual parts of the outcrop. This can, for instance, be contacts between different lithologies, alteration sequences, diagenetic variations or structures such as beddings or fractures. Commonly, such variations are, at a first distant glance of an outcrop, expressed as colour and/or texture variations. In, for instance, the case of colour variations, the geologist may decide that it is a certain colour that stands out from a more common background that should be given special attention, i.e. as ‘interest points’. The next step for the geologists is, thus, usually to investigate the contacts between these first-order ‘interest points’ and the background. When further advancing on the outcrop, new ‘interest points’ of more detailed scale are selected based upon the same principle in a somewhat iterative way. The ‘colour-only’ Cyborg Astrobiologist has, in our first tests, been employed in unvegetated, desert-like environments where it was able to identify novel or uncommon areas in image sequences, ranging from mostly white-coloured gypsum to mostly red-coloured ‘red bed’ sandstones. The system was able to identify, for example, bleached areas and hematite concretions in the red beds of Spain, and lichens of varying colours within the desert landscapes of Spain and Utah as being novel features (when first observed) of those landscapes (Gross et al. Reference Gross, Wendt, McGuire, Bonnici, Foing, Souza-Egipsy, Bose, Walter, Ormö and Diaz-Martinez2009, Reference Gross, Wendt, McGuire, Bonnici, Foing, Souza-Egipsy, Bose, Walter, Ormö, Díaz-Martínez, Oesker, Ontrup, Haschke and Ritter2010; Wendt et al. Reference Wendt, Gross, McGuire, Bonnici, Foing, Souza-Egipsy, Bose, Walter, Ormö, Díaz-Martínez, Oesker, Ontrup, Haschke and Ritter2009; McGuire et al. Reference McGuire, Gross, Wendt, Bonnici, Souza-Egipsy, Ormö, Diaz-Martinez, Foing, Bose, Walter, Oesker, Ontrup, Haschke and Ritter2010). However, another important factor in addition to colour variations in the decision-making of the geologist when approaching an outcrop is texture, the objective of this study.

Herein, using data acquired by the Astrobiology Phone-cam at a geological field site (Fig. 1), we test an image-comparison technique of Heidemann and Ritter (Reference Heidemann and Ritter2008a, Reference Heidemann and Ritterb), which uses image compression and is capable of (i) detecting novel (coloured) textures in a series of images as well as of (ii) alerting the user to the similarity of a new image to a previously observed texture. We first implemented this technique in 2010 (Bonnici et al. Reference Bonnici, Gross, McGuire, Ormö, Walter and Wendt2010), but did not test this software at a geological field site until now. Such a capability could be useful in giving more scientific autonomy to robotic planetary rovers (Volpe Reference Volpe2003; Fink et al. Reference Fink, Dohm, Tarbell, Hare and Baker2005, Castano et al. Reference Castano, Estlin, Gaines, Chouinard, Bomstein, Anderson, Burl, Thompson, Castano and Judd2007; Halatci et al. Reference Halatci, Brooks and Iagnemma2007, Reference Halatci, Brooks and Iagnemma2008), and perhaps in assisting human astronauts in their geological explorations. For example, suppose that a semi-autonomous planetary rover equipped with texture-based novelty detection is observing a long series of textures corresponding to hematite concretions embedded in mineral deposits. With texture-based novelty detection (Gulick et al. Reference Gulick, Morris, Ruzon and Roush2001, Reference Gulick, Hart, Shi and Siegel2004; Thompson et al. Reference Thompson, Allwood, Bekker, Cabrol, Estlin, Fuchs and Wagstaff2012), this rover would report, during stage 1 of its approach towards the outcrop, that a particular previously unobserved horizontally layered texture is novel, and hence merits further investigation. In stage 2, it would detect smaller spherical objects, possibly even of a different colour relative to the background, within the previously observed layers. Depending on the resolution of the camera, this process could continue iteratively, with further observations of onion-layering of the spherules and so forth. A certain amount of autonomous decision-making of the rover would, in this way, greatly reduce the amount of data to be transferred between the rover and the scientists.

Fig. 1. The Cyborg Astrobiologist using its image-comparison software with the Astrobiology Phone-cam to study lichens (left picture) and coal with cleats (right picture) at the geological field site (photos by K.R. Bruner).

Heidemann and Ritter's image-compression technique

Following Heidemann and Ritter (2008a, b), we ‘calculate the similarity of two images I 1, I 2 as:

(1)$$ D_{{\rm SIM}} (I_1, I_2 ) = S(I_1 ) + S(I_2 ) - S(I_{12} ),$$

where S(·) denotes the [(single-number) byte] size of a compressed image. I 12 is the ‘joint’ image obtained as juxtaposition of pixel arrays I 1 and I 2’. If the two images are very similar, then S(I 12) will be small and D SIM(I 1, I 2) will be large. If the two images are very different (due to textural and colour differences), then S(I 12) will be large and D SIM(I 1, I 2) will be small.

The juxtaposition in our implementation is left–right, but we have not compared our implementation to an up–down juxtaposition. Various image-compression programmes were investigated by Heidemann & Ritter (Reference Heidemann and Ritter2008a, Reference Heidemann and Ritterb), and their optimal image compressor for texture classification was gzip, which we have chosen as our image compressorFootnote 1. Gzip relies on Huffman entropy coding (Huffman Reference Huffman1952; see also Saravanan & Ponalagusamy Reference Saravanan and Ponalagusamy2010) and the Lempel–Ziv algorithm (Lempel & Ziv Reference Lempel and Ziv1977; see also Kärkkäinen et al. Reference Kärkkäinen, Kempa, Puglisi, Bonifaci, Demetrescu and Marchetti-Spaccamela2013). Huffman coding is based on how frequent the symbolsFootnote 2 are in a data stream, assigning shorter bit-length representations for more commonly used symbols. The Lempel–Ziv algorithm (LZ77) eliminates duplicate strings of symbols using ‘sliding-window compression’ (referring to the buffer window that records the previously observed symbols in the data stream)Footnote 3. Quoting Heidemann & Ritter (Reference Heidemann and Ritter2008b): ‘The Lempel–Ziv algorithm (LZ77) (Lempel & Ziv Reference Lempel and Ziv1977), which detects repeatedly occurring symbol sequences within the data, such that a dictionary can be established. A repeated symbol sequence can then be replaced by the symbol defined in the dictionary.’Footnote 4 The gzip algorithm outperforms (Heidemann & Ritter Reference Heidemann and Ritter2008b) the correlation technique and the histogram-based matching technique for the texture-classification, object-recognition and image-retrieval tasks.

By employing a simple image-compression technique such as Heidemann and Ritter's, we can avoid the complexities and ambiguities of more reductionist textural comparison techniques (i.e. Haralick et al. Reference Haralick, Shanmugam and Dinstein1973; Rao Reference Rao2012), wherein a significant number of different textural indicators (at different spatial scales and orientations) are computed and mapped for each image.

In order to simplify interpretation, we have taken the logarithm of D SIM and added a normalization factor, so that the maximal range of D SIM is 0–100% for the current database of images. We accomplish this by comparing the newest image to itself and setting the value obtained as the 100% similarity value. One would expect this value of D sim(I N,I N) to be slightly smaller than S(I N) since the two images being compared are the same, but the changes at the boundary (between the two identical juxtaposed images) introduce a discrepancy in the value of D sim(I N,I N).

Field tests

The Astrobiology Phone-cam system sends images wirelessly by Bluetooth to a nearby laptop computer, dynamically building an image library with examples from different terrain. The Phone-cam for this field test is a Samsung SGH-A767 with 1280×960 RGB colour images, and the laptop is a Dell Inspiron 9300. Each incoming image I N is processed by a compiled MATLAB code by the laptop using equation (1) to compute the similarity D SIM(I N, I J), with each previous image, I J. If D SIM(I N, I J) is less than a chosen threshold for all previous images I J (for all J<N), then the computer considers image I N as ‘novel’ and returns a text message to the Phone-cam by Bluetooth informing the explorer that the image is novel. If the image is novel, the explorer might decide to perform a more detailed analysis of the ground or rocky outcrop that image I N represents. If D SIM(I N, I J) is greater than the chosen threshold for one or more of the previous images I J, then the image I K, which has the highest similarity score (highest value of D SIM(I N, I K)), is returned to the phone camera via Bluetooth, juxtaposed with I N, in order for the user to assess the similarity visually. We have performed tests of this procedure for detecting novelty or similarity; a set of example images in the test sequence and their best-matching prior images are juxtaposed in Fig. 2. Image pairs such as in Fig. 2 are the visual information that the image-compression algorithm on the laptop computer outputs and then sends to the phone camera for inspection.

Fig. 2. Good matches: Incoming image I N is on the left, and best-matching image is on the right. Each image is about 0.5 m across. (a) Platy-rock texture (laminated shale); (b) yellow-sporing bodies of lichens; (c) coal bed.

Owing to its proximity to three of the authors (McGuire, Smosna and Bruner) at the time of the survey, we chose a geological field site near the northern end of the Morgantown Mall in Morgantown, West Virginia, USA. This mall is the former site of a coal mine, so there are several exposed geological cuts, including one of a coal bed. These rocksFootnote 5 comprise a part of West Virginia's coal measures, in particular the lowest section of the Pittsburgh Formation, and have a late Carboniferous age. The outcrop exposes a heterogeneous mix of coal, sandstone, shale, mudstone and limestone, typically occurring in beds of 30–150 cm thickness. The photographed coal is the Pittsburgh coal, about 3 m thick with well-developed cleats (fractures) and elemental sulphur on the face. The sandstone ranges from being thin-bedded to being massive with planar bedding and cross-bedding. The shale is characteristically thinly laminated. The mudstone shows a nodular character and locally contains ironstone concretions (composed of iron carbonate and iron silicate minerals). The limestone exhibits spherical weathering and solution features.

Our original plan was to use the Cyborg Astrobiologist's image-compression system at the geological field site, acquiring and analysing about 25 images in the 1.5 h battery lifetime of the laptop server computer (it takes an average of 3–4 min to analyse each image). However, at the field site, we decided to acquire 55 imagesFootnote 6 with the Astrobiology Phone-cam in 1 h, and then to analyse the images offline after the field work was completed. This offline analysisFootnote 7 lasted for about 4 h, longer than the battery life of the laptop. Therefore, this was not the ideal real-time test, but all the capabilities of the system were otherwise tested and more images could be analysed than during a real-time test.

Three of the 55 result images are shown in Fig. 2. Figure 2(a) shows the platy-rock texture on the left, and the best-matching of the prior images on the right. This is indeed a successful match. Likewise, it was a successful match for the yellow-sporing bodies of the lichens in Fig. 2(b), and for the coal bed in Fig. 2(c). Generally, for each image which was not novel, the computer successfully matched it best with another image of similar texture and/or colour statistics.

For those images which were novel, the computer either gave a low D SIM score (<39%) or matched it with a higher score with an image that even the human analyst would say is similar. See Fig. 3 for examples of the novel images and their best matches in the database. Figure 3(a) shows the low D SIM score for the novel platy texture with the best-match smooth iron-stained texture. Figure 3(b) shows the low D SIM score for the novel yellow-sporing bodies of the lichens with the best-match platy texture with red clast. And Fig. 3(c) shows the higher D SIM score for the novel coal bed with the best-match yellow-sporing bodies of the lichens. In Fig. 3(c), the score is likely to be higher because the substrate rock for the yellow lichen is black, much like a large part of the coal bed, and because the coal bed has some pale yellow colouring due to sulphur content.

Fig. 3. Novel images: Incoming image I N is on the left, and best-matching image is on the right. Each image is about 0.5 m across. (a) Novel platy-rock texture; (b) novel yellow-sporing bodies of lichens; (c) novel coal bed.

In Table 1, we estimate the number of images that are novel, similar or different, as judged by three members of our human teamFootnote 8, and the number of images that are novel but have high similarity scores, the number of images that are similar and the number of images that are similar with low similarity scores. A more detailed compilation of our results is shown in Table 2, listing for each of the 55 incoming images: (i) which prior image matched best, as well as (ii) notes about novelty and mismatching. The range of observed scores for matching images is generally about 39–50%. Lower scores generally signify novelty, although there are some novel images with scores above 39%, due to similarity to other geological units (see Table 1). The threshold of 39% between similarity and novelty was chosen after the experiment, and it is not a perfect threshold, but it is approximately the right value for these images.

Table 1. Summary of results.

Accuracy of Novelty detection: 9/14=64%.

Accuracy of Similarity detection: 31/34=91%.

Accuracy of Difference detection: 2/6=33%.

Overall accuracy: 42/54=78%.

Table 2. Detailed results.

Note: a new image database was created with the first image, which was image #264.

There are examples where the ∼48% score appears rather low (for instance, the pair 273/272, where the two images appear to the human observer identical, apart from a small vertical shift). We would expect that the algorithm should be able to detect this better: if the strategy is to compare horizontal scan line (pieces), then the vertical shift should leave many scan line similarities intact and a high value should result (but there might still be high sensitivity to rotation). Based upon prior tests, we believe that the reason the highest scores are not near 100% is because of the discontinuity between the juxtaposed images.

Discussion

The intent of our algorithm is to analyse a series of images made in a homogeneous environment (a common outcrop), which slightly changes over time. The scope here is to automatically detect sudden variations in the outcrop's visual appearance, for the purpose of minimizing interactions by the human scientists who are supervising the robotic rover from afar. We do not expect our current algorithm to be able to extrapolate to a new field site in order to identify lichens or coal or other coloured textures that were observed at a prior field site, for the following reasons:

First, these coloured textures would most likely have significant intrinsic variability in their colour or textural qualities from one site to the next. However, if the colour and textural qualities were the same or similar, then the similarity matching and novelty detection would work well.

Second, we have not controlled for lighting conditions or photometric angles of observation. We do not attempt such control in the present work, since our main objective is to test the similarity-matching and novelty-detection capabilities of the image comparison algorithm. Nonetheless, in a much more advanced system, real-time photometric and atmospheric correction could be added, similar to the post-processing correction system for the CRISM camera on the Mars Reconnaissance Orbiter (McGuire et al. Reference McGuire, Wolff, Smith, Arvidson, Murchie, Clancy, Roush, Cull, Lichtenberg, Wiseman, Green, Martin, Milliken, Cavender, Humm, Seelos, Seelos, Taylor, Ehlmann, Mustard, Pelkey, Titus, Hash and Malaret2008). We did not need such photometric or atmospheric control for the field tests completed here, since the lighting conditions were the same (overcast) throughout the one-hour field test.

Third, at the new field site, the colour and textural properties of the images may have different spatial scales than at the original field site. Careful control of the camera distance to the rock outcrop may lessen the differences in spatial scale of the coloured textures in the images from one field site to another. In the current field tests, we have only tested the system at a single image scale or resolution, given by the human operator locating the camera about 1 m away from the outcrop. This single-scale property of the images could be approximated on a Mars rover, for example by using an imaging camera held by the robotic arm at a fixed distance from the outcrop. We do not have plans currently to develop an algorithm that is capable of distinguishing between coloured textures at different spatial scales.

Fourth, this algorithm is an unsupervised algorithm; it does not need to know prior image classifications in order to identify novel or familiar coloured textures. The algorithm does not know that an image is one of lichen or a coal bed – it just can classify an image as familiar or novel, based upon similarity to previous images. So, the algorithm cannot really identify lichens or coal as being lichens or coal. The algorithm does not need to know what it is looking at ahead of time. The classifications were made by human geologists a couple of days after the field test. The intent of the algorithm is to allow the robotic astrobiologist to find those areas of a rocky outcrop that are similar to each other (for mapping purposes) and those areas of a rocky outcrop that are novel (for further investigation and/or sample acquisition).

These conditions exactly correspond to the requirements for an extraterrestrial robotic image-acquisition system, where many images are autonomously taken in a mostly homogeneous environment. Sudden changes or exceptions in the environment could be flagged and reported to the human scientists on Earth, effectively scanning the immense number of images for anomalies (possibly of astrobiological interest).

Conclusions and future

Our initial tests at a geological field site of this texture-based algorithm for image comparison show promise for both novelty detection and for similarity matching. The similarity matching was superb (91%), especially of images of the yellow lichens and of images of the coal bed. Our next step is for the Cyborg Astrobiologist to perform more extensive field testing of this algorithm with the Astrobiology Phone-cam at other geological/astrobiological field sites.

Testing with grey-scale cameras and multispectral cameras are possible extensions of this work. To simulate images taken by a rover, one could take videos and randomly select frames from the video. We also hope to speed up the algorithm so that it can analyse ∼100 images in a laptop-battery-limited real-time test of 1.5 h. This speed-up can be accomplished perhaps by using the cluster characteristics of the image database so that each image is only compared with one image of each cluster of the database. The speed-up could also be accomplished by eliminating the delays of Bluetooth transmission and developing an app for smartphones which acts directly on the smartphone as soon as the image is captured; this would at least reduce the bottleneck to one battery life rather than two, and carrying a spare phone battery is much easier than carrying around a spare laptop battery. Another extension of the system could be useful for offline analysis: allowing for an image to be segmented for texture with this compression algorithm. This could lead the Cyborg Astrobiologist to understand further the images, so that the software could be used by a Robotic Astrobiologist to zero in on novel astrobiological/geological areas of outcrops.

Acknowledgements

We thank Helge Ritter for suggesting the image-compression technique to us, and Tim Warner for help with identifying a field site in West Virginia. PCM acknowledges past support by a research fellowship from the Alexander von Humboldt Foundation, support from the Freie Universität Berlin, and computer resources for the preparation of this manuscript at the library of West Virginia University. CG acknowledges the Helmholtz Association through the research alliance ‘Planetary Evolution and Life’. The work by JO is supported by grants AYA2008-03467⁄ESP and AYA2011-24780/ESP from the Spanish Ministry of Economy and Competitiveness.

Footnotes

1 In our implementation, we have used the gzip functionality of MATLAB 7.10.0 (R2010a), using the MATLAB function java.util.zip.GZIPOutputStream(), which calls a Java function that was derived from the Java Deflater compression algorithm. Our version of Java is 7.0.110.21. The Java Deflater algorithm is based upon the Compressed Data Format specification for DEFLATE in the Request for Comments #1951 (http://www.ietf.org/rfc/rfc1951.txt) of the Internet Engineering Task Force and the Internet Society.

2 A symbol in this context is a numeric byte.

3 We use gzip's standard sliding window size of 32kB. In future work, by modifying the gzip source code, we plan to study how texture-matching performance depends on this sliding window size.

4 Bytes of similar numeric value will not be automatically classified as a single symbol. Therefore, digital noise or slight, intrinsic variations of pixel values may affect the compression results. This is one shortcoming of this dictionary-based technique for data compression. In the future, as suggested by Heidemann and Ritter (Reference Heidemann and Ritter2008a,Reference Heidemann and Ritterb), other data-compression techniques could be used (or developed) that can handle better the inherent variability of RGB pixel values. Such techniques would be in principle lossy, though lossy compression in the JPEG implementation does not work well, due to the discrete cosine transform that was utilized by JPEG (Heidemann and Ritter, Reference Heidemann and Ritter2008a).

5 unpublished data from the West Virginia Geological & Economic Survey.

6 The images were acquired without horizon, or rock-debris aprons, or vegetation in the view. This is not the way a robot or rover would have taken the images, unless it was trained to do so. However, it allows us to focus our investigation on the rocky outcrops. We also tried to focus as much as possible on the ‘end-member homogeneous textures’, as opposed to images that contain multiple, different, mixed textures. In a limited investigation which we do not present here, our software seems to work also with mixed colored textures, wherein the best matching images are of similar mixing amounts of the different colored textures. Much more work needs to be done for these mixed textures.

7 The testing procedure consisted of (i) an empty image database at the beginning of the test, and subsequently, (ii) each image was compared to the previous ones as it is added to the database, and (iii) to test the performance of the algorithm, the images were later classified based on expert knowledge as one of the following classes: weathered rock, laminar shale… (and so on). The match was considered successful if the algorithm returned an image from the same rock unit.

8 Novelty is based upon the history of images, and difference is only for the non-novel and non-similar images and is based upon only the labels of the best matching pair.

References

Bartolo, A., McGuire, P.C., Camilleri, K.P., Spiteri, C., Borg, J.C., Farrugia, P.J., Ormö, J., Gomez-Elvira, J., Rodriguez-Manfredi, J.A., Diaz-Martinez, E., Ritter, H., Haschke, R., Oesker, M., Ontrup, J. (2007). The Cyborg Astrobiologist: porting from a wearable computer to the Astrobiology Phone-cam. Int. J. Astrobiol. 6, 255261.Google Scholar
Bonnici, A., Gross, C., McGuire, P.C., Ormö, J., Walter, S. & Wendt, L. (2010). The Cyborg Astrobiologist: compressing images for the matching of prior textures and for the detection of novel textures. In European Planetary Science Congress (EPSC), vol. 5. Potsdam, Germany, extended abstract #162.Google Scholar
Castano, R., Estlin, T., Gaines, D., Chouinard, C., Bomstein, B., Anderson, R.C., Burl, M., Thompson, D., Castano, A. & Judd, M. (2007). Onboard autonomous rover science. In IEEE Aerospace Conf. Big Sky, Montana.Google Scholar
Fink, W., Dohm, J.M., Tarbell, M.A., Hare, T.M. & Baker, V.R. (2005). Next-generation robotic planetary reconnaissance missions: a paradigm shift. Planet. Space Sci. 53, 14191426.CrossRefGoogle Scholar
Gross, C., Wendt, L., McGuire, P.C., Bonnici, A., Foing, B.H., Souza-Egipsy, V., Bose, R., Walter, S., Ormö, J., Diaz-Martinez, E. (2009). Testing the Cyborg Astrobiologist at the Mars Desert Research Station (MDRS), Utah. In European Planetary Science Congress (EPSC), vol. 4. Potsdam, Germany, extended abstract #548.Google Scholar
Gross, C., Wendt, L., McGuire, P.C., Bonnici, A., Foing, B.H., Souza-Egipsy, V., Bose, R., Walter, S., Ormö, J., Díaz-Martínez, E., Oesker, M., Ontrup, J., Haschke, R., Ritter, H. (2010). The Cyborg Astrobiologist: testing a novelty detection algorithm at the Mars Desert Research Station (MDRS), Utah. In LPSCXLI, Lunar and Planetary Science Conf. Houston, Texas, extended abstract #2457.Google Scholar
Gulick, V.C., Morris, R.L., Ruzon, M.A. & Roush, T.L. (2001). Autonomous image analyses during the 1999 Marsokhod rover field test. J. Geophys. Res. 106, 77457764.Google Scholar
Gulick, V.C., Hart, S.D., Shi, X. & Siegel, V.L. (2004). Developing an automated science analysis system for Mars surface exploration for MSL and beyond. In Lunar and Planetary Institute Conf. XXXV, extended abstract #2121.Google Scholar
Halatci, I, Brooks, C.A. & Iagnemma, K. (2007). Terrain classification and classifier fusion for planetary exploration rovers. In IEEE Aerospace Conf., pp. 111. Big Sky, MT.Google Scholar
Halatci, I., Brooks, C.A. & Iagnemma, K. (2008). A study of visual and tactile terrain classification and classifier fusion for planetary exploration rovers. Robotica 26, 767779.Google Scholar
Haralick, R.M., Shanmugam, K. & Dinstein, I.H. (1973). Textural features for image classification. IEEE Trans. Syst. Man Cybern. 6, 610621.Google Scholar
Heidemann, G. & Ritter, H. (2008a). Compression for visual pattern recognition. In Proc. Third Int. Symp. Communications, Control and Signal Processing (ISCCSP 2008), pp. 15201523. IEEE, St. Julians, Malta.Google Scholar
Heidemann, G. & Ritter, H. (2008b). On the contribution of compression to visual pattern recognition. In Proc. Third Int'l Conf. on Comp. Vision Theory and Applications, Funchal, Madeira-Portugal, vol. 2, pp. 8389.Google Scholar
Huffman, D.A. (1952). A method for the construction of minimum-redundancy codes. In Proc. Institute of Radio Engineers (IRE), pp. 10981102.Google Scholar
Kärkkäinen, J., Kempa, D. and Puglisi, S.J. (2013). Lightweight Lempel–Ziv parsing. In Proceedings of the 12th International Symposium, SEA 2013, Rome, Italy, June 5–7, 2013, Experimental Algorithms, ed. Bonifaci, V., Demetrescu, C., Marchetti-Spaccamela, A., pp. 139150.Springer, Berlin, Heidelberg; http://arxiv.org/abs/1302.1064Google Scholar
Lempel, A. & Ziv, J. (1977). A universal algorithm for sequential data compression. IEEE Trans. Inf. Theory 23, 337343.Google Scholar
McGuire, P.C., Ormö, J.O., Diaz, Martinez E., Rodriguez, Manfredi J.A., Gomez, Elvira J., Ritter, H., Oesker, M. & Ontrup, J. (2004). The Cyborg Astrobiologist: first field experience. Int. J. Astrobiol. 3, 189207.CrossRefGoogle Scholar
McGuire, P.C., Diaz, Martinez E., Ormö, J.O., Gomez, Elvira J., Rodriguez, Manfredi J.A., Sebastian, Martinez E., Ritter, H., Haschke, R., Oesker, M. & Ontrup, J. (2005a). The Cyborg Astrobiologist: scouting red beds for uncommon features with geological significance. Int. J. Astrobiol. 4, 101113.Google Scholar
McGuire, P.C., Gomez, Elvira J., Rodriguez, Manfredi J.A., Sebastian, Martinez E., Ormö, J., Diaz, Martinez E., Ritter, H., Oesker, M., Haschke, R. & Ontrup, J. (2005b). Field geology with a wearable computer: first results of the Cyborg Astrobiologist system. In Proc. ICINCO'2005 (Int. Conf. Informatics in Control, Automation and Robotics), vol. 3, pp. 283291. 14–17 September, 2005, Barcelona, Spain.Google Scholar
McGuire, P.C., Wolff, M.J., Smith, M.D., Arvidson, R.E., Murchie, S.L., Clancy, R.T., Roush, T.L., Cull, S.C., Lichtenberg, K.A., Wiseman, S.M., Green, R.O., Martin, T.Z., Milliken, R.E., Cavender, P.J., Humm, D.C., Seelos, F.P., Seelos, K.D., Taylor, H.W., Ehlmann, B.L., Mustard, J.F., Pelkey, S.M., Titus, T.N., Hash, C.D., Malaret, E.R. (2008). MRO/CRISM retrieval of surface lambert albedos for multispectral mapping of Mars with DISORT-based radiative transfer modeling: Phase 1 – using historical climatology for temperatures, aerosol optical depths, and atmospheric pressures. Trans. Geosci. Remote Sens., 46, 40204040.Google Scholar
McGuire, P.C., Gross, C., Wendt, L., Bonnici, A., Souza-Egipsy, V., Ormö, J., Diaz-Martinez, E., Foing, B.H., Bose, R., Walter, S., Oesker, M., Ontrup, J., Haschke, R., Ritter, H. (2010). The Cyborg Astrobiologist: testing a novelty-detection algorithm on two mobile exploration systems at Rivas Vaciamadrid in Spain and at the Mars Desert Research Station in Utah. Int. J. Astrobiol. 9, 1127.Google Scholar
Rao, A.R. (2012). A Taxonomy for Texture Description and Identification. Springer Publishing Company, Incorporated.Google Scholar
Saravanan, C. & Ponalagusamy, R. (2010). Lossless grey-scale image compression using source symbols reduction and Huffman coding. Int. J. Image Process. (IJIP) 3, 246251.Google Scholar
Thompson, D.R., Allwood, A., Bekker, D., Cabrol, N.A., Estlin, T., Fuchs, T. & Wagstaff, K.L. (2012). TextureCam: autonomous image analysis for astrobiology survey. In LPSCXLIII Lunar and Planetary Institute Conf., Houston, Texas, extended abstract #1659.Google Scholar
Volpe, R. (2003). Rover functional autonomy development for the Mars Mobile Science Laboratory. In IEEE Aerospace Conf., vol. 2, pp. 2_643–2_652. Big Sky, Montana.Google Scholar
Wendt, L., Gross, C., McGuire, P.C., Bonnici, A., Foing, B.H., Souza-Egipsy, V., Bose, R., Walter, S., Ormö, J., Díaz-Martínez, E., Oesker, M., Ontrup, J., Haschke, R., Ritter, H. (2009). The Cyborg Astrobiologist: teaching computers to find uncommon or novel areas of geological scenery in real-time. In European Space Agency Int. Conf. Comparative Planetology: Venus – Earth – Mars, ESTEC, Noordwijk, The Netherlands.Google Scholar
Figure 0

Fig. 1. The Cyborg Astrobiologist using its image-comparison software with the Astrobiology Phone-cam to study lichens (left picture) and coal with cleats (right picture) at the geological field site (photos by K.R. Bruner).

Figure 1

Fig. 2. Good matches: Incoming image IN is on the left, and best-matching image is on the right. Each image is about 0.5 m across. (a) Platy-rock texture (laminated shale); (b) yellow-sporing bodies of lichens; (c) coal bed.

Figure 2

Fig. 3. Novel images: Incoming image IN is on the left, and best-matching image is on the right. Each image is about 0.5 m across. (a) Novel platy-rock texture; (b) novel yellow-sporing bodies of lichens; (c) novel coal bed.

Figure 3

Table 1. Summary of results.

Figure 4

Table 2. Detailed results.