Having described the basic mode of operation of the Itrax core-scanner countless times in the past few months, I thought it time for a proper poster with nice diagrams, so I recently made this poster for the Itrax facility I’m looking after. It has a description of how x-ray beams are created and controlled in the scanner, the principles of measurement, and an explanation of how deriving chemical compositional data from fluorescence spectra can be done. It’s available from the resources page of this website. Let me know if you print one and use it – I love seeing photos of resources I’ve created being used!
This year the Technical Excellence at Manchester (TEaM) group at The University of Manchester did a short profile on me, which is online here.
I took a trip with our second-year students enrolled on Phil Hughes’ GEOG20351 “Glaciers” course to the Arenigs in North Wales. I delivered a short presentation on work I did as an undergraduate in 2008 on a nearby lacustrine sequence with an excellent chironomid temperature proxy record that seems to date to the late-glacial and early Holocene. I’ve uploaded the document here if anyone wants to take a closer look.
Ben Bell, a PhD candidate has just published his latest collaborative research on cedar pollen and climate variability. His research is focussed on ways in which Cedrus atlantica might be used as a proxy for (palaeo)climate in the Atlas. This paper examines a previously postulated link between pollen grain size and moisture availability, and concludes that moisture availability is not a significantly related to grain size in this context.
The study makes use of a number of methods for determining grain size – light microscopy, scanning electron microscopy, and laser granulometry. I was involved in the laser granulometry aspect, which Ben proves experimentally is comparable to the microscopic methods. Laser granulometry is considerably less time consuming than microscopic examination, so allowed for the large sample size used in this study.
The paper is published in Palynology, and is open access, available here.
I’ve recently become a Registered Scientist. This is a promotion from my previous registration, as a Registered Science Technician, and is all part of the Science Council’s scheme. I completed my registration via the Institute for Science and Technology, although there are subject-specific bodies that might be more suitable in the future.
I knocked this app up for students who were struggling to draw diagrams for their reports on particle size. It’s my first app written and published with shiny, and I’m looking forward to using this interface for more of my code in the future.
How to Use This App:
- Your data should be saved as a comma-separated-value, or *.csv file – NOT an Excel file. Remember to select this option in “Save as…” if you are using Excel.
- Your data should have sieve sizes in rows, and sample names in columns. The pan should have a size of zero (0). Here’s an example to download and use as a template.
- Your sieve sizes should be expressed in microns.
- Visit https://tombishop.shinyapps.io/histogramR/.
- If you’ve included multiple samples (in columns), select the column you’d like to plot.
- If you want to use units of phi, rather than mm, toggle that setting.
- To copy the image, just right-click it and save it.
- If you need statistics calculating for your data, you could use Regis Gallon’s excellent G2Sd program – just remember to set the “sep” parameter to “,” and the “dec” parameter to “.”.
- Remember to label your axes as appropriate.
I’ve got a beta version of my CIC and CRS model functions up on my GitHub page – let me know if you spot any glaring errors!
I’ve been working on ways to make Itrax data more useful to casual users – I figured one way to do this would be to provide some kind of standard report for each scan (or core sequence), with a stratigraphic diagram, some zonation and multivariate analysis. I’ve decided to do this in R, as it is freely available, cross-platform, handles large datasets and has some existing packages that are useful in manipulating scanning XRF data. At present the functionality is very basic (a bit like my understanding of R). I’ve made the following functions available on my Github repository:
- Import: A function for importing Itrax data into R and cleaning it up a bit on the way. Can also plot the data.
- Ordination: Performs correspondence analysis, with various options for preparing the data. Also provides biplots.
- Correlation: Generates correlation matrixes for Itrax data, and some visualisation.
- Average: Averages Itrax data into a smaller dataset.
I’ll update as I add or modify functionality and documentation. I’m particularly interested to hear from others who are writing code for working with Itrax data, as I think it would make sense to collaborate and work towards a single, powerful suite of tools. Currently my plan is to begin to incorporate some of Menno Bloemsma’s methodology (parts of Itraxelerate) into R, whilst also working on a printable “standard” core data report that can be generated in batches from raw data.
I’ve just discovered the very useful counter function in PAST. PAST is a statistical software package designed specifically for palaeontological data, and can do all sorts of tests and exploratory data processing. I’ve recently moved to version 3.14. One function I’ve just noticed is the counter – this enables you to input counts directly into a spreadsheet using the keys on your keyboard. It also provides auditory feedback and a total count. The software and instructions for its use are available from Øyvind Hammer’s website.
Interpreting palaeoecological data can be a opaque process, differing considerably between workers, and oftentimes scholars have some difficulty describing their own decision making process, or interpreting that of others, particularly in formal written formats like journal articles. This probably has a lot to do with the nature of multi-proxy palaeoecological investigations, where sources of information can be multiple, conflicting, incomplete, imprecise, and unreliable.
Often palaeoecological investigators don’t know exactly what information they will find in a palaeoecological archive before they analyse it, or what the quality of that information will be. This limits the use of statistical hypothesis test – for example, defining a hypothesis (and null hypothesis) to test for significance, although it has some limited application with quantitative data. Traditional hypothesis testing tends to focus on the most likely scenarios, rather than all of the proposed hypotheses. This got me thinking of other ways of testing hypotheses with palaeoecological data.
In many ways, palaeoecological data is a lot like intelligence, medical, or forensic data – information is derived from multiple, different, incomplete, unreliable sources, and can be interpreted in different ways. It is comprised of imperfect evidence preserved after some event or epoch, and it is up to the researcher(s) to compose different information sources into some coherent, plausible sequence of events, causal explanation and/or quantitative information about a past environment. This led me to take a look at methodical ways of testing hypotheses used in other fields.
For example, anyone who has seen the medical drama “House, M.D.” will be familiar with the fast-paced “differential diagnosis” sessions Gregory House (Hugh Laurie) holds with his team. The system is commonly taught in medical schools to assist medical practitioners to come to a diagnosis of a patient’s condition when the symptoms presented are similar. It also allows a medical practitioner to select an appropriate diagnostic if they are unable to discriminate between two or more diagnoses. The process can be broadly summarised as:
- Gather all information.
- List all possible causes.
- Prioritise the list by risk to the patient’s health.
- Working from the highest priority to the lowest, rule out each condition using the available information.
This simple model perhaps mirrors the approach informally adopted by many palaeoecological workers – collect data, hypothesise, rule out until settled on answers. This model fails to accommodate the possibility of competing hypotheses that cannot be adequately differentiated because of limitations in the information available. The intelligence analytical community has developed a way of reasoning and testing hypotheses called the “Analysis of Competing Hypotheses” (ACH). This approach can accommodate the various imperfections of the information available, and can indicate (qualitatively) the likelihood of a particular hypothesis being false. To summarise, the process goes something like this:
- Identify the possible hypotheses.
- List information and arguments (inc. assumptions and deductions) both for and against each hypothesis.
- Assess the relative “diagnosticity” of each piece of information.
- Prepare a matrix with hypotheses in columns, and all evidence and/or arguments in rows.
- Assess how consistent each piece of information or argument is with each hypothesis, attempting to refute each hypothesis.
- Reconsider the hypotheses, removing sources that don’t help discriminate, and identify further evidence required.
- Iterate steps 2-7 as required.
- Draw tentative conclusions about the relative likelihood of each hypothesis (rank them).
- Consider how sensitive your conclusion is the a few critical items of information, and the consequences thereof.
- Report conclusions, discussing all hypotheses.
ACH was introduced by Richard Heuer in “The Psychology of Intelligence Analysis” (CIA) to combat confirmation bias in the field of intelligence analysis, to facilitate multiple workers to address a common problem with multiple lines of evidence, and to create an audit trail for intelligence decisions.
It’s clear that with multiple lines of evidence, weighting, and the iterations, this could quickly become more complicated than just muddling through the data. This is perhaps why there is a growing market for consultants marketing their software and services in intelligence, forensics and criminal investigation. Fortunately both Richard Heuer’s treatise on the subject, alongside some powerful software to assist, are available gratis online.
I’d be interested to hear from anyone who’d like to try (or has tried) using ACH in their analysis of palaeoenvironmental data. I’ll happily configure a portable web-server if you’d like to try the software based version in a group meeting.