Summary

  • Raw data files should be stored in a single directory named data.
  • data should be centrally accessible.
  • Files in data should have unique names.
  • data should have a flat structure (no subdirectories).
  • Files in data should remain pristine as they came off the instrument and never be altered.
  • Do not use data as a working directory. I.e. don't put code, manuscripts, etc. in data; use a project directory instead.

Introduction

If you are like me, in the course of your research you generate data files from the various apparatus used in your experiments. These may be SEM images in TIFF format, files containing I-V data for your Schottky diodes in ASCII format, or AFM images in some proprietary binary format. How should you organize these files? How should you store them so others can get access when they need to (without bothering you)? How can you keep track of them and any changes during data analysis?

The solution: unique filenames, read-only files, and a single, flat data directory

Most of your data management issues can be resolved by adopting a single, flat (no subdirectories), centrally accessable directory containing read-only copies of your group's data files which are uniquely named.

The workflow for this kind of system is straightforward: data is taken on an instrument and gets copied to data. If analysis is to be done on files in data, the person analyzing the data would first copy the files into a different project directory and then perform the analysis.

Having all data in a single, centrally accessible data directory has a number of advantages. First, this approach eliminates findability costs. People will never have to go around to different computers looking for the data they took. Second, this approach decreases transaction costs. If you need data that I took, simply find it in data. There's no need to ask me to email or otherwise send you the data.

This approach to data management is strongly coupled to choosing unique, metadata-rich filenames. To recap, use

YYYYMMDD-HHMM_experiment_sample_experimenter.extension

on systems that support long filenames and the nested directory structure

experimenter/YYYYMMDD/experiment/sample/HHMM.extension

on older systems that do not support long filenames.

Having a data directory filled with uniquely and descriptively named files is the next best thing to having a URI for each file and having the files in some kind of relational database. In addition to the advantage of findability I mentioned above, uniquely naming files in a single data directory yields addressability. Cross-referencing a file in a lab notebook or sample log is as simple as writing the filename. To find the cross-referenced file, simply look for it in data.

Using metadata-rich filenames gives you another advantage: it is easy to perform searches for specific files or sets of files using filename search functions in your file browser or shell. More advanced searches can be done via simple python scripts. For example, lets say I wanted to find all the XPS data files I've taken

gamma:data jrsmith3$ ls | grep -i jrs | grep -i xps
20100202-0844_xps_tfan25_jrs.dat
20100202-0850_xps_tfan25_jrs.dat
20100202-1008_xps_tfan25_jrs.dat
20100202-1019_xps_tfan25_jrs.dat
20100202-1127_xps_tfan25_jrs.dat
20100202-1133_xps_tfan25_jrs.dat
20100202-1223_xps_tfan25_jrs.dat
20100202-1227_xps_tfan25_jrs.dat
20100202-1324_xps_tfan25_jrs.dat
20100202-1330_xps_tfan25_jrs.dat
20100203-0823_xps_tfan24_jrs.dat
20100203-0829_xps_tfan24_jrs.dat
20100203-0932_xps_tfan24_jrs.dat
20100203-0937_xps_tfan24_jrs.dat
20100203-0956_xps_tfan24_jrs.dat
20100203-1033_xps_tfan24_jrs.dat
20100203-1039_xps_tfan24_jrs.dat

The same search can be performed from OSX Finder. I don't use Windows, but I would be happy to post instructions on the same search from powershell or Windows file Explorer, email me or pull request.

Using the file naming scheme I suggest above gives one final advantage. File browsers and the shell will automatically list files in chronological order. In the bash shell:

gamma:data jrsmith3$ ls
20110520-1543_stm_jrs0075_jrs.sm4
20110520-1614_stm_jrs0075_jrs.sm4
20110520-1622_oscilloscope_jrs0075_jrs.txt
20110520-1623_stm_jrs0075_jrs.sm4
20110520-1700_stm_jrs0075_jrs.sm4
20110520-1709_stm_jrs0075_jrs.sm4
20110520-1721_stm_jrs0075_jrs.sm4
20110520-1731_stm_jrs0075_jrs.sm4
20110520-1742_stm_jrs0075_jrs.sm4
20110520-1819_stm_jrs0076_jrs.sm4
20110520-1830_stm_jrs0076_jrs.sm4
20110520-1840_stm_jrs0076_jrs.sm4
20110520-1840_stm_jrs0076_jrs.txt
20110520-1916_stm_jrs0076_jrs.sm4
20110520-1925_stm_jrs0076_jrs.sm4
20110520-2002_stm_jrs0076_jrs.sm4
20110523-1052_xps_jrs0076_jrs.dat
20110523-1110_xps_jrs0076_jrs.dat
20110527-1614_TEM_JRST27_ATW.dm3
20110527-1615_TEM_JRST27_ATW.dm3
20110527-1617_TEM_JRST27_ATW.dm3
20110527-1619_TEM_JRST27_ATW.dm3
20110527-162412_TEM_JRST26_ATW.dm3
20110527-162430_TEM_JRST26_ATW.dm3
20110527-1627_TEM_JRST26_ATW.dm3
20110527-1628_TEM_JRST26_ATW.dm3
20110527-1635_TEM_JRST26_ATW.dm3

Workflow

The workflow is pretty simple: Data files are created on the instrument and are named appropriately. Data then moves into the data directory either directly or via an intermediate medium such as a USB flash drive. If you are on Windows, I believe WinSCP has a nice synchronization capability.

Cautions

There are two related issues that you should avoid. First, putting all data in a single directory creates a single point of failure. I suggest having a robust backup and recovery plan for this directory.

The second thing you should avoid is altering any data that is located in the data directory. Oftentimes you will need to do some kind of analysis that changes the data file itself; if so, make a copy of the data you need and modify the copy. To reiterate: once data lands in the data directory, it should never change.

Putting data under version control might not be a bad idea. Some people might balk at having a big repo, but apparently some companies do it.

I would recommend using git or another vcs if your lab's data is a few hundred MBs.

Short filenames

If you've been paying attention, you will have noticed a contradiction. According to my file naming rubric, short filenames are necessarily in a nested directory structure but I advocate a flat structure for data.

The best case scenario would be to have some scripts for various platforms which would flatten the nested structure in data. I will post links to these scripts when they become available. In the meantime, the nested structure I suggest makes drag-and-drop additions to the data directory easy. To keep things in data clean, use a single subdirectory named nested to store all of the data in nested directories coming from computers constrained to short filenames. According to my nested directory naming scheme, the nested directory will contain subdirectories corresponding to each experimenter. Each experimenter's directory will end up having a series of YYYYMMDD subdirectories. In this way, everyone can simply drag-and-drop newly acquired data into a new YYYYMMDD subdirectory in their own directory.

Here's a visualization:

$ tree data/
data/
└── nested
    ├── jrs
    │   └── 20140622
    │       └── xps
    │           └── jrs0014
    │               ├── 0943.dat
    │               ├── 0950.dat
    │               ├── 1002.dat
    │               ├── 1307.dat
    │               └── 1320.dat
    └── rpt

6 directories, 5 files