1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
85
86
87
88
89
90
91
92
93
94
95
96
97
98
99
100
101
102
103
104
105
106
107
108
109
110
111
112
113
114
115
116
117
118
119
120
121
122
123
124
125
126
127
128
129
130
131
132
133
134
135
136
137
138
139
140
141
142
143
144
145
146
147
148
149
150
151
152
153
154
155
156
157
158
159
160
161
162
163
164
165
166
167
168
169
170
171
172
173
174
175
176
177
178
179
180
181
182
183
184
185
186
187
188
189
190
191
192
193
194
195
196
197
198
199
200
201
202
203
204
205
206
207
208
209
210
211
212
213
214
215
216
217
218
219
220
221
222
223
224
225
226
227
228
229
230
231
232
233
234
235
236
237
238
239
240
241
242
243
244
245
246
247
248
249
250
251
252
253
254
255
256
257
258
259
260
261
262
263
264
265
266
267
268
269
270
271
272
273
274
275
276
277
278
279
280
281
282
283
284
285
286
287
288
289
290
291
292
293
294
295
296
297
298
299
300
301
302
303
304
305
306
307
308
309
310
311
312
313
314
315
316
317
318
319
320
321
322
323
324
325
326
327
328
329
330
331
332
333
334
335
336
337
338
339
340
341
342
343
344
345
346
347
348
349
350
351
352
353
354
355
356
357
358
359
360
361
362
363
364
365
366
367
368
369
370
371
372
373
374
375
376
377
378
379
380
381
382
383
384
385
386
387
388
389
390
391
392
393
394
395
396
397
398
399
400
401
402
403
404
405
406
407
408
409
410
411
412
413
414
415
416
417
418
419
420
421
422
423
424
425
426
427
428
429
430
431
432
433
434
435
436
437
438
439
440
441
442
443
444
445
446
447
448
449
450
451
452
453
454
455
456
457
458
459
460
461
462
463
464
465
466
467
468
469
470
471
472
473
474
475
476
477
478
479
480
481
482
483
484
485
486
487
488
489
490
491
492
493
494
495
496
497
498
499
500
501
502
503
504
505
506
507
508
509
510
511
512
513
514
515
516
517
518
519
520
521
522
523
524
525
526
527
528
529
530
531
532
533
534
535
536
537
538
539
540
541
542
543
544
545
546
547
548
549
550
551
552
553
554
555
556
557
558
559
560
561
562
563
564
565
566
567
568
569
570
571
572
573
574
575
576
577
578
579
580
581
582
583
584
585
586
587
588
589
590
591
592
593
594
595
596
597
598
599
600
601
602
603
604
605
606
607
608
609
610
611
612
613
614
615
616
617
618
619
620
621
622
623
624
625
626
627
628
629
630
631
632
633
634
635
636
637
638
639
640
641
642
643
644
645
646
647
648
649
650
651
652
653
654
655
656
657
658
659
660
661
662
663
664
665
666
667
668
669
670
671
672
673
674
675
676
677
678
679
680
681
682
683
684
685
686
687
688
689
690
691
692
693
694
695
696
697
698
699
700
701
702
703
704
705
706
707
708
709
710
711
712
713
714
715
716
717
718
719
720
721
722
723
724
725
726
727
728
729
730
731
732
733
734
735
736
737
738
739
740
741
742
743
744
745
746
747
748
749
750
751
752
753
754
755
756
757
758
759
760
761
762
763
764
765
766
767
768
769
770
771
772
773
774
775
776
777
778
779
780
781
782
783
784
785
786
787
788
789
790
791
792
793
794
795
796
797
798
799
800
801
802
803
804
805
806
807
808
809
810
811
812
813
814
815
816
817
818
819
820
821
822
823
824
825
826
827
828
829
830
831
832
833
834
835
836
837
838
839
840
841
842
843
844
845
846
847
848
849
850
851
852
853
854
855
856
857
858
859
860
|
Introduction
============
This description is for *creators* of the reproduction pipeline. See
`README` for instructions on running it.
This project contains a **fully working template** for a high-level
research reproduction pipeline, or reproducible paper, as defined in the
link below. If the link below is not accessible at the time of reading,
please see the appendix at the end of this file for a portion of its
introduction. Some
[slides](http://akhlaghi.org/pdf/reproducible-paper.pdf) are also
available to help demonstrate the concept implemented here.
http://akhlaghi.org/reproducible-science.html
This template is created with the aim of supporting reproducible research
by making it easy to start a project in this framework. As shown below, it
is very easy to customize this template reproducible paper pipeline for any
particular research/job and expand it as it starts and evolves. It can be
run with no modification (as described in `README`) as a demonstration and
customized for use in any project as fully described below. The [final
reproducible paper
output](https://gitlab.com/makhlaghi/reproducible-paper-output/raw/master/paper.pdf)
of this pipeline is also present in [a separate
repository](https://gitlab.com/makhlaghi/reproducible-paper-output).
Below, we start with a discussion of why Make was chosen as the high-level
language/framework for this research reproduction pipeline and how to learn
and master Make easily (and freely). The general architecture and design of
the pipeline is then discussed to help you navigate the files and their
contents. This is followed by a checklist for the easy/fast customization
of this pipeline to your exciting research. We continue with some tips and
guidelines on how to manage or extend the pipeline as your research grows
based on our experiences with it so far. The main body concludes with a
description of possible future improvements that are planned for the
pipeline (but not yet implemented). As discussed above, we end with a short
introduction on the necessity of reproducible science in the appendix.
Please don't forget to share your thoughts, suggestions and criticisms on
this pipeline. Maintaining and designing this pipeline is itself a separate
project, so please join us if you are interested. Once it is mature enough,
we will describe it in a paper (written by all contributors) for a formal
introduction to the community.
Why Make?
---------
When batch processing is necessary (no manual intervention, as in a
reproduction pipeline), shell scripts are usually the first solution that
come to mind. However, the inherent complexity and non-linearity of
progress in a scientific project (where experimentation is key) make it
hard to manage the script(s) as the project evolves. For example, a script
will start from the top/start every time it is run. So if you have already
completed 90% of a research project and want to run the remaining 10% that
you have newly added, you have to run the whole script from the start
again. Only then will you see the effects of the last new steps (to find
possible errors, or better solutions and etc).
It is possible to manually ignore/comment parts of a script to only do a
special part. However, such checks/comments will only add to the complexity
of the script and will discourage you to play-with/change an already
completed part of the project when an idea suddenly comes up. It is also
prone to very serious bugs in the end (when trying to reproduce from
scratch). Such bugs are very hard to notice during the work and frustrating
to find in the end.
The Make paradigm, on the other hand, starts from the end: the final
*target*. It builds a dependency tree internally, and finds where it should
start each time the pipeline is run. Therefore, in the scenario above, a
researcher that has just added the final 10% of steps of her research to
her Makefile, will only have to run those extra steps. With Make, it is
also trivial to change the processing of any intermediate (already written)
*rule* (or step) in the middle of an already written analysis: the next
time Make is run, only rules that are affected by the changes/additions
will be re-run, not the whole analysis/pipeline.
This greatly speeds up the processing (enabling creative changes), while
keeping all the dependencies clearly documented (as part of the Make
language), and most importantly, enabling full reproducibility from scratch
with no changes in the pipeline code that was working during the
research. This will allow robust results and let the scientists get to what
they do best: experiment and be critical to the methods/analysis without
having to waste energy and time on technical problems that come up as a
result of that experimentation in scripts.
Since the dependencies are clearly demarcated in Make, it can identify
independent steps and run them in parallel. This further speeds up the
processing. Make was designed for this purpose. It is how huge projects
like all Unix-like operating systems (including GNU/Linux or Mac OS
operating systems) and their core components are built. Therefore, Make is
a highly mature paradigm/system with robust and highly efficient
implementations in various operating systems perfectly suited for a complex
non-linear research project.
Make is a small language with the aim of defining *rules* containing
*targets*, *prerequisites* and *recipes*. It comes with some nice features
like functions or automatic-variables to greatly facilitate the management
of text (filenames for example) or any of those constructs. For a more
detailed (yet still general) introduction see the article on Wikipedia:
https://en.wikipedia.org/wiki/Make_(software)
Many implementations of Make exist and all should be usable with this
pipeline. This pipeline has been created and tested mainly with GNU Make
which is the most common implementation. But if you see parts specific to
GNU Make, please inform us to correct it.
How can I learn Make?
---------------------
The best place to learn Make from scratch is the GNU Make manual. It is an
excellent and non-technical (in its first chapters) book to help get
started. It is freely available and always up to date with the current
release. It also clearly explains which features are specific to GNU Make
and which are general in all implementations. So the first few chapters
regarding the generalities are useful for all implementations.
The first link below points to the GNU Make manual in various formats and
in the second, you can download it in PDF (which may be easier for a first
time reading).
https://www.gnu.org/software/make/manual/
https://www.gnu.org/software/make/manual/make.pdf
If you use GNU Make, you also have the whole GNU Make manual on the
command-line with the following command (you can come out of the "Info"
environment by pressing `q`).
```shell
$ info make
```
If you aren't familiar with the Info documentation format, we strongly
recommend running `$ info info` and reading along. In less than an hour,
you will become highly proficient in it (it is very simple and has a great
manual for itself). Info greatly simplifies your access (without taking
your hands off the keyboard!) to many manuals that are installed on your
system, allowing you to be much more efficient as you work. If you use the
GNU Emacs text editor (or any of its variants), you also have access to all
Info manuals while you are writing your projects (again, without taking
your hands off the keyboard!).
Published works using this pipeline
-----------------------------------
The links below will guide you to some of the works that have already been
published using the method of this pipeline. Note that this pipeline is
evolving, so some small details may be different in them, but they can be
used as a good working model to build your own.
- Section 7.3 of Bacon et
al. ([2017](http://adsabs.harvard.edu/abs/2017A%26A...608A...1B), A&A
608, A1): The version controlled reproduction pipeline is available [on
Gitlab](https://gitlab.com/makhlaghi/muse-udf-origin-only-hst-magnitudes)
and a snapshot of the pipeline along with all the necessary input
datasets and outputs is available in
[zenodo.1164774](https://doi.org/10.5281/zenodo.1164774).
- Section 4 of Bacon et
al. ([2017](http://adsabs.harvard.edu/abs/2017A%26A...608A...1B), A&A,
608, A1): The version controlled reproduction pipeline is available [on
Gitlab](https://gitlab.com/makhlaghi/muse-udf-photometry-astrometry) and
a snapshot of the pipeline along with all the necessary input datasets
is available in
[zenodo.1163746](https://doi.org/10.5281/zenodo.1163746).
- Akhlaghi & Ichikawa
([2015](http://adsabs.harvard.edu/abs/2015ApJS..220....1A), ApJS, 220,
1): The version controlled reproduction pipeline is available [on
Gitlab](https://gitlab.com/makhlaghi/NoiseChisel-paper). This is the
very first (and much less mature) implementation of this pipeline: the
history of this template pipeline started more than two years after that
paper was published. It is a very rudimentary/initial implementation,
thus it is only included here for historical reasons. However, the
pipeline is complete and accurate and uploaded to arXiv along with the
paper. See the more recent implementations if you want to get ideas for
your version of this pipeline.
Citation
--------
A paper will be published to fully describe this reproduction
pipeline. Until then, if this pipeline is useful in your work, please cite
the paper that implemented the first version of this pipeline: Akhlaghi &
Ichikawa ([2015](http://adsabs.harvard.edu/abs/2015ApJS..220....1A), ApJS,
220, 1).
The experience gained with this template after several more implementations
will be used to make this pipeline robust enough for a complete and useful
paper to introduce to the community afterwards.
Also, when your paper is published, don't forget to add a notice in your
own paper (in coordination with the publishing editor) that the paper is
fully reproducible and possibly add a sentence or paragraph in the end of
the paper shortly describing the concept. This will help spread the word
and encourage other scientists to also publish their reproduction
pipelines.
Reproduction pipeline architecture
==================================
In order to adopt this pipeline to your research, it is important to first
understand its architecture so you can navigate your way in the directories
and understand how to implement your research project within its
framework. In short, when the user runs `make` to start the processing, the
first file that is read is the top-level `Makefile`. Therefore, we'll start
our navigation with this file. This file is heavily commented so hopefully
the descriptions in each comment will be enough to understand the general
details. As you read this section, please also look at the contents of the
mentioned files and directories to fully understand what is being said.
Before starting to look into the top Makefile, it is important to recall
that Make defines dependencies by files. Therefore, the input and output of
every step must be a file. Also recall that Make will use the modification
date of the prerequisite and target files to see if the a target must be
re-built or not. Therefore during the processing _many_ intermediate files
will be created (see the tips section below on a good strategy to deal with
large/huge files). Therefore, in configuration time, the user can define a
top-level build directory variable (or `$(BDIR)`) to host all the
intermediate files. This directory doesn't need to be version controlled or
even synchronized or backed-up in other servers: its contents are all
products of the pipeline after all, and can be easily re-created any
time. As you define targets, it is thus important to place them all under
sub-directories of `$(BDIR)`.
Let's start reviewing the processing with the top Makefile. Please open and
inspect it as we go along here. The first step is to define the ultimate
target (`paper.pdf`). You shouldn't modify this line. The rule to build
`paper.pdf` is in another Makefile that will be imported into this top
Makefile later. Don't forget that Make goes over all the process once (to
define dependencies and etc) and then starts its execution. So it is fine
to define the rule to build `paper.pdf` at a later stage (this is the
beauty of Make after all).
Having defined the top target, we will import all the necessary
Makefiles. As you see in `Makefile`, first we include all
`reproduce/config/pipeline/*.mk`. The configuration of each logical step of
the pipeline is placed here as a separate file. These Makefiles must only
contain raw Make variables (pipeline configurations). By raw we mean that
the Make variables in these files must not depend on any other variables
because we don't want to assume any order in reading them. It is very
important to *not* define any rule or other Make construct in any of these
_configuration-Makefiles_ (see the next paragraph for Makefiles with
rules). This will enable you to set the respective files in this directory
as a prerequisite to any target that depends on their variable
values. Therefore, if you change any of the values, all targets that depend
on those values will be re-built.
Once all the raw variables have been imported into the top Makefile, we are
ready to import the Makefiles containing the details of the processing
steps (Makefiles containing rules, let's call these
_workhorse-Makefiles_). But *order is important* in this phase because the
prerequisites of most rules will be other rules that will be defined at a
lower level (not a fixed name like `paper.pdf`). The lower-level rules must
be imported into Make before the higher-level ones. Hence, we can't use a
simple wildcard like when we imported configuration-Makefiles above. All
these Makefiles are defined in `reproduce/src/make`, therefore, the top
Makefile uses the `foreach` function to read them in a specific order.
The main body of this pipeline is thus going to be managed within the
workhorse-Makefiles of `reproduce/src/make`. If you set clear-to-understand
names for these workhorse-Makefiles and follow the convention of the top
Makefile that you only include one workhorse-Makefile per line, the
`foreach` loop of the top Makefile that imports them will become very easy
to read and understand by eye. This will let you know which step you are
taking before or after another without much thought (in a few months
especially). Projects will scale up very fast. Thus if you don't start and
continue with a clean and robust management strategy, in the end it will
become very dirty and hard to manage/understand (even for yourself). As a
general rule of thumb, break your rules into as many logically-similar but
independent steps as possible.
All processing steps ultimately (usually after many rules) end up in some
number, image, figure, or table that must be included in the paper. After
all, if you don't want to report the value of a processing, why would you
do it in the first place? Therefore if the targets in a workhorse-Makefile
aren't directly a prerequisite of other workhorse-Makefile targets, they
should be a pre-requisite of an intermediate LaTeX macro file in
`$(BDIR)/tex/macros` (the highest-level target of that workhorse-Makefile).
The last part of the top-level Makefile is the rule to build
`tex/pipeline.tex`. This file is the connection between the processing
steps of the pipeline and the creation of the final PDF. In
`reproduce/src/make/paper.mk`, you will notice that `paper.pdf` (final
target of the whole reproduction pipeline) depends on
`tex/pipeline.tex`. This file is thus the connection of these two very
high-level different phases of the reproduction pipeline. Therefore, to
keep the over-all management clean, the rule to create this bridge between
the processing and paper-writing phases is defined in the top-level
Makefile.
But `tex/pipeline.tex` is only a merging/concatenation of LaTeX macros
defined as the output of each high-level processing step. In some cases you
want tables and images to also be included in the final PDF. To keep these
necessary LaTeX inputs, you can define other directories under
`$(BDIR)/tex` in the relevant workhorse-Makefile. One of the LaTeX macros
that `reproduce/src/make/initialize.mk` creates is the location of the
build directory, so you can easily guide LaTeX to look into the proper
directory through the `\bdir` macro. If the target of the rule that creates
these other LaTeX inputs isn't a prerequisite of other rules, add it as a
pre-requisite of `tex/pipeline.tex`.
During the research, it often happens that you want to test a step that is
not a prerequisite of any higher-level operation. In such cases, you can
(temporarily) define the target of that rule as a prerequisite of
`tex/pipeline.tex`. If your test gives a promising result and you want to
include it in your research, set it as prerequisites to other rules and
remove it from the list of prerequisites for `tex/pipeline.tex`. In fact,
this is how a project is designed to grow in this framework.
Summary
-------
Based on the explanation above, some major design points you should have in
mind are listed below.
- Define new `reproduce/src/make/XXXXXX.mk` workhorse-Makefile(s) with
good and human-friendly name(s) replacing `XXXXXX`.
- Add `XXXXXX`, as a new line, to the loop which includes the
workhorse-Makefiles in the top-level `Makefile`.
- Do not use any constant numbers (or important names like filter names)
in the workhorse-Makefiles. Define such constants as logically-grouped
separate configuration-Makefiles in `reproduce/config/pipeline`. Then
set the respective configuration-Makefiles file as a pre-requisite to
any rule that uses the variable defined in it.
- To be executed, any target should either be a prerequisite of another
rule (possibly in another Makefile), or a file that is directly imported
into LaTeX as fixed macros for inclusion in text or LaTeX settings (in
`$(BDIR)/tex/macros`), images, plots or tables (in other `$(BDIR)/tex`
sub-directories). In any cases, through any number of intermediate
prerequisites, all processing steps should end in (be a prerequisite of)
`tex/pipeline.tex`.
Checklist to customize the pipeline
===================================
Take the following steps to fully customize this pipeline for your research
project. After finishing the list, be sure to run `./configure` and `make`
to see if everything works correctly before expanding it. If you notice
anything missing or any in-correct part (probably a change that has not
been explained here), please let us know to correct it.
- **Get this repository** (if you don't already have it): Arguably the
easiest way to start is to clone this repository as shown below:
```shell
$ git clone https://gitlab.com/makhlaghi/reproducible-paper.git
$ mv reproducible-paper my-project-name
$ cd my-project-name
```
- **Copyright**, **name** and **date**: Go over the following files and
add your name and email to the copyright notice: `configure`,
`Makefile` and `reproduce/src/make/*.mk`. When making new files,
always remember to add a similar copyright statement at the top of the
file.
- **Title**, **short description** and **author** in source files: In this
raw skeleton, the title or short description of your project should be
added in the following two files: `Makefile` (the first line), and
`tex/preamble-style.tex` (the last few lines, along with the names of
you and your colleagues). In both cases, the texts you should replace
are all in capital letters to make them easier to identify. Of course,
if you use a different LaTeX method of managing the title and authors,
please feel free to use your own methods, just find a way to keep the
pipeline version in a nicely visible place.
- **Gnuastro**: GNU Astronomy Utilities (Gnuastro) is currently a
dependency of the pipeline and without it, the pipeline will complain
and abort. The main reason for this is to demonstrate how critically
important it is to version your software. If you don't want to install
Gnuastro please follow the instructions in the list below. If you have
installed Gnuastro and tried the pipeline, but don't need Gnuastro in
your pipeline, also follow the list below. If you do want to use
Gnuastro in your pipeline, be sure to un-comment the `onlyversion`
option in `reproduce/config/gnuastro/gnuastro.conf` file and set it to
your version of Gnuastro. This will force you to keep the pipeline in
match with the version of Gnuastro you are using all the time and also
allow commits to be exactly reproducible also (for example if you
update to a new version of Gnuastro during your research project). If
you will be using Gnuastro, you can also remove the "marks" (comments)
put in the relevant files of the list below to make them more
readable.
- Delete the description about Gnuastro in `README`.
- Delete marked part(s) in `configure`.
- Delete marked parts in `reproduce/src/make/initialize.mk`.
- Delete `and Gnuastro \gnuastroversion` from `tex/preamble-style.tex`.
- **Other dependencies**: If there are any more of the dependencies that
you don't use (or others that you need), then remove (or add) them in
the respective parts of `configure`. It is commented thoroughly and
reading over the comments should guide you on what to add/remove and
where. Note that it is always good to have an option to download the
necessary datasets in case the user doesn't have them. But in case
your pipeline doesn't need any downloads, you can also remove the
sections of `configure` that are for `flock` and the downloader.
- **`README`**: Go through this top-level instruction file and make it fit
to your pipeline: update the text and etc. Don't forget that your
colleagues or anyone else, will first be drawn to read this file, so
make it as easy as possible for them to understand your
work. Therefore, also check and update `README` one last time when you
are ready to publish your work (and its reproduction pipeline).
- **First input dataset**: The user manages the top-level directory of the
input data through the variables set in
`reproduce/config/pipeline/LOCAL.mk.in` (the user actually edits a
`LOCAL.mk` file that is created by `configure` from the `.mk.in` file,
but the `.mk` file is not under version control). So open this file
and replace `SURVEY` in the variable name and value with the name of
your input survey or dataset (all in capital letters), for example if
you are working on data from the XDF survey, replace `SURVEY` with
`XDF`. Don't change anything else in the value, just the the all-caps
name. Afterwards, change any occurrence of `SURVEY` in the whole
pipeline with the new name. You can find the occurrences with a simple
command like the ones shown below. We follow the Make convention here
that all `ONLY-CAPITAL` variables are those directly set by the user
and all `small-caps` variables are set by the pipeline designer. All
variables that also depend on this survey have a `survey` in their
name. Hence, also correct all these occurrences to your new name in
small-caps. Of course, ignore those occurrences that are irrelevant,
like those in this file. Note that in the raw version of this template
no target depends on these files, so they are ignored. Afterwards, set
the webpage and correct the filenames in
`reproduce/src/make/download.mk` if necessary.
```shell
$ grep -r SURVEY ./
$ grep -r survey ./
```
- **Other input datasets**: Add any other input datasets that may be
necessary for your research to the pipeline based on the example
above.
- **Paper title and authors**: The final paper's title, authors and other
information are defined in the last two sections of the
`tex/preamble-style.tex` file (section that are written in
ONLY-CAPITAL characters). Correct these for your project or use any
other LaTeX title and author management package that you prefer
instead. As you add more packages to the preambles, it is recommended
to follow this convention of having five empty lines between each group
of package importing and configuration along with comments for each
package. This will greatly help you in readability later.
- **Delete dummy parts**: The template pipeline contains some parts that
are only for the initial/test run, not for any real analysis. The
respective files to remove and parts to fix are discussed here.
- `paper.tex`: Delete the main body and abstract of text and
generally.
- `tex/preamble-header.tex`: Correct/add the titles, headers and
authors list of the paper.
- `Makefile`: Delete the two occurrences of `delete-me` in the
`foreach` loops.
- Delete the following files: `README.md`, all `delete-me*` files (in
`reproduce/config/pipeline`, `reproduce/src/make`, and `tex`).
- **Initiate a new Git repo**: You probably don't want to mix the history
of this template reproduction pipeline with your own reproduction
pipeline. You have already made some small changes in the previous
step, so let's re-initiate history before continuing. But before doing
that, keep the output of `git describe` in a place and write it in
your first commit message to document what point in this pipeline's
history you started from. Since the pipeline is highly integrated with
your particular research, it may not be easy to merge the changes
later. Having the commit information that you started from, will allow
you to check and manually apply any changes that don't interfere with
your implemented pipeline. After this step, you can commit your
changes into your newly initiated history as you like.
```shell
$ git describe # The point in this history you started from.
$ git clean -fxd # Remove any possibly created extra files.
$ rm -rf .git # Completely remove this history.
$ git init # Initiate a new history.
$ git add --all # Stage everything that is here.
$ git commit # Your first commit (mention the first output).
$ git tag -a v0 # Tag this as the zero-th version of your pipeline.
```
- **Notice on reproducibility**: Add a notice somewhere prominent in the
first page within your paper, informing the reader that your research
is fully reproducible. For example in the end of the abstract, or
under the keywords with a title like "reproducible paper". This will
encourage them to publish their own works in this manner also and also
will help spread the word.
- **Start your exciting research**: You are now ready to add flesh and
blood to this raw skeleton by further modifying and adding your
exciting research steps. You can use the "published works" section in
the introduction as some fully working models to learn from.
- **Feedback**: As you use the pipeline you will notice many things that
if implemented from the start would have been very useful for your
work. This can be in the actual file structure of the pipeline or in
useful implementation and usage tips like below. In any case, since
this concept is still evolving, please share your thoughts and
suggestions, so we can add them here for everyone's benefit.
Usage tips: designing your pipeline/workflow
============================================
The following is a list of design points, tips, or recommendations that
have been learned after some experience with this pipeline. Please don't
hesitate to share any experience you gain after using this pipeline with
us. In this way, we can add it here for the benefit of others.
- **Modularity**: Modularity is the key to easy and clean growth of a
project. So it is always best to break up a job into as many
sub-components as reasonable. Here are some tips to stay modular.
- *Short recipes*: if you see the recipe of a rule becoming more than a
handful of lines which involve significant processing, it is probably
a good sign that you should break up the rule into its main
components. Try to only have one major processing step per rule.
- *Context-based (many) Makefiles*: This pipeline is designed to allow
the easy inclusion of many Makefiles (in `reproduce/src/make/*.mk`)
for maximal modularity. So keep the rules for closely related parts
of the processing in separate Makefiles.
- *Descriptive names*: Be very clear and descriptive with the naming of
the files and the variables because a few months after the
processing, it will be very hard to remember what each one was
for. Also this helps others (your collaborators or other people
reading the pipeline after it is published) to more easily understand
your work and find their way around.
- *Naming convention*: As the project grows, following a single standard
or convention in naming the files is very useful. Try best to use
multiple word filenames for anything that is non-trivial (separating
the words with a `-`). For example if you have a Makefile for
creating a catalog and another two for processing it under models A
and B, you can name them like this: `catalog-create.mk`,
`catalog-model-a.mk` and `catalog-model-b.mk`. In this way, when
listing the contents of `reproduce/src/make` to see all the
Makefiles, those related to the catalog will all be close to each
other and thus easily found. This also helps in auto-completions by
the shell or text editors like Emacs.
- *Source directories*: If you need to add files in other languages for
example in shell, Python, AWK or C, keep them in a separate directory
under `reproduce/src`, with the appropriate name.
- *Configuration files*: If your research uses special programs as part
of the processing, put all their configuration files in a devoted
directory (with the program's name) within
`reproduce/config`. Similar to the `reproduce/config/gnuastro`
directory (which is put in the template as a demo in case you use GNU
Astronomy Utilities). It is much cleaner and readable (thus less
buggy) to avoid mixing the configuration files, even if there is no
technical necessity.
- **Contents**: It is good practice to follow the following
recommendations on the contents of your files, whether they are source
code for a program, Makefiles, scripts or configuration files
(copyrights aren't necessary for the latter).
- *Copyright*: Always start a file containing programming constructs
with a copyright statement like the ones that this template starts
with (for example in the top level `Makefile`).
- *Comments*: Comments are vital for readability (by yourself in two
months, or others). Describe everything you can about why you are
doing something, how you are doing it, and what you expect the result
to be. Write the comments as if it was what you would say to describe
the variable, recipe or rule to a friend sitting beside you. When
writing the pipeline it is very tempting to just steam ahead with
commands and codes, but be patient and write comments before the
rules or recipes. This will also allow you to think more about what
you should be doing. Also, in several months when you come back to
the code, you will appreciate the effort of writing them. Just don't
forget to also read and update the comment first if you later want to
make changes to the code (variable, recipe or rule). As a general
rule of thumb: first the comments, then the code.
- *File title*: In general, it is good practice to start all files with
a single line description of what that particular file does. If
further information about the totality of the file is necessary, add
it after a blank line. This will help a fast inspection where you
don't care about the details, but just want to remember/see what that
file is (generally) for. This information must of course be commented
(its for a human), but this is kept separate from the general
recommendation on comments, because this is a comment for the whole
file, not each step within it.
- **Make programming**: Here are some experiences that we have come to
learn over the years in using Make and are useful/handy in research
contexts.
- *Automatic variables*: These are wonderful and very useful Make
constructs that greatly shrink the text, while helping in
read-ability, robustness (less bugs in typos for example) and
generalization. For example even when a rule only has one target or
one prerequisite, always use `$@` instead of the target's name, `$<`
instead of the first prerequisite, `$^` instead of the full list of
prerequisites and etc. You can see the full list of automatic
variables
[here](https://www.gnu.org/software/make/manual/html_node/Automatic-Variables.html). If
you use GNU Make, you can also see this page on your command-line:
```shell
$ info make "automatic variables
```
- *Debug*: Since Make doesn't follow the common top-down paradigm, it
can be a little hard to get accustomed to why you get an error or
un-expected behavior. In such cases, run Make with the `-d`
option. With this option, Make prints a full list of exactly which
prerequisites are being checked for which targets. Looking
(patiently) through this output and searching for the faulty
file/step will clearly show you any mistake you might have made in
defining the targets or prerequisites.
- *Large files*: If you are dealing with very large files (thus having
multiple copies of them for intermediate steps is not possible), one
solution is the following strategy. Set a small plain text file as
the actual target and delete the large file when it is no longer
needed by the pipeline (in the last rule that needs it). Below is a
simple demonstration of doing this, where we use Gnuastro's
Arithmetic program to add all pixels of the input image with 2 and
create `large1.fits`. We then subtract 2 from `large1.fits` to create
`large2.fits` and delete `large1.fits` in the same rule (when its no
longer needed). We can later do the same with `large2.fits` when it
is no longer needed and so on.
```
large1.fits.txt: input.fits
astarithmetic $< 2 + --output=$(subst .txt,,$@)
echo "done" > $@
large2.fits.txt: large1.fits.txt
astarithmetic $(subst .txt,,$<) 2 - --output=$(subst .txt,,$@)
rm $(subst .txt,,$<)
echo "done" > $@
```
A more advanced Make programmer will use Make's [call
function](https://www.gnu.org/software/make/manual/html_node/Call-Function.html)
to define a wrapper in `reproduce/src/make/initialize.mk`. This
wrapper will replace `$(subst .txt,,XXXXX)`. Therefore, it will be
possible to greatly simplify this repetitive statement and make the
code even more readable throughout the whole pipeline.
- **Dependencies**: It is critically important to exactly document, keep
and check the versions of the programs you are using in the pipeline.
- *Check versions*: In `reproduce/src/make/initialize.mk`, check the
versions of the programs you are using.
- *Keep the source tarball of dependencies*: keep a tarball of the
necessary version of all your dependencies (and also a copy of the
higher-level libraries they depend on). Software evolves very fast
and only in a few years, a feature might be changed or removed from
the mainstream version or the software server might go down. To be
safe, keep a copy of the tarballs. Software tarballs are rarely over
a few megabytes, very insignificant compared to the data. If you
intend to release the pipeline in a place like Zenodo, then you can
create your submission early (before public release) and upload/keep
all the necessary tarballs (and data)
there. [zenodo.1163746](https://doi.org/10.5281/zenodo.1163746) is
one example of how the data, Gnuastro (main software used) and all
major Gnuastro's dependencies have been uploaded with the pipeline.
- *Keep your input data*: The input data is also critical to the
pipeline, so like the above for software, make sure you have a backup
of them.
- **Version control**: It is important (and extremely useful) to have the
history of your pipeline under version control. So try to make commits
regularly (after any meaningful change/step/result), while not
forgetting the following notes.
- *Tags*: To help manage the history, tag all major commits. This helps
make a more human-friendly output of `git describe`: for example
`v1-4-gaafdb04` states that we are on commit `aafdb04` which is 4
commits after tag `v1`. The output of `git describe` is included in
your final PDF as part of this pipeline. Also, if you use
reproducibility-friendly software like Gnuastro, this value will also
be included in all output files, see the description of `COMMIT` in
[Output
headers](https://www.gnu.org/software/gnuastro/manual/html_node/Output-headers.html).
In the checklist above, we tagged the first commit of your pipeline
with `v0`. Here is one suggestion on when to tag: when you have fully
adopted the pipeline and have got the first (initial) results, you
can make a `v1` tag. Subsequently when you first start reporting the
results to your colleagues, you can tag the commit as `v2`. Afterwards
when you submit to a paper, it can be tagged `v3` and so on.
- *Pipeline outputs*: During your research, it is possible to checkout a
specific commit and reproduce its results. However, the processing
can be time consuming. Therefore, it is useful to also keep track of
the final outputs of your pipeline (at minimum, the paper's PDF) in
important points of history. However, keeping a snapshot of these
(most probably large volume) outputs in the main history of the
pipeline can unreasonably bloat it. It is thus recommended to make a
separate Git repo to keep those files and keep this pipeline's volume
as small as possible. For example if your main pipeline is called
`my-exciting-project`, the name of the outputs pipeline can be
`my-exciting-project-output`. This enables easy sharing of the output
files with your co-authors (with necessary permissions) and not
having to bloat your email archive with extra attachments (you can
just share the link to the online repo in your communications). After
the research is published, you can also release the outputs pipeline,
or you can just delete it if it is too large or un-necessary (it was
just for convenience, and fully reproducible after all). This
pipeline's output is available for demonstration in the separate
[reproducible-paper-output](https://gitlab.com/makhlaghi/reproducible-paper-output)
repository.
Future improvements
===================
This is an evolving project and as time goes on, it will evolve and become
more robust. Here are the list of features that we plan to add in the
future.
- *Containers*: It is important to have better/full control of the
environment of the reproduction pipeline. Our current reproducible
paper pipeline simply assumes that the necessary software are already
installed on the host system. So it ignores details of how they are
built or the versions of their dependencies (which is not good). As in
[zenodo.1164774](https://doi.org/10.5281/zenodo.1164774) or
[zenodo.1163746](https://doi.org/10.5281/zenodo.1163746), the best we
can currently do is distribute the tarballs of the necessary
software. The solution here is based on [an interesting
tutorial](https://mozillafoundation.github.io/2017-fellows-sf/re-papers/index.html)
by the Mozilla science lab to build reproducible papers. It suggests
using the [Nix package manager](https://nixos.org/nix/about.html) to
build the necessary software for the pipeline and run the pipeline in
its completely closed environment. This is a great solution because
using Nix or [Guix](https://www.gnu.org/software/guix/) (which is based
on Nix, but uses the [Scheme
language](https://en.wikipedia.org/wiki/Scheme_(programming_language)),
not a custom language for the management) will allow a fully working
closed environment on the host system which contains the instructions
on how to build the environment. These package managers can also
co-exist with the package manager of the host's operating system and
they allow separate versions of a software to be present. Thus it is
not necessary to change existing programs on the host system (that have
been updated for example) to run a particular reproduction
pipeline. The availability of the instructions to build the programs
and environment with Nix or Guix, makes them a better solution than
binary containers like [docker](https://www.docker.com/) which are
essentially just a binary (not human readable) black box and only
usable on the given CPU architecture. The initial running of Nix or
Guix and setting up of the environment can also be included in a
`Makefile` of this pipeline, and thus be fully automatic. The software
tarballs (to be used by Nix or Guix) can also be uploaded/archived, as
we do now. These package managers can then be instructed to get the
tarballs for building the environment from there.
Appendix: Necessity of exact reproduction in scientific research
================================================================
In case [the link above](http://akhlaghi.org/reproducible-science.html) is
not accessible at the time of reading, here is a copy of the introduction
of that link, describing the necessity for a reproduction pipeline like
this (copied on February 7th, 2018):
The most important element of a "scientific" statement/result is the fact
that others should be able to falsify it. The Tsunami of data that has
engulfed astronomers in the last two decades, combined with faster
processors and faster internet connections has made it much more easier to
obtain a result. However, these factors have also increased the complexity
of a scientific analysis, such that it is no longer possible to describe
all the steps of an analysis in the published paper. Citing this
difficulty, many authors suffice to describing the generalities of their
analysis in their papers.
However, It is impossible to falsify (or even study) a result if you can't
exactly reproduce it. The complexity of modern science makes it vitally
important to exactly reproduce the final result. Because even a small
deviation can be due to many different parts of an analysis. Nature is
already a black box which we are trying so hard to comprehend. Not letting
other scientists see the exact steps taken to reach a result, or not
allowing them to modify it (do experiments on it) is a self-imposed black
box, which only exacerbates our ignorance.
Other scientists should be able to reproduce, check and experiment on the
results of anything that is to carry the "scientific" label. Any result
that is not reproducible (due to incomplete information by the author) is
not scientific: the readers have to have faith in the subjective experience
of the authors in the very important choice of configuration values and
order of operations: this is contrary to the scientific spirit.
|