diff --git a/Paper/Makefile b/Paper/Makefile
new file mode 100644
index 0000000000000000000000000000000000000000..9c0c202082280561c2a74c6b9f07fd7753a5247a
--- /dev/null
+++ b/Paper/Makefile
@@ -0,0 +1,21 @@
+SHELL = /bin/bash
+
+main = paper_cw_mcmc.pdf
+
+.PHONY : main clean figclean arxiv jones
+
+main : $(main)
+
+git_tag.tex :
+	./git-tag.sh $@
+
+$(main): *.tex git_tag.tex
+	pdflatex ${@:.pdf=} && bibtex ${@:.pdf=} && pdflatex ${@:.pdf=} && pdflatex ${@:.pdf=}
+
+clean :
+	rm -f git_tag.tex $(main:.pdf=){.aux,.bbl,.blg,.log,.out,.pdf,Notes.bib} $(texfigs) $(texonlyfigs)
+
+arxiv :
+	rm -f $(main)
+	rm -f git_tag.tex
+	tar -cvf arxiv_submission.tar *tex *bib bibliography.bib *pdf *.bbl
diff --git a/Paper/bibliography.bib b/Paper/bibliography.bib
new file mode 100644
index 0000000000000000000000000000000000000000..5db48ad2842c06e85eb7cfd2f02397ea2696bc1a
--- /dev/null
+++ b/Paper/bibliography.bib
@@ -0,0 +1,380 @@
+@ARTICLE{jks1998,
+   author = {{Jaranowski}, P. and {Kr{\'o}lak}, A. and {Schutz}, B.~F.},
+    title = "{Data analysis of gravitational-wave signals from spinning neutron stars: The signal and its detection}",
+  journal = {\prd},
+   eprint = {gr-qc/9804014},
+ keywords = {Gravitational radiation detectors, mass spectrometers, and other instrumentation and techniques, Gravitational wave detectors and experiments, Mathematical procedures and computer techniques, Pulsars},
+     year = 1998,
+    month = sep,
+   volume = 58,
+   number = 6,
+      eid = {063001},
+    pages = {063001},
+      doi = {10.1103/PhysRevD.58.063001},
+   adsurl = {http://adsabs.harvard.edu/abs/1998PhRvD..58f3001J},
+  adsnote = {Provided by the SAO/NASA Astrophysics Data System}
+}
+
+@ARTICLE{brady1998,
+   author = {{Brady}, P.~R. and {Creighton}, T. and {Cutler}, C. and {Schutz}, B.~F.
+	},
+    title = "{Searching for periodic sources with LIGO}",
+  journal = {\prd},
+   eprint = {gr-qc/9702050},
+ keywords = {Gravitational radiation detectors, mass spectrometers, and other instrumentation and techniques, Gravitational wave detectors and experiments, Mathematical procedures and computer techniques, Pulsars},
+     year = 1998,
+    month = feb,
+   volume = 57,
+    pages = {2101-2116},
+      doi = {10.1103/PhysRevD.57.2101},
+   adsurl = {http://adsabs.harvard.edu/abs/1998PhRvD..57.2101B},
+  adsnote = {Provided by the SAO/NASA Astrophysics Data System}
+}
+
+@ARTICLE{brady2000,
+   author = {{Brady}, P.~R. and {Creighton}, T.},
+    title = "{Searching for periodic sources with LIGO. II. Hierarchical searches}",
+  journal = {\prd},
+   eprint = {gr-qc/9812014},
+ keywords = {Gravitational wave detectors and experiments, Gravitational radiation detectors, mass spectrometers, and other instrumentation and techniques, Mathematical procedures and computer techniques, Pulsars},
+     year = 2000,
+    month = apr,
+   volume = 61,
+   number = 8,
+      eid = {082001},
+    pages = {082001},
+      doi = {10.1103/PhysRevD.61.082001},
+   adsurl = {http://adsabs.harvard.edu/abs/2000PhRvD..61h2001B},
+  adsnote = {Provided by the SAO/NASA Astrophysics Data System}
+}
+
+@ARTICLE{shaltev2013,
+   author = {{Shaltev}, M. and {Prix}, R.},
+    title = "{Fully coherent follow-up of continuous gravitational-wave candidates}",
+  journal = {\prd},
+archivePrefix = "arXiv",
+   eprint = {1303.2471},
+ primaryClass = "gr-qc",
+ keywords = {Gravitational waves: theory},
+     year = 2013,
+    month = apr,
+   volume = 87,
+   number = 8,
+      eid = {084057},
+    pages = {084057},
+      doi = {10.1103/PhysRevD.87.084057},
+   adsurl = {http://adsabs.harvard.edu/abs/2013PhRvD..87h4057S},
+  adsnote = {Provided by the SAO/NASA Astrophysics Data System}
+}
+
+
+@ARTICLE{prix2012,
+   author = {{Prix}, R. and {Shaltev}, M.},
+    title = "{Search for continuous gravitational waves: Optimal StackSlide method at fixed computing cost}",
+  journal = {\prd},
+archivePrefix = "arXiv",
+   eprint = {1201.4321},
+ primaryClass = "gr-qc",
+ keywords = {Gravitational waves: theory, Gravitational-wave astrophysics, Gravitational wave detectors and experiments, Data analysis: algorithms and implementation, data management},
+     year = 2012,
+    month = apr,
+   volume = 85,
+   number = 8,
+      eid = {084010},
+    pages = {084010},
+      doi = {10.1103/PhysRevD.85.084010},
+   adsurl = {http://adsabs.harvard.edu/abs/2012PhRvD..85h4010P},
+  adsnote = {Provided by the SAO/NASA Astrophysics Data System}
+}
+
+@ARTICLE{krishnan2004,
+   author = {{Krishnan}, B. and {Sintes}, A.~M. and {Papa}, M.~A. and {Schutz}, B.~F. and 
+	{Frasca}, S. and {Palomba}, C.},
+    title = "{Hough transform search for continuous gravitational waves}",
+  journal = {\prd},
+   eprint = {gr-qc/0407001},
+ keywords = {Gravitational wave detectors and experiments, Wave generation and sources, Data analysis: algorithms and implementation, data management, Gravitational radiation detectors, mass spectrometers, and other instrumentation and techniques},
+     year = 2004,
+    month = oct,
+   volume = 70,
+   number = 8,
+      eid = {082001},
+    pages = {082001},
+      doi = {10.1103/PhysRevD.70.082001},
+   adsurl = {http://adsabs.harvard.edu/abs/2004PhRvD..70h2001K},
+  adsnote = {Provided by the SAO/NASA Astrophysics Data System}
+}
+
+
+@ARTICLE{astone2014,
+   author = {{Astone}, P. and {Colla}, A. and {D'Antonio}, S. and {Frasca}, S. and 
+	{Palomba}, C.},
+    title = "{Method for all-sky searches of continuous gravitational wave signals using the frequency-Hough transform}",
+  journal = {\prd},
+archivePrefix = "arXiv",
+   eprint = {1407.8333},
+ primaryClass = "astro-ph.IM",
+ keywords = {Gravitational waves: theory, Gravitational wave detectors and experiments, Neutron stars},
+     year = 2014,
+    month = aug,
+   volume = 90,
+   number = 4,
+      eid = {042002},
+    pages = {042002},
+      doi = {10.1103/PhysRevD.90.042002},
+   adsurl = {http://adsabs.harvard.edu/abs/2014PhRvD..90d2002A},
+  adsnote = {Provided by the SAO/NASA Astrophysics Data System}
+}
+
+@ARTICLE{cutler2005,
+   author = {{Cutler}, C. and {Gholami}, I. and {Krishnan}, B.},
+    title = "{Improved stack-slide searches for gravitational-wave pulsars}",
+  journal = {\prd},
+   eprint = {gr-qc/0505082},
+ keywords = {Gravitational wave detectors and experiments, Mathematical procedures and computer techniques, Pulsars},
+     year = 2005,
+    month = aug,
+   volume = 72,
+   number = 4,
+      eid = {042004},
+    pages = {042004},
+      doi = {10.1103/PhysRevD.72.042004},
+   adsurl = {http://adsabs.harvard.edu/abs/2005PhRvD..72d2004C},
+  adsnote = {Provided by the SAO/NASA Astrophysics Data System}
+}
+
+@ARTICLE{prix2009,
+   author = {{Prix}, R. and {Krishnan}, B.},
+    title = "{Targeted search for continuous gravitational waves: Bayesian versus maximum-likelihood statistics}",
+  journal = {Classical and Quantum Gravity},
+archivePrefix = "arXiv",
+   eprint = {0907.2569},
+ primaryClass = "gr-qc",
+     year = 2009,
+    month = oct,
+   volume = 26,
+   number = 20,
+      eid = {204013},
+    pages = {204013},
+      doi = {10.1088/0264-9381/26/20/204013},
+   adsurl = {http://adsabs.harvard.edu/abs/2009CQGra..26t4013P},
+  adsnote = {Provided by the SAO/NASA Astrophysics Data System}
+}
+
+@Article{	  foreman-mackay2013,
+  author	= {{Foreman-Mackey}, D. and {Hogg}, D.~W. and {Lang}, D. and
+		  {Goodman}, J. },
+  title		= "{emcee: The MCMC Hammer}",
+  journal	= {PASP},
+  archiveprefix	= "arXiv",
+  eprint	= {1202.3665},
+  primaryclass	= "astro-ph.IM",
+  keywords	= {Data Analysis and Techniques},
+  year		= 2013,
+  month		= mar,
+  volume	= 125,
+  pages		= {306-312},
+  doi		= {10.1086/670067},
+  adsurl	= {http://adsabs.harvard.edu/abs/2013PASP..125..306F},
+  adsnote	= {Provided by the SAO/NASA Astrophysics Data System}
+}
+
+@Article{	  goodman2010,
+  author	= {{Goodman}, J. and {Weare}, J.},
+  title		= {Ensemble samplers with affine invariance},
+  journal	= {Comm. App. Math. Comp. Sci.},
+  year		= 2010,
+  volume	= 5,
+  pages     = {65-80},
+}
+
+@article{     swendsen1986,
+  title = {Replica Monte Carlo Simulation of Spin-Glasses},
+  author = {Swendsen, Robert H. and Wang, Jian-Sheng},
+  journal = {Phys. Rev. Lett.},
+  volume = {57},
+  issue = {21},
+  pages = {2607--2609},
+  numpages = {0},
+  year = {1986},
+  month = {Nov},
+  publisher = {American Physical Society},
+  doi = {10.1103/PhysRevLett.57.2607},
+  url = {http://link.aps.org/doi/10.1103/PhysRevLett.57.2607}
+}
+
+@Article{goggans2004,
+  author = "Goggans, Paul M. and Chi, Ying",
+  title = "Using Thermodynamic Integration to Calculate the Posterior Probability in Bayesian Model Selection Problems",
+  journal = "AIP Conference Proceedings",
+  year = "2004",
+  volume = "707",
+  number = "1", 
+  pages = "59-66",
+  url = "http://scitation.aip.org/content/aip/proceeding/aipcp/10.1063/1.1751356",
+  doi = "http://dx.doi.org/10.1063/1.1751356" 
+}
+
+@ARTICLE{wette2013,
+   author = {{Wette}, K. and {Prix}, R.},
+    title = "{Flat parameter-space metric for all-sky searches for gravitational-wave pulsars}",
+  journal = {\prd},
+archivePrefix = "arXiv",
+   eprint = {1310.5587},
+ primaryClass = "gr-qc",
+ keywords = {Gravitational wave detectors and experiments, Gravitational radiation detectors, mass spectrometers, and other instrumentation and techniques, Mathematical procedures and computer techniques, Neutron stars},
+     year = 2013,
+    month = dec,
+   volume = 88,
+   number = 12,
+      eid = {123005},
+    pages = {123005},
+      doi = {10.1103/PhysRevD.88.123005},
+   adsurl = {http://adsabs.harvard.edu/abs/2013PhRvD..88l3005W},
+  adsnote = {Provided by the SAO/NASA Astrophysics Data System}
+}
+
+@ARTICLE{wette2015,
+   author = {{Wette}, K.},
+    title = "{Parameter-space metric for all-sky semicoherent searches for gravitational-wave pulsars}",
+  journal = {\prd},
+archivePrefix = "arXiv",
+   eprint = {1508.02372},
+ primaryClass = "gr-qc",
+ keywords = {Gravitational wave detectors and experiments, Gravitational radiation detectors, mass spectrometers, and other instrumentation and techniques, Mathematical procedures and computer techniques, Neutron stars},
+     year = 2015,
+    month = oct,
+   volume = 92,
+   number = 8,
+      eid = {082003},
+    pages = {082003},
+      doi = {10.1103/PhysRevD.92.082003},
+   adsurl = {http://adsabs.harvard.edu/abs/2015PhRvD..92h2003W},
+  adsnote = {Provided by the SAO/NASA Astrophysics Data System}
+}
+
+@ARTICLE{prix2007,
+   author = {{Prix}, R.},
+    title = "{Template-based searches for gravitational waves: efficient lattice covering of flat parameter spaces}",
+  journal = {Classical and Quantum Gravity},
+archivePrefix = "arXiv",
+   eprint = {0707.0428},
+ primaryClass = "gr-qc",
+     year = 2007,
+    month = oct,
+   volume = 24,
+    pages = {S481-S490},
+      doi = {10.1088/0264-9381/24/19/S11},
+   adsurl = {http://adsabs.harvard.edu/abs/2007CQGra..24S.481P},
+  adsnote = {Provided by the SAO/NASA Astrophysics Data System}
+}
+
+@article{prix2007metric,
+    archivePrefix = {arXiv},
+    arxivId = {gr-qc/0606088},
+    author = {Prix, Reinhard},
+    doi = {10.1103/PhysRevD.75.023004},
+    eprint = {0606088},
+    file = {:home/greg/Dropbox/Papers/Prix{\_}2007.pdf:pdf},
+    issn = {15507998},
+    journal = {Physical Review D - Particles, Fields, Gravitation and Cosmology},
+    number = {2},
+    pages = {1--20},
+    primaryClass = {gr-qc},
+    title = {{Search for continuous gravitational waves: Metric of the multidetector F-statistic}},
+    volume = {75},
+    year = {2007}
+}
+
+
+@ARTICLE{cutlershutz2005,
+   author = {{Cutler}, C. and {Schutz}, B.~F.},
+    title = "{Generalized F-statistic: Multiple detectors and multiple gravitational wave pulsars}",
+  journal = {\prd},
+   eprint = {gr-qc/0504011},
+ keywords = {Gravitational radiation detectors, mass spectrometers, and other instrumentation and techniques, Gravitational wave detectors and experiments, Mathematical procedures and computer techniques, Pulsars},
+     year = 2005,
+    month = sep,
+   volume = 72,
+   number = 6,
+      eid = {063006},
+    pages = {063006},
+      doi = {10.1103/PhysRevD.72.063006},
+   adsurl = {http://adsabs.harvard.edu/abs/2005PhRvD..72f3006C},
+  adsnote = {Provided by the SAO/NASA Astrophysics Data System}
+}
+
+@Article{	  prix2005,
+  author	= {{Prix}, R. and {Itoh}, Y.},
+  title		= "{Global parameter-space correlations of coherent searches
+		  for continuous gravitational waves}",
+  journal	= {Classical and Quantum Gravity},
+  eprint	= {gr-qc/0504006},
+  year		= 2005,
+  month		= sep,
+  volume	= 22,
+  pages		= {1003},
+  doi		= {10.1088/0264-9381/22/18/S14},
+  adsurl	= {http://adsabs.harvard.edu/abs/2005CQGra..22S1003P},
+  adsnote	= {Provided by the SAO/NASA Astrophysics Data System}
+}
+
+@misc{lalsuite,
+  title = {{LALSuite: FreeSoftware (GPL) Tools for Data-Analysis}},
+  author = {{LIGO Scientific Collaboration}},
+  year   = 2014,
+  note    = {\url{https://www.lsc-group.phys.uwm.edu/daswg/projects/lalsuite.html}}
+}
+
+@ARTICLE{allyskyS42008,
+   author = {{Abbott}, B. and {Abbott}, R. and {Adhikari}, R. and {Agresti}, J. and 
+	{Ajith}, P. and {Allen}, B. and {Amin}, R. and {Anderson}, S.~B. and 
+	{Anderson}, W.~G. and {Arain}, M. and et al.},
+    title = "{All-sky search for periodic gravitational waves in LIGO S4 data}",
+  journal = {\prd},
+archivePrefix = "arXiv",
+   eprint = {0708.3818},
+ primaryClass = "gr-qc",
+ keywords = {Gravitational wave detectors and experiments, Data analysis: algorithms and implementation, data management, Gravitational radiation detectors, mass spectrometers, and other instrumentation and techniques, Pulsars},
+     year = 2008,
+    month = jan,
+   volume = 77,
+   number = 2,
+      eid = {022001},
+    pages = {022001},
+      doi = {10.1103/PhysRevD.77.022001},
+   adsurl = {http://adsabs.harvard.edu/abs/2008PhRvD..77b2001A},
+  adsnote = {Provided by the SAO/NASA Astrophysics Data System}
+}
+
+@article{pletsch2010,
+    archivePrefix = {arXiv},
+    arxivId = {1005.0395},
+    author = {Pletsch, Holger J.},
+    doi = {10.1103/PhysRevD.82.042002},
+    eprint = {1005.0395},
+    file = {:home/greg/Dropbox/Papers/Pletsch{\_}2014.pdf:pdf},
+    issn = {15507998},
+    journal = {Physical Review D},
+    number = {4},
+    title = {{Parameter-space metric of semicoherent searches for continuous gravitational waves}},
+    volume = {82},
+    year = {2010}
+}
+
+@article{leaci2015,
+    archivePrefix = {arXiv},
+    arxivId = {1502.00914},
+    author = {Leaci, Paola and Prix, Reinhard},
+    doi = {10.1103/PhysRevD.91.102003},
+    eprint = {1502.00914},
+    file = {:home/greg/Dropbox/Papers/Leaci{\_}Prix{\_}2015.pdf:pdf},
+    issn = {15502368},
+    journal = {Physical Review D},
+    number = {10},
+    pages = {1--25},
+    title = {{Directed searches for continuous gravitational waves from binary systems: Parameter-space metrics and optimal scorpius X-1 sensitivity}},
+    volume = {91},
+    year = {2015}
+}
+
diff --git a/Paper/definitions.tex b/Paper/definitions.tex
new file mode 100644
index 0000000000000000000000000000000000000000..5bfa6393e43bb32b5b2c5375fc252b3fb4a1d5d6
--- /dev/null
+++ b/Paper/definitions.tex
@@ -0,0 +1,26 @@
+\newcommand{\F}{{\mathcal{F}}}
+\newcommand{\A}{\boldsymbol{\mathcal{A}}}
+\newcommand{\blambda}{\boldsymbol{\mathbf{\lambda}}}
+\newcommand{\blambdaSignal}{\boldsymbol{\mathbf{\lambda}}^{\rm s}}
+\newcommand{\Tspan}{T_{\rm span}}
+\newcommand{\Tcoh}{T_{\rm coh}}
+\newcommand{\tref}{t_{\rm ref}}
+\newcommand{\Nseg}{N_{\rm seg}}
+\newcommand{\Nsteps}{N_{\rm steps}}
+\newcommand{\Ntemps}{N_{\rm temps}}
+\newcommand{\Nspindown}{N_{\rm spindowns}}
+\renewcommand{\H}{\mathcal{H}}
+\newcommand{\Hs}{\H_{\rm s}}
+\newcommand{\Hn}{\H_{\rm n}}
+\newcommand{\ho}{h_0}
+\newcommand{\homax}{\ho^{\rm max}}
+\newcommand{\Bsn}{B_{\rm S/N}}
+\newcommand{\Pic}{\Pi_{\rm c}}
+\newcommand{\mutilde}{\tilde{\mu}}
+\newcommand{\Sn}{S_{\rm n}}
+\newcommand{\V}{\mathcal{V}}
+\newcommand{\Vsky}{\V_{\rm Sky}}
+\newcommand{\Vpe}{\V_{\rm PE}}
+\newcommand{\smax}{s_{\textrm{max}}}
+\newcommand{\fmax}{f_{\textrm{max}}}
+
diff --git a/Paper/fully_coherent_search_using_MCMC_walkers.png b/Paper/fully_coherent_search_using_MCMC_walkers.png
new file mode 100644
index 0000000000000000000000000000000000000000..589a99d81a18deee317b051c95df1199492729ae
Binary files /dev/null and b/Paper/fully_coherent_search_using_MCMC_walkers.png differ
diff --git a/Paper/git-tag.sh b/Paper/git-tag.sh
new file mode 100755
index 0000000000000000000000000000000000000000..e6c8b6b93e0077397d998e496a2d8cb335515b7c
--- /dev/null
+++ b/Paper/git-tag.sh
@@ -0,0 +1,29 @@
+#!/bin/bash
+
+FILE="$1"
+if [ "x$FILE" = x ]; then
+    echo "usage: $0 <file.tex>" >&2
+    exit 1
+fi
+TMPFILE="$FILE.tmp"
+
+gitLOG=`git log -1 --pretty="format:%% generated by $0
+\\newcommand{\\commitDATE}{%ai}
+\\newcommand{\\commitID}{commitID: %H}
+\\newcommand{\\commitIDshort}{commitID: %h}
+"`
+echo "$gitLOG" > $TMPFILE
+
+diffcmd="git diff -- .";
+diff=`$diffcmd`;
+if test -n "$diff"; then
+    echo "\\newcommand{\\commitSTATUS}{UNCLEAN}" >> $TMPFILE
+else
+    echo "\\newcommand{\\commitSTATUS}{CLEAN}" >> $TMPFILE
+fi
+
+if cmp -s $TMPFILE $FILE; then
+    rm -f $TMPFILE
+else
+    mv -f $TMPFILE $FILE
+fi
diff --git a/Paper/grided_frequency_search_1D.png b/Paper/grided_frequency_search_1D.png
new file mode 100644
index 0000000000000000000000000000000000000000..7df348058ea5b974603b629a9680e50ebc8893fe
Binary files /dev/null and b/Paper/grided_frequency_search_1D.png differ
diff --git a/Paper/paper_cw_mcmc.tex b/Paper/paper_cw_mcmc.tex
new file mode 100644
index 0000000000000000000000000000000000000000..d32706b7aed68a5fb8b6e391bc39910a89a81163
--- /dev/null
+++ b/Paper/paper_cw_mcmc.tex
@@ -0,0 +1,755 @@
+\documentclass[aps, prd, twocolumn, superscriptaddress, floatfix, showpacs, nofootinbib, longbibliography]{revtex4-1}
+
+% Fixes Adbobe error (131)
+\pdfminorversion=4
+
+% Ams math
+\usepackage{amsmath}
+\usepackage{amssymb}
+
+% Packages for setting figures
+\usepackage[]{graphicx}
+\usepackage{subfig}
+
+% Colors
+\usepackage{color}
+
+% Bibliography
+\usepackage{natbib}
+
+\usepackage{placeins}
+
+\usepackage{multirow}
+
+\usepackage{hhline}
+
+% get hyperlinks
+%\usepackage{hyperref}
+
+% Tables
+\newcommand{\specialcell}[2][c]{%
+  \begin{tabular}[#1]{@{}c@{}}#2\end{tabular}}
+
+\input{definitions.tex}
+
+% For editing purposes: remove before submition
+\usepackage[normalem]{ulem}	%% only added for 'strikeout' \sout
+\usepackage[usenames,dvipsnames]{xcolor}
+
+\newcommand{\dcc}{LIGO-{\color{red}}}
+
+%% ---------- editing/commenting macros: make sure all a cleared at the end! ----------
+\newcommand{\mygreen}{\color{green!50!black}}
+\newcommand{\addtext}[1]{\textcolor{green!50!black}{#1}}
+\newcommand{\meta}[1]{\addtext{#1}}
+\newcommand{\CHECK}[1]{\textcolor{red}{#1}}
+\newcommand{\strike}[1]{\textcolor{red}{\sout{#1}}}
+\newcommand{\comment}[1]{\textcolor{red}{[#1]}}
+\newcommand{\replace}[2]{\strike{#1}\addtext{$\rightarrow$ #2}}
+%% ---------- end: editing/commenting macros ----------------------------------------
+
+\begin{document}
+
+\title{MCMC follow-up methods for continuous gravitational wave candidates including
+glitching and transient waveforms}
+
+    \author{G. Ashton}
+    \email[E-mail: ]{gregory.ashton@ligo.org}
+    \affiliation{Max Planck Institut f{\"u}r Gravitationsphysik
+                 (Albert Einstein Institut) and Leibniz Universit\"at Hannover,
+                 30161 Hannover, Germany}
+    \author{R. Prix}
+    \affiliation{Max Planck Institut f{\"u}r Gravitationsphysik
+                 (Albert Einstein Institut) and Leibniz Universit\"at Hannover,
+                 30161 Hannover, Germany}
+
+\date{\today}
+
+\begin{abstract}
+We detail methods to follow-up potential CW signals (as identified by
+wide-parameter space semi-coherent searches) leverging MCMC optimisation of the
+$\mathcal{F}$-statistic. First, we demonstrate the advantages of such an
+optimisation whilst increasing the coherence time, namely the ability to
+efficiently sample an evolving distrubution and consider multiple modes.
+Subsequently, we illustrate estimation of parameters and the Bayes factor which
+can be used to understand the signficance of the candidate. Finally, we explain
+how the methods can be simply generalised to allow the waveform model to be
+transient or undergo glitches.
+
+\end{abstract}
+
+\pacs{04.80.Nn, 97.60.Jd, 04.30.Db}
+\input{git_tag.tex}
+\date{\commitDATE; \commitIDshort-\commitSTATUS, \dcc}
+
+\maketitle
+
+
+\section{Introduction}
+
+A possible target for the advanced gravitational wave detector network of LIGO
+and Virgo are long-lived periodic sources called continuous waves (CWs).
+Rapidly rotating nonaxisymmetric neutron stars are potentially capable of
+producing detectable CWs which last much longer than typical observation spans.
+There exists three well known sources of the nonaxisymmetry: `mountains',
+precession, and r-mode oscillations; each of which make a prediction for the
+scaling between $\nu$, the NS spin frequency and $f$, the gravitational wave
+frequency. In any case, observing neutron stars through their gravitational
+wave emmission would provide a unique astrophysical insight and has hence
+motivated numerous searches.
+
+As shown by \citet{jks1998}, the gravitational wave signal from a
+nonaxisymmetric source produces a strain in the detector $h(t, \A, \blambda)$;
+where $\A{=}\{h_0, \cos\iota, \psi, \phi_0\}$ is a vector of the four
+\emph{amplitude-parameters} (expressed in `physical coordinates') and
+$\blambda$ is a vector of the \emph{Doppler-parameters} consisting of the
+sky-location, frequency $f$, and any spindown terms required by the search.
+
+CW searches typically use a fully-coherent matched-filtering methods whereby a
+template (the signal model at some specific set of parameters) is convolved
+against the data resulting in a detection statistic. Since the signal
+parameters are unknown, it is usual to perform this matched filtering over a
+grid of points.  Three search categories can be identified: \emph{targeted}
+searches for a signal from a known electromagnetic pulsar where the Doppler
+parameters are considered `known'; \emph{directed} searches in which the
+location is known, but not the frequency and spin-down (i.e.  searching for the
+neutron star in a supernova remnant which does not have a known pulsar); and
+\emph{all-sky} searches where none of the parameters are considered known.
+Searching over more parameters amounts to an increase in the dimension of the
+search space. Since the density of grid points required to resolve a signal
+scales inversely with total coherence time $\Tcoh$ (the span of data used in
+a fully-coherent matched filter), wide-parameter searches (such as the all-sky)
+with many search dimensions over long durations of data are computationally
+demanding.
+
+At a fixed computing cost, it has been shown (see for example \citep{brady1998,
+prix2012}) that a semi-coherent search is more sensitive for unknown signals
+than a fully-coherent search. While the exact details of how the search works
+depends on the implementation, semi-coherent search work by splitting the total
+observation span $\Tspan$ into $\Nseg$ segments (each lasting for $\Tcoh$) and
+in each segment computes the fully-coherent detection statistic; the
+semi-coherent detection statistic is then computed by some combination of all
+segments summed at the same point in parameter space. Fundamentally, this gain
+in sensitivity is because the width of a peak in the detection statistic due to
+a signal is inversely propotional to the cohrence time: shorter coherence times
+make the peak wider and hence the a lower density of templates. This idea was
+first proposed by \citet{brady2000} along with the first implementation, the
+`Stack-slide' search. Since then, several modifications such as the
+`Hough-transform' \citep{krishnan2004, astone2014}, and the `Powerflux' method
+(first described in \citet{allyskyS42008}) have been proposed, implemented and
+applied to gravitational wave searches.
+
+Wide parameter space searches produce a list of candidates with an associated
+detection statistic which passes some threshold. In order to verify these
+candidates, they are subjected to a \emph{followed-up}: a process of increasing
+the coherence time, eventually aiming to calculate a fully-coherent detection
+statistic over the maximal span of data. In essense, the semi-coherent search
+is powerful as it spreads the significance of a candidate over a wider area of
+parameter space, so a follow-up attempts to reverse this process and recover
+the maximum significance and tightly constrain the candidate parameters. The
+original hierarchical follow-up of \citet{brady2000} proposed a two-stage method
+(an initial semi-coherent  stage followed directly by a fully-soherent search.
+However, it was shown in a numerical study by \citet{cutler2005} that allowing
+an arbitrary number of semi-coherent stages before the final fully-coherent
+stage can significantly improve the efficiency: ultimately they concluded that
+three semi-coherent stages provide the best trade-off between sensitivity and
+computational cost.
+
+The first implementation of a two-stage follow-up was given by
+\citet{shaltev2013} and used the Mesh Adaptive Direct Search algorithm for
+optimisation. At the time of publication, this method was limited to two stages
+and could not handle binary parameters \comment{(I think these limitations have
+now been removed, but I can't find a publication)}, however these are practical
+limitations which can \comment{(have?)} be overcome.
+\comment{Add something on multiple modes?}
+
+In this paper, we propose an alternative hierarchical follow-up procudure using
+Markov-Chain Monte-Carlo (MCMCM) as the optimisation tool. In terms of the
+semi-coherent to follow-up procedure, an MCMC tool is advantages due to it's
+ability to trace the evolution multiple modes simultaneuosly through the
+follow-up procudedure and allow the optimisation to decide between them without
+arbitrary cuts. In addition, MCMC methods also provide two further
+advatanges: they can calculate directly calculate Bayes factors, the significance
+of a candidate and because they are `gridless' one can arbitrarily vary the
+waveform model without requring an understanding of the underlying topology.
+We will exploit this latter property to propose an additional step in the
+follow-up procudure which allows for the CW signal to be either a transient-CW
+(a periodic signal lasting $\mathcal{O}(\textrm{hours-weeks})$) or to undergo
+glitches (as seen in pulsars).
+
+We begin in Section~\ref{sec_hypothesis_testing} with a review of search
+methods from a Bayesian perspective. Then in
+Section~\ref{sec_MCMC_and_the_F_statistic} we introduce the MCMC optimisation
+producedure and give details of our particular implementation. In
+Section~\ref{sec_follow_up} we will illustrate applications of the method and
+provide a prescription for choosing the setup. In Sections~\ref{sec_transients}
+and \ref{sec_glitches} we demonstrate how searches can be performed for either
+transient or glitches CWs before we finally conclude in
+Section~\ref{sec_conclusion}.
+
+\section{Hypothesis testing}
+\label{sec_hypothesis_testing}
+
+\subsection{Bayes factors}
+Given some data $x$ and a set of background assumptions $I$, we formulate
+two hypotheses: $\Hn$, the data contains solely Gaussian noise and $\Hs$, the
+data contains an additive mixture of noise and a signal $h(t; \A, \blambda)$.
+In order to make a quantitative comparison, we use Bayes theorum in the usual
+way to write the odds as
+\begin{equation}
+O_{\rm S/N} \equiv \frac{P(\Hs| x I)}{P(\Hn| x I)} =
+\Bsn(x| I) \frac{P(\Hs| I)}{P(\Hn | I)},
+\end{equation}
+where the second factor is the prior odds while the first factor is the
+\emph{Bayes factor}:
+\begin{equation}
+\Bsn(x| I) = \frac{P(x| \Hs I)}{P(x| \Hn I)}.
+\end{equation}
+Typically, we set the prior odds to unity such that it is the Bayes factor
+which determines our confidence in the signal hypothesis. In this work we will
+therefore discuss the Bayes factor with the impplied assumption this is
+equivalent to the odds, unless we have a good reason to change the prior odds.
+
+We can rewrite this Bayes factor in terms of the two sets of signal parameters
+as
+\begin{equation}
+\Bsn(x| I) = \frac{P(x, \A, \blambda|\Hs, I)}
+{P(\A| \Hs, I)P(\blambda| \Hs, I)P(x| \Hn, I)}.
+\end{equation}
+Marginalising over the two sets of parameters we find that
+\begin{equation}
+\Bsn(x| I)= \iint
+\mathcal{L}(x; \A, \blambda)
+P(\A| \Hs, I)P(\blambda| \Hs, I)
+d\blambda d\A
+\label{eqn_full_bayes}
+\end{equation}
+where
+\begin{equation}
+\mathcal{L}(x; \A, \blambda) \equiv \frac{P(x |\A, \blambda, \Hs, I)}{P(x| \Hn, I)},
+\label{eqn_likelihood}
+\end{equation}
+is the \emph{likelihood-ratio}.
+
+At this point, we can appreciate the problems of searching for unknown signals:
+one has four amplitude parameters and several doppler parameters (three plus
+the number of spin-down and binary parameters) over which this integral must be
+performed. If a single signal exists in the data, this corresponds to a single
+peak in the likelihood-ratio, but at an unknown location. Therefore, one must
+must first search for peaks (candidates), and then subsequently analyse their
+significance. If one has prior knowledge this can be used; for example in
+targeted searches for gravitational waves from known pulsars emitting CWs from
+a mountain the Doppler parameters are considered known collapsing their
+integral to a single evaluation.
+
+\subsection{The $\F$-statistic}
+
+For directed and all-sky searches, a common method introduced by
+\citet{jks1998} to reduce the parameter space is the maximum-likelihood
+approach. In this approach (often referred to as `frequentist'), one starts by
+defining the likelihood-ratio, Equation~\eqref{eqn_likelihood}, which in this
+context is a \emph{matched-filtering} amplitude. Then, analytically maximising
+this likelihood-ratio with respect to the four amplitude parameters results
+(c.f.~\citet{prix2009}) in a maximised log-likelihood given by $\F(x|
+\blambda)$: the so-called $\F$-statistic. Picking a particular set of Doppler
+parameters $\blambda$ (the template) one can then compute a detection statistic
+(typicaly $2\F$ is used) which can be used to quantify the significance of the
+template. Usually this is done by calculating a corresponding false alarm rate,
+the probability of seeing such a detection statistic in Gaussian noise.
+
+Calculations of the significance are possible due to the properties of the
+detection statistic: in Gaussian noise, it can be shown \citep{jks1998,
+cutlershutz2005} that $\widetilde{2\F}$ follows a chi-squared distribution with
+4 degrees of freedom. In the presence of a signal, $\widetilde{2\F}$ is
+still chi-squared with 4 degrees of freedom, but has a non-centrality parameter
+$\tilde{\rho}^{2}$ such that its expectation value is
+\begin{equation}
+\textrm{E}[\widetilde{2\F}(x; \blambda)] = 4 + \tilde{\rho}(x; \blambda)^2.
+\label{eqn_twoF_expectation}
+\end{equation}
+The non-centrality parameter in this context is the SNR of the matched-filter
+given by
+\begin{equation}
+\rho^{2} = (h | h) \propto \frac{h_0^2}{\Sn}\Tcoh \mathcal{N}
+\end{equation}
+where $(h|h)$ is the inner product of the signal with itself (see for example
+\citet{prix2009}), $\Sn$ is a (frequency-dependent) measure of the noise in
+the detector and $\mathcal{N}$ is the number of detectors.
+
+
+\subsection{Using the $\F$-statistic to compute a Bayes factor}
+At first, it appeared that the $\F$-statistic was independent of the Bayesian
+framework. However, it was shown by \citet{prix2009} that if we marginalise
+over the four amplitude parameters of Equation~\eqref{eqn_full_bayes}, choosing
+a prior $\Pi_{\rm c}$ such that
+\begin{equation}
+P(\A| \Hs, \Pi_{\rm c}, I) \equiv \left\{
+\begin{array}{ll}
+C & \textrm{ for } \ho < \homax \\
+0 & \textrm{ otherwise}
+\end{array}
+\right..
+\end{equation}
+then the integral, when $\homax \gg 1$, is a Gaussian integral and can be
+computed analytically as
+\begin{align}
+B_{\rm S/N}(x| \Pi_{\rm c}, \blambda) & \equiv
+\int
+\mathcal{L}(x ;\A, \blambda)
+P(\A| \Hs, I) d\A
+\\
+& = \frac{C (2\pi)^{2} e^{\F(x| \blambda)}}
+{\sqrt{\textrm{det} \mathcal{M}}},
+\end{align}
+where $C$ is a normalisation constant, $\textrm{det}\mathcal{M}$ is an antenna
+pattern factor dependent on the sky-position and observation period, and
+$\mathcal{F}$ is the frequentist log-likelihood of \citet{jks1998}. This result
+demonstrates that the $\mathcal{F}$-statistic is proportional to the log-Bayes
+factors when calculated with a uniform prior on the amplitude parameters and
+fixed Doppler parameters.
+
+As such, we can define the Bayes-factor of Equation~\eqref{eqn_full_bayes} as
+\begin{equation}
+B_{\rm S/N}(x| \Pi_{\rm c}, I) = \int
+B_{\rm S/N}(x| \Pi_{\rm c}, \blambda) P(\blambda| \Hs, I)
+ d\blambda,
+\end{equation}
+or neglecting the constants
+\begin{equation}
+B_{\rm S/N}(x| \Pi_{\rm c}, I) \propto \int
+e^{\F(x| \blambda)} P(\blambda| \Hs, I)
+ d\blambda.
+\label{eqn_bayes_over_F}
+\end{equation}
+
+Formulating the significance of a CW candidate in this way is pragmatic in that
+there exists a wealth of well-tested tools \citep{lalsuite} capable of
+computing the $\mathcal{F}$-statistic for CW signals, transient-CWs, and CW
+signals from binary systems; these can be levereged to compute
+Equation~\eqref{eqn_bayes_over_F}, or adding in the constant
+$B_{\rm S/N}(x| \Pi_{\rm c})$ itself. The disadvantage to this method is that
+we are forced to use the prior $\Pic$, which was shown by \citet{prix2009} to
+be unphysical.
+
+\section{MCMC and the $\mathcal{F}$-statistic}
+\label{sec_MCMC_and_the_F_statistic}
+
+The MCMC class of optimisation tools are formulated to solve the problem of
+infering the posterior distribution of some general model parameters $\theta$
+given given some data $x$ for some hypothesis $\H$. Namely, Bayes rule
+\begin{equation}
+P(\theta| x, \H, I) \propto P(x| \theta, \H, I)P(\theta| \H, I),
+\label{eqn_bayes_for_theta}
+\end{equation}
+is used to evaluate proposed jumps from one point in parameter to other points;
+jumps which increase this probabily are accepted with some probability. The
+algorithm, proceeding in this way, is highly effective at resolving peaks in
+the high-dimension parameter spaces.
+
+At this point, we note the equivalence of Equation~\eqref{eqn_bayes_for_theta}
+to the integrand of Equation~\eqref{eqn_bayes_over_F}:
+\begin{equation}
+P(\blambda | x, \Pi_{\rm c}, \Hs, I)
+%=B_{\rm S/N}(x| \Pi_{\rm c}, \blambda) P(\blambda| \Hs I).
+\propto e^{\F(x| \blambda)} P(\blambda| \Hs I),
+\label{eqn_lambda_posterior}
+\end{equation}
+where $e^{\F}$ is the likelihood.
+In this work, we will focus on using MCMC methods to sample this, the
+posterior distribution of the Doppler parameters and moreover compute the
+final Bayes factor.
+
+\subsection{The \texttt{emcee} sampler}
+
+In this work we will use the \texttt{emcee} ensemble sampler
+\citep{foreman-mackay2013}, an implementation of the affine-invariant ensemble
+sampler proposed by \citet{goodman2010}. This choice addresses a key issue with
+the use of MCMC sampler, namely the choice of \emph{proposal distribution}. At
+each step of the MCMC algorithm, the sampler generates from some distribution
+(known as the proposal-distribution) a jump in parameter space. Usualy, this
+proposal distribution must be `tuned' so that the MCMC sampler effeciently
+walks the parameter space without either jumping too far off the peak, or
+taking such small steps that it takes a long period of time to traverse the
+peak. The \texttt{emcee} sampler addresses this by using an ensemble, a large
+number ${\sim}100$ parallel `walkers', in which the proposal for each walker
+is generated from the current distribution of the other walkers. Moreover, by
+applying an an affine transformation, the efficiency of the algorithm is not
+diminished when the parameter space is highly anisotropic. As such, this
+sampler requires little in the way of tuning: a single proposal scale and the
+Number of steps to take.
+
+Beyond the standard ensemble sampler, we will often use one further
+modification, namely the parallel-tempered ensemble sampler. A parallel
+tempered MCMC simulation, first proposed by \citet{swendsen1986}, runs
+$\Ntemps$ simulations in parallel with the likelihood in the $i^{\rm th}$
+parallel simulation is raied to a power of $1/T_{i}$ (where $T_i$ is referred
+to as the temperature) such that Equation~\eqref{eqn_lambda_posterior} becomes
+\begin{equation}
+P(\blambda | T_i, x, \Pi_{\rm c}, \Hs, I)
+%=B_{\rm S/N}(x| \Pi_{\rm c}, \blambda)^{T_i} P(\blambda| \Hs I).
+\propto (e^{\F(x| \blambda)})^{T_i} P(\blambda| \Hs I).
+\end{equation}
+Setting $T_0=1$ with $T_i > T_0 \forall i > 1$, such that the lowest
+temperature recovers Equation~\eqref{eqn_lambda_posterior} while for higher
+temperatures the likelihood is broadened (for a Gaussian likelihood, the
+standard devitation is larger by a factor of $\sqrt{T_i}$). Periodically, the
+different tempereates swap elements. This allows the $T_0$ chain (from which we
+draw samples of the posterior) to efficiently sample from multi-modal
+posteriors. This does however introduce two additional tuning parameters, the
+number and range of the set of temperatures $\{T_i\}$.
+
+\subsection{Parallel tempering}
+In addition, parallel-tempering also offers a robust method to estimate the
+Bayes factor of Equation~\eqref{eqn_bayes_over_F}. If we define
+$\beta\equiv1/T$, the inverse temperature and $Z(\beta)\equiv B_{\rm S/N}(x| \Pi_{\rm
+c}, I)$, then as noted by \citet{goggans2004} for the general case, we may
+write
+\begin{align}
+\frac{1}{Z} \frac{\partial Z}{\partial \beta}=
+\frac{
+\int B_{\rm S/N}(x| \Pi_{\rm c}, \blambda)^{\beta}
+\log(B_{\rm S/N}(x| \Pi_{\rm c}, \blambda))P(\blambda| I)
+}
+{
+\int B_{\rm S/N}(x| \Pi_{\rm c}, \blambda)^{\beta})P(\blambda| I)
+}
+\end{align}
+The right-hand-side expresses the average of the log-likelihood at $\beta$. As
+such, we have
+\begin{align}
+\frac{\partial \log Z}{\partial \beta} = 
+\langle \log(\Bsn(x| \Pic, \blambda) \rangle_{\beta}
+\end{align}
+The log-likelihood are a calculated during the MCMC sampling. As such, one
+can numerically integrate to get the Bayes factor, i.e.
+\begin{align}
+\log \Bsn(x| \Pic, I) = \log Z = \int_{0}^{1}
+\langle \log(\Bsn(x| \Pic, \blambda) \rangle_{\beta} d\beta.
+\end{align}
+In practise, we use a simple numerical quadrature over a finite ladder of
+$\beta_i$ with the smallest chosen such that extending closer to zero does not
+change the result beyond other numerical uncertainties. Typically, getting
+accurate results for the Bayes factor requires a substantially larger number of
+temperatures than are required for effeciently sampling multi-modal
+distributions.  Therefore, it is recomended that one uses a small number of
+temperatures during the search stage, and subsequently a larger number of
+temperatures (suitably initialised close to the target peak) when estimating
+the Bayes factor.
+
+\subsection{The topology of the likelihood}
+As discussed, we intend to use the $\F$-statistic as our log-likelihood in the
+MCMC simulations. Before continuing, it is worthwhile to understand the behaviour
+of the log-likelihood. As shown in Equation~\eqref{eqn_twoF_expectation},
+$\widetilde{2\F}$ has a expectation value of 4 (corresponding to the
+4 degrees of freedom of the underlying chi-square distribution) in Gaussian
+noise, but in the presence of a signal larger value are expected proportional
+to the squared SNR. 
+
+To illustrate this, let us consider $\widetilde{2\F}$ (the log-likelihood)
+as a function of $f$ (the template frequency) if there exists a signal in the
+data with frequency $f_0$. We will assume that all other Doppler parameters
+are perfectly matched.
+This can be calculated analytically, taking the matched-filtering amplitude
+(Equation~(11) of \citep{prix2005}) with $\Delta\Phi(t) = 2\pi(f - f_0) t$
+the expectation of $\widetilde{2\F}$ as a function of the template frequency
+$f$ is given by
+\begin{equation}
+\textrm{E}[\widetilde{2\F}](f) = 4 + 
+(\textrm{E}[\widetilde{2\F_0}] -4)\textrm{sinc}^{2}(\pi(f-f_0)\Tcoh))
+\label{eqn_grid_prediction}
+\end{equation}
+where $\textrm{E}[\widetilde{2\F_0}]$ is the expected $\widetilde{2\F}$ for
+a perfectly matched signal (when $f=f_0$).
+
+In Figure~\ref{fig_grid_frequency} we compare the analytic prediction of
+Equation~\eqref{eqn_grid_prediction} with the value computed numerically
+from simulating a signal in Gaussian noise. As expected, close to the signal
+frequency $f_0$ the detection statistic peaks with a a few local secondary
+maxima. Away from this frequency, in Gaussian noise, we see many local maxima
+centered around the expected value of 4.
+\begin{figure}[htb]
+\centering \includegraphics[width=0.45\textwidth]{grided_frequency_search_1D}
+\caption{Comparison of the analytic prediction of
+Equation~\eqref{eqn_grid_prediction} (in red) with the value computed
+numerically from simulating a signal in Gaussian noise (in black).
+\comment{Need to explain ticks}}
+\label{fig_grid_frequency}
+\end{figure}
+
+\subsection{Limitations of use}
+
+In general, MCMC samplers are highly effective in generating samples of the
+posterior in multi-dimensional parameter spaces. However, they will perform
+poorly if the posterior has multiple small maxima with only a small number of
+large maxima which occupy a small fraction of the prior volume. Since we will
+use $\F$ as our log-likelihood, Figure~\ref{fig_grid_frequency} provides an
+example of the space we will ask the sampler to explore. Clearly, if the width
+of the signal peak is small compared to the prior volume, the sampler will get
+`stuck' on the local maxima and be ineffecient at finding the global maxima.
+This problem is excabated in higher-dimensional search spaces where the volume
+fraction of the signal scales with the exponent of the number of dimensions.
+
+In a traditional CW search which uses a grid of templates (also known as a
+template bank), the spacings of the grid are chosen such that the loss of
+signal to noise ratio (SNR) is bounded to less than $u$, the template-bank
+mismatch. The optimal choice of grid then consists of minimising the computing
+cost while respecting this minimum template-bank mismatch or vice-verse (for
+examples see \citet{pletsch2010, prix2012, wette2013, wette2015}). We will now
+discuss how the work on setting up these grids can be applied to the problem of
+determining whether the setup is appropriate for an MCMC method: i.e. given the
+prior volume do we expect a signal to occupy a non-negligible volume?
+
+For a fully-coherent $\F$-statistic search on data containing Gaussian noise
+and a signal with Doppler parameters $\blambdaSignal$, the template-bank
+mismatch at the grid point $\blambda_{l}$ is defined to be
+\begin{align}
+\mutilde(\blambdaSignal, \blambda_{l}) \equiv 1 - 
+\frac{\tilde{\rho}(\blambda_l;\blambdaSignal)^{2}}
+{\tilde{\rho}(\blambdaSignal; \blambdaSignal)^{2}},
+\end{align}
+where $\tilde{\rho}(\blambda_l; \blambdaSignal)$ is the non-centrality
+parameter (c.f. Equation~\ref{eqn_twoF_expectation}) at $\blambda_l$, given
+that the signal is at $\blambdaSignal$. As such
+$\widetilde{\textrm{SNR}}(\blambdaSignal; \blambdaSignal)$ is the
+perfectly-matched non-centrality parameter, for which the mismatch is zero.
+For a fully-coherent search, this non-centrality parameter is equivalent to
+fully-coherent matched-filter signal to noise ratio SNR. However,
+as noted by \citet{leaci2015}, this is true for the fully-coherent case only.
+Therefore, we will use the non-centrality parameter which easily generalised to
+the semi-coherent case.
+
+To make analytic calculations of the mismatch possible, as first shown by
+\citet{brady1998}, the mismatch can be approximated by
+\begin{equation}
+\mutilde(\blambda, \Delta\blambda) \approx 
+\tilde{g}_{\alpha \beta}^{\phi} \Delta\lambda^{\alpha}\Delta\lambda^{\beta}
++ \mathcal{O}\left(\Delta\blambda^{3}\right)
+\end{equation}
+where we switch to using index notation for which we sum over repeated indices.
+Here, $\tilde{g}_{\alpha\beta}^{\phi}$ is the `phase-metric' given by
+\begin{align}
+\tilde{g}^{\phi}_{\alpha \beta}(\blambda) = 
+\langle 
+\partial_{\Delta\lambda^{\alpha}}\phi
+\partial_{\Delta\lambda^{\beta}}\phi
+\rangle
+-
+\langle 
+\partial_{\Delta\lambda^{\alpha}}\phi
+\rangle
+\langle
+\partial_{\Delta\lambda^{\beta}}\phi
+\rangle,
+\label{eqn_metric}
+\end{align}
+where $\langle \cdot \rangle$ denotes the time-average over $\Tcoh$ and
+$\phi(t; \blambda)$ is the phase evolution of the source. The phase metric is
+in fact an approximation of the full metric which includes modulations of the
+amplitude parameters $\A$; it was shown by \citet{prix2007metric} that it is a
+good approximation when using data spans longer than a day and data from
+multiple detectors. 
+
+The phase metric, Equation~\eqref{eqn_metric} provides the neccesery tool to
+measure distances in the Doppler parameter space in units of mismatch. To
+calculate it's components, we define the phase evolution
+of the source as \citep{wette2015}
+\begin{align}
+\phi(t; \blambda) \approx 2\pi\left(
+\sum_{s=0}^{\smax} f^{(s)}\frac{(t-\tref)^{s+1}}{(s+1)!}
++ \frac{r(t)\cdot\mathbf{n}}{c} \fmax\right),
+\label{eqn_phi}
+\end{align}
+where $\mathbf{n}(\alpha, \delta)$ is the fixed position of the source with
+respect to the solar system barycenter (with coordinates $\alpha, \delta$ the
+right ascension and declination), $f^(s)\equiv d^{s}\phi/dt^s$, and $\fmax$
+a constant chosen conservatively to be the maximum frequency over the data
+span.
+
+The frequency and spin-down components of the metric can be easily calculated
+due to their linearity in Equation~\eqref{eqn_phi} and for the special case in
+which $\tref$ is in the middle of the data span, the frequency and spin-down
+parts of the metric are diagonal. Accurately approximating the sky components
+of the metric is non-trivial, but was accomplished by \citet{wette2013} for the
+fully-coherent case. In \citet{wette2015} it was shown how the calculate the
+equivalent semi-coherent metric $\hat{g}_{\alpha\beta}^{\phi}(\blambda,
+\Nseg)$.  In the following, we will work with this calculation with the
+understanding that $\hat{g}_{\alpha\beta}^{\phi}(\blambda, \Nseg{=}1)=
+\tilde{g}_{\alpha\beta}^{\phi}(\blambda)$.
+
+To understand the volume of parameter space which a true signal would occupy,
+we can make use of the \emph{metric-volume} \citep{prix2007}, given by
+\begin{align}
+\mathcal{V} = \int 
+\sqrt{\textrm{det}\hat{g}^{\phi}_{\alpha\beta}(\blambda, \Nseg)} d\blambda \approx 
+\sqrt{\textrm{det}\hat{g}^{\phi}_{\alpha\beta}(\blambda, \Nseg)} \Delta\blambda
+\end{align}
+where in the second step we assume a constant coefficient metric. Here, $\Delta
+\blambda$ is the volume element which is given by 
+\begin{equation}
+\Delta\lambda = \frac{\Delta\Omega}{2}
+%\frac{1}{2}\sin\delta\Delta\delta\Delta\alpha
+\prod_{s=0}^{\smax} \Delta f^{(s)},
+\end{equation}
+where $\Delta\Omega$ is the solid angle of the sky-patch which is searched,
+$\Delta f^(s)$ is the extend of the frequency and spin-down range(s) searched,
+and the factor of $1/2$ comes from converting the normalised determinant which
+is computed over the whole sky to the solid angle of the directed search.
+\comment{Not sure I fully understand this yet, or have really derived it properly}.
+
+The metric volume $\V$ is the approximate number of templates required to cover
+the the given Doppler parameter volume at a fixed mismatch of $\approx 1$. As
+such, its inverse gives the approximate (order of magnitude) volume fraction of
+the search volume which would be occupied by a signal. This can therefore be
+used as a proxy for determing if an MCMC search will operate in a regime where
+it is effecicient (i.e. where the a signal occupes a reasonable fraction of the
+search volume).
+
+The volume $\V$ combines the search volume from all search dimensions. However,
+let us know discuss how we can delve deeper to understand how each dimension
+contributes to the total product. This is done by noticing that the metric has
+a particular block form:
+\begin{align}
+g_{ij} = \left[
+\begin{array}{cc}
+g^{\rm Sky} & 0 \\
+0 & g^{\rm PE}
+\end{array}
+\right]
+\end{align}
+where $g^{\rm Sky}$ is the $2\times2$ sky-metric, while $g^{\rm PE}$ is the
+$(\smax{+}1){\times}(\smax{+}1)$ phase-evolution metric.
+As such, the volume can be decomposed as
+\begin{align}
+\mathcal{V} & =
+\sqrt{\textrm{det}g^{\rm Sky}}\frac{\Delta\Omega}{2} \times
+\sqrt{\textrm{det}g^{\rm PE}}\prod_{s=0}^{\smax}\Delta f^{(s)} \\
+& = \Vsky \times \Vpe.
+\end{align}
+Moreover, if $\tref$ is in the middle of the observation span, the diagonal
+nature of $g^{\rm PE}$ means that one can further identify
+\begin{align}
+\Vpe = \prod_{s=0}^{\smax}\sqrt{g^{\rm PE}_{ss}} \Delta f^{(s)}
+= \prod_{s=0}^{\smax}\Vpe^{(s)}
+\end{align}
+This decomposition may be useful in setting up MCMC searches.
+
+\subsection{An example}
+
+In order to familiarise the reader with the features of the search, we will now
+describe a simple directed search (over $f$ and $\dot{f}$) for a strong
+simulated signal in Gaussian noise. The setup of the search consists in defining
+the following
+\begin{itemize}
+\item The prior for each search parameter. Typically, we recomend either a uniform
+prior bounding the area of interest, or a normal distribution centered on the
+target and with some well defined width. In this example we will use a uniform
+prior.
+\item The initialisation of the walkers. If the whole prior volume is to be explored,
+the walkers should be initialised from the prior (i.e. random drawn from the
+prior distributions) as we will do here. However, it is possible that only a
+small region of parameter space requires exploration, therefore we provide
+functionality to initialise the walkers subject to an independent distribution
+if needed.
+\item The number of burn-in and production steps to take. This is a tuning
+parameter of the MCMC algorithm. First we allow the walkers to run for a number
+of `burn-in' steps which are discarded and then a number of `production' steps
+are taken from which one makes estimates of the posterior
+\item The parallel tempering set-up. If used, one must specify the number of
+temperatures and their arrangement. Typically, we use 3 or so temperatures
+with arranged linearly in log-space from some zero to some maximum temperature.
+\end{itemize}
+
+Using these choices, the simulation is run. To illustrate the full MCMC process,
+in Figure~\ref{fig_MCMC_simple_example} we plot the progress of all the individual
+walkers (each represented by an individual line) as a function of the total
+number of steps. The red portion of steps are `burn-in' and hence discarded,
+from this plot we see why: the walkers are initialised from the uniform prior
+and initially spend some time exploring the whole parameter space before congerging.
+The production samples, colored black, are only taken once the sampler has
+converged - these can be used to generate posterior plots.
+
+\begin{figure}[htb]
+\centering
+\includegraphics[width=0.5\textwidth]{fully_coherent_search_using_MCMC_walkers}
+\caption{}
+\label{fig:}
+\end{figure}
+
+\section{Follow-up}
+\label{sec_follow_up}
+
+Incoherent detection statistics trade significance (the height of the peak) for
+sensitivity (the width of the peak). We will now discuss the advantages of
+using an MCMC sampler to follow-up a candidate found incoherently, increasing
+the coherence time until finally estimating it's parameters and significance
+fully-coherently. We begin by rewritting Equation~\eqref{eqn_lambda_posterior},
+the posterior distribution of the Doppler parameters, with the explicit
+dependence on the coherence time $\Tcoh$:
+\begin{equation}
+P(\blambda | \Tcoh, x, \Pi_{\rm c}, \Hs, I)
+%=B_{\rm S/N}(x| \Tcoh, \Pi_{\rm c}, \blambda) P(\blambda| \Hs I).
+\propto e^{\hat{\F}(x| \Tcoh, \blambda)} P(\blambda| \Hs I).
+\end{equation}
+
+Introducing the coherent time $\Tcoh$ as a variable provides an ability to
+adjust the likelihood. Therefore, a natural way to perform a follow-up is to
+start the MCMC simulations with a short coherence time (such that the signal
+peak occupies a substantial fraction of the prior volume). Subsequently,
+incrementally increase this coherence time in a controlled manner, aiming to
+allow the MCMC walkers to converge to the new likelihood before again
+increasing the coherence time. This can be considered analogous to simulated
+annealing (where the likelihood is raised to a power $1/T$ and subseuqntly
+`cooled') with the important difference that the semi-coherent likelihood is
+wider at short coherence times (rather than flatter as in the case of
+high-temperature simulated annealing stages).
+
+To illustrate the utility of this method, in Figure~\ref{fig_follow_up} we show
+the progress of the MCMC sampler during such a follow-up. The data, 100 days
+from a single detector, consists of Gaussian noise with
+$\sqrt{\Sn}=10^{-23}$~Hz$^{-1/2}$ (at the fiducial frequency of the signal) and
+a signal. The signal has an amplitude $h_0=1.4\times10^{25}$ such that the
+signal has a depth of $\sqrt{\Sn}/h_0=70$ in the noise. The search setup is
+outlined in Table~\ref{tab_weak_signal_follow_up}.
+
+\begin{table}[htb]
+\caption{The search setup used in Figure~\ref{fig_follow_up}}
+\label{tab_weak_signal_follow_up}
+\input{weak_signal_follow_up_run_setup}
+\end{table}
+
+\begin{figure}[htb]
+\centering
+\includegraphics[width=0.5\textwidth]{weak_signal_follow_up_walkers}
+
+\caption{In the top three panels we show the progress of the 500 parallel
+walkers (each of which is an individual thin line) during the MCMC simulation
+for each of the search parameters, frequency $f$, right-ascension $\alpha$ and
+declination $\delta$. Each vertical dashed line indicates the start of a new
+stage of the search, the parameters for all stages are listed in
+Table~\ref{tab_weak_signal_follow_up}. Samples for use in estimating
+posteriors, or other processes are taken from those samples colored black,
+which we call the production samples.  The period for which the lines are
+coloured red, the samples are discarded either because they are taken from the
+posterior when $\Tcoh < \Tspan$, or they are from the burn-in of the final
+stage. In the final panel we plot the estimated distribution of
+$\widetilde{2\F}$ taken from the production samples.}
+
+\label{fig_follow_up}
+\end{figure}
+
+\section{Alternative waveform models: transients}
+\label{sec_transients}
+
+\section{Alternative waveform models: glitches}
+\label{sec_glitches}
+
+\section{Conclusion}
+\label{sec_conclusion}
+
+
+
+\section{Acknowledgements}
+
+\bibliography{bibliography}
+
+\end{document}
diff --git a/Paper/weak_signal_follow_up_run_setup.tex b/Paper/weak_signal_follow_up_run_setup.tex
new file mode 100644
index 0000000000000000000000000000000000000000..9039eace0a5e1596093df2d92bf8c9fce2189478
--- /dev/null
+++ b/Paper/weak_signal_follow_up_run_setup.tex
@@ -0,0 +1,9 @@
+\begin{tabular}{c|cccccc}
+Stage & $\Nseg$ & $\Tcoh^{\rm days}$ &$\Nsteps$ & $\V$ & $\Vsky$ & $\Vpe$ \\ \hline
+0 & 80 & 1.25 & 100 & 20.0 & 2.0 & 10.0 \\
+1 & 40 & 2.5 & 100 & $2{\times}10^{2}$ & 7.0 & 20.0 \\
+2 & 20 & 5.0 & 100 & $1{\times}10^{3}$ & 30.0 & 50.0 \\
+3 & 10 & 10.0 & 100 & $1{\times}10^{4}$ & $1{\times}10^{2}$ & 90.0 \\
+4 & 5 & 20.0 & 100 & $7{\times}10^{4}$ & $4{\times}10^{2}$ & $2{\times}10^{2}$ \\
+5 & 1 & 100.0 & 100,100 & $1{\times}10^{6}$ & $1{\times}10^{3}$ & $9{\times}10^{2}$ \\
+\end{tabular}
diff --git a/Paper/weak_signal_follow_up_walkers.png b/Paper/weak_signal_follow_up_walkers.png
new file mode 100644
index 0000000000000000000000000000000000000000..96097988a61fd8e354d2fbac5665eccbc02ec074
Binary files /dev/null and b/Paper/weak_signal_follow_up_walkers.png differ