ViewVC Help
View File | Revision Log | Show Annotations | View Changeset | Root Listing
root/group/trunk/mattDisertation/Introduction.tex
(Generate patch)

Comparing trunk/mattDisertation/Introduction.tex (file contents):
Revision 1008 by mmeineke, Tue Feb 3 17:41:56 2004 UTC vs.
Revision 1106 by mmeineke, Tue Apr 13 21:13:53 2004 UTC

# Line 1 | Line 1
1  
2 + \chapter{\label{chapt:intro}INTRODUCTION AND THEORETICAL BACKGROUND}
3  
3 \chapter{\label{chapt:intro}Introduction and Theoretical Background}
4  
5
5   The techniques used in the course of this research fall under the two
6   main classes of molecular simulation: Molecular Dynamics and Monte
7 < Carlo. Molecular Dynamic simulations integrate the equations of motion
8 < for a given system of particles, allowing the researcher to gain
9 < insight into the time dependent evolution of a system. Diffusion
10 < phenomena are readily studied with this simulation technique, making
11 < Molecular Dynamics the main simulation technique used in this
12 < research. Other aspects of the research fall under the Monte Carlo
13 < class of simulations. In Monte Carlo, the configuration space
14 < available to the collection of particles is sampled stochastically,
15 < or randomly. Each configuration is chosen with a given probability
16 < based on the Maxwell Boltzmann distribution. These types of simulations
17 < are best used to probe properties of a system that are only dependent
18 < only on the state of the system. Structural information about a system
19 < is most readily obtained through these types of methods.
7 > Carlo. Molecular Dynamics simulations are carried out by integrating
8 > the equations of motion for a given system of particles, allowing the
9 > researcher to gain insight into the time dependent evolution of a
10 > system. Transport phenomena are readily studied with this simulation
11 > technique, making Molecular Dynamics the main simulation technique
12 > used in this research. Other aspects of the research fall in the
13 > Monte Carlo class of simulations. In Monte Carlo, the configuration
14 > space available to the collection of particles is sampled
15 > stochastically, or randomly. Each configuration is chosen with a given
16 > probability based on the Boltzmann distribution. These types
17 > of simulations are best used to probe properties of a system that are
18 > dependent only on the state of the system. Structural information
19 > about a system is most readily obtained through these types of
20 > methods.
21  
22   Although the two techniques employed seem dissimilar, they are both
23 < linked by the overarching principles of Statistical
24 < Thermodynamics. Statistical Thermodynamics governs the behavior of
23 > linked by the over-arching principles of Statistical
24 > Mechanics. Statistical Mechanics governs the behavior of
25   both classes of simulations and dictates what each method can and
26 < cannot do. When investigating a system, one most first analyze what
27 < thermodynamic properties of the system are being probed, then chose
26 > cannot do. When investigating a system, one must first analyze what
27 > thermodynamic properties of the system are being probed, then choose
28   which method best suits that objective.
29  
30   \section{\label{introSec:statThermo}Statistical Mechanics}
31  
32   The following section serves as a brief introduction to some of the
33 < Statistical Mechanics concepts present in this dissertation.  What
34 < follows is a brief derivation of Boltzmann weighted statistics, and an
33 > Statistical Mechanics concepts presented in this dissertation.  What
34 > follows is a brief derivation of Boltzmann weighted statistics and an
35   explanation of how one can use the information to calculate an
36   observable for a system.  This section then concludes with a brief
37   discussion of the ergodic hypothesis and its relevance to this
# Line 39 | Line 39 | Consider a system, $\gamma$, with some total energy,,
39  
40   \subsection{\label{introSec:boltzman}Boltzmann weighted statistics}
41  
42 < Consider a system, $\gamma$, with some total energy,, $E_{\gamma}$.
43 < Let $\Omega(E_{\gamma})$ represent the number of degenerate ways
44 < $\boldsymbol{\Gamma}$, the collection of positions and conjugate
45 < momenta of system $\gamma$, can be configured to give
46 < $E_{\gamma}$. Further, if $\gamma$ is in contact with a bath system
47 < where energy is exchanged between the two systems, $\Omega(E)$, where
48 < $E$ is the total energy of both systems, can be represented as
42 > Consider a system, $\gamma$, with total energy $E_{\gamma}$.  Let
43 > $\Omega(E_{\gamma})$ represent the number of degenerate ways
44 > $\boldsymbol{\Gamma}\{r_1,r_2,\ldots r_n,p_1,p_2,\ldots p_n\}$, the
45 > collection of positions and conjugate momenta of system $\gamma$, can
46 > be configured to give $E_{\gamma}$. Further, if $\gamma$ is a subset
47 > of a larger system, $\boldsymbol{\Lambda}\{E_1,E_2,\ldots
48 > E_{\gamma},\ldots E_n\}$, the total degeneracy of the system can be
49 > expressed as
50   \begin{equation}
51 < \Omega(E) = \Omega(E_{\gamma}) \times \Omega(E - E_{\gamma})
51 > \Omega(\boldsymbol{\Lambda}) = \Omega(E_1) \times \Omega(E_2) \times \ldots
52 >        \Omega(E_{\gamma}) \times \ldots \Omega(E_n).
53 > \label{introEq:SM0.1}
54 > \end{equation}
55 > This multiplicative combination of degeneracies is illustrated in
56 > Fig.~\ref{introFig:degenProd}.
57 >
58 > \begin{figure}
59 > \centering
60 > \includegraphics[width=\linewidth]{omegaFig.eps}
61 > \caption[An explanation of the combination of degeneracies]{Systems A and B both have three energy levels and two indistinguishable particles. When the total energy is 2, there are two ways for each system to disperse the energy. However, for system C, the superset of A and B, the total degeneracy is the product of the degeneracy of each system. In this case $\Omega(\text{C})$ is 4.}
62 > \label{introFig:degenProd}
63 > \end{figure}
64 >
65 > Next, consider the specific case of $\gamma$ in contact with a
66 > bath. Exchange of energy is allowed between the bath and the system,
67 > subject to the constraint that the total energy
68 > ($E_{\text{bath}}+E_{\gamma}$) remain constant. $\Omega(E)$, where $E$
69 > is the total energy of both systems, can be represented as
70 > \begin{equation}
71 > \Omega(E) = \Omega(E_{\gamma}) \times \Omega(E - E_{\gamma}).
72   \label{introEq:SM1}
73   \end{equation}
74 < Or additively as
74 > Or additively as,
75   \begin{equation}
76 < \ln \Omega(E) = \ln \Omega(E_{\gamma}) + \ln \Omega(E - E_{\gamma})
76 > \ln \Omega(E) = \ln \Omega(E_{\gamma}) + \ln \Omega(E - E_{\gamma}).
77   \label{introEq:SM2}
78   \end{equation}
79  
80   The solution to Eq.~\ref{introEq:SM2} maximizes the number of
81 < degenerative configurations in $E$. \cite{Frenkel1996}
81 > degenerate configurations in $E$. \cite{Frenkel1996}
82   This gives
83   \begin{equation}
84   \frac{\partial \ln \Omega(E)}{\partial E_{\gamma}} = 0 =
85          \frac{\partial \ln \Omega(E_{\gamma})}{\partial E_{\gamma}}
86           +
87          \frac{\partial \ln \Omega(E_{\text{bath}})}{\partial E_{\text{bath}}}
88 <        \frac{\partial E_{\text{bath}}}{\partial E_{\gamma}}
88 >        \frac{\partial E_{\text{bath}}}{\partial E_{\gamma}},
89   \label{introEq:SM3}
90   \end{equation}
91 < Where $E_{\text{bath}}$ is $E-E_{\gamma}$, and
91 > where $E_{\text{bath}}$ is $E-E_{\gamma}$, and
92   $\frac{\partial E_{\text{bath}}}{\partial E_{\gamma}}$ is
93   $-1$. Eq.~\ref{introEq:SM3} becomes
94   \begin{equation}
95   \frac{\partial \ln \Omega(E_{\gamma})}{\partial E_{\gamma}} =
96 < \frac{\partial \ln \Omega(E_{\text{bath}})}{\partial E_{\text{bath}}}
96 > \frac{\partial \ln \Omega(E_{\text{bath}})}{\partial E_{\text{bath}}}.
97   \label{introEq:SM4}
98   \end{equation}
99  
# Line 83 | Line 104 | S = k_B \ln \Omega(E)
104   process is the partitioning of energy among the two systems.  This
105   allows the following definition of entropy:
106   \begin{equation}
107 < S = k_B \ln \Omega(E)
107 > S = k_B \ln \Omega(E),
108   \label{introEq:SM5}
109   \end{equation}
110 < Where $k_B$ is the Boltzmann constant.  Having defined entropy, one can
111 < also define the temperature of the system using the relation
110 > where $k_B$ is the Boltzmann constant.  Having defined entropy, one can
111 > also define the temperature of the system using the Maxwell relation,
112   \begin{equation}
113 < \frac{1}{T} = \biggl ( \frac{\partial S}{\partial E} \biggr )_{N,V}
113 > \frac{1}{T} = \biggl ( \frac{\partial S}{\partial E} \biggr )_{N,V}.
114   \label{introEq:SM6}
115   \end{equation}
116   The temperature in the system $\gamma$ is then
117   \begin{equation}
118   \beta( E_{\gamma} ) = \frac{1}{k_B T} =
119 <        \frac{\partial \ln \Omega(E_{\gamma})}{\partial E_{\gamma}}
119 >        \frac{\partial \ln \Omega(E_{\gamma})}{\partial E_{\gamma}}.
120   \label{introEq:SM7}
121   \end{equation}
122   Applying this to Eq.~\ref{introEq:SM4} gives the following
123   \begin{equation}
124 < \beta( E_{\gamma} ) = \beta( E_{\text{bath}} )
124 > \beta( E_{\gamma} ) = \beta( E_{\text{bath}} ).
125   \label{introEq:SM8}
126   \end{equation}
127 < Showing that the partitioning of energy between the two systems is
128 < actually a process of thermal equilibration.\cite{Frenkel1996}
127 > Eq.~\ref{introEq:SM8} shows that the partitioning of energy between
128 > the two systems is actually a process of thermal
129 > equilibration.\cite{Frenkel1996}
130  
131 < An application of these results is to formulate the form of an
132 < expectation value of an observable, $A$, in the canonical ensemble. In
133 < the canonical ensemble, the number of particles, $N$, the volume, $V$,
134 < and the temperature, $T$, are all held constant while the energy, $E$,
135 < is allowed to fluctuate. Returning to the previous example, the bath
131 > An application of these results is to formulate the expectation value
132 > of an observable, $A$, in the canonical ensemble. In the canonical
133 > ensemble, the number of particles, $N$, the volume, $V$, and the
134 > temperature, $T$, are all held constant while the energy, $E$, is
135 > allowed to fluctuate. Returning to the previous example, the bath
136   system is now an infinitely large thermal bath, whose exchange of
137   energy with the system $\gamma$ holds the temperature constant.  The
138 < partitioning of energy in the bath system is then related to the total
139 < energy of both systems and the fluctuations in $E_{\gamma}$:
138 > partitioning of energy between the bath and the system is then related
139 > to the total energy of both systems and the fluctuations in
140 > $E_{\gamma}$:
141   \begin{equation}
142 < \Omega( E_{\gamma} ) = \Omega( E - E_{\gamma} )
142 > \Omega( E_{\gamma} ) = \Omega( E - E_{\gamma} ).
143   \label{introEq:SM9}
144   \end{equation}
145   As for the expectation value, it can be defined
146   \begin{equation}
147   \langle A \rangle =
148          \int\limits_{\boldsymbol{\Gamma}} d\boldsymbol{\Gamma}
149 <        P_{\gamma} A(\boldsymbol{\Gamma})
149 >        P_{\gamma} A(\boldsymbol{\Gamma}),
150   \label{introEq:SM10}
151   \end{equation}
152 < Where $\int\limits_{\boldsymbol{\Gamma}} d\boldsymbol{\Gamma}$ denotes
153 < an integration over all accessible phase space, $P_{\gamma}$ is the
154 < probability of being in a given phase state and
155 < $A(\boldsymbol{\Gamma})$ is some observable that is a function of the
152 > where $\int\limits_{\boldsymbol{\Gamma}} d\boldsymbol{\Gamma}$ denotes
153 > an integration over all accessible points in phase space, $P_{\gamma}$
154 > is the probability of being in a given phase state and
155 > $A(\boldsymbol{\Gamma})$ is an observable that is a function of the
156   phase state.
157  
158   Because entropy seeks to maximize the number of degenerate states at a
# Line 138 | Line 161 | P_{\gamma} \propto \Omega( E_{\text{bath}} ) =
161   states the coupled system is able to assume. Namely,
162   \begin{equation}
163   P_{\gamma} \propto \Omega( E_{\text{bath}} ) =
164 <        e^{\ln \Omega( E - E_{\gamma})}
164 >        e^{\ln \Omega( E - E_{\gamma})}.
165   \label{introEq:SM11}
166   \end{equation}
167 < With $E_{\gamma} \ll E$, $\ln \Omega$ can be expanded in a Taylor series:
167 > Because $E_{\gamma} \ll E$, $\ln \Omega$ can be expanded in a Taylor series:
168   \begin{equation}
169   \ln \Omega ( E - E_{\gamma}) =
170          \ln \Omega (E) -
171          E_{\gamma}  \frac{\partial \ln \Omega }{\partial E}
172 <        + \ldots
172 >        + \ldots.
173   \label{introEq:SM12}
174   \end{equation}
175   Higher order terms are omitted as $E$ is an infinite thermal
176   bath. Further, using Eq.~\ref{introEq:SM7}, Eq.~\ref{introEq:SM11} can
177   be rewritten:
178   \begin{equation}
179 < P_{\gamma} \propto e^{-\beta E_{\gamma}}
179 > P_{\gamma} \propto e^{-\beta E_{\gamma}},
180   \label{introEq:SM13}
181   \end{equation}
182 < Where $\ln \Omega(E)$ has been factored out of the proportionality as a
182 > where $\ln \Omega(E)$ has been factored out of the proportionality as a
183   constant.  Normalizing the probability ($\int\limits_{\boldsymbol{\Gamma}}
184   d\boldsymbol{\Gamma} P_{\gamma} = 1$) gives
185   \begin{equation}
186   P_{\gamma} = \frac{e^{-\beta E_{\gamma}}}
187 < {\int\limits_{\boldsymbol{\Gamma}} d\boldsymbol{\Gamma} e^{-\beta E_{\gamma}}}
187 > {\int\limits_{\boldsymbol{\Gamma}} d\boldsymbol{\Gamma} e^{-\beta E_{\gamma}}}.
188   \label{introEq:SM14}
189   \end{equation}
190   This result is the standard Boltzmann statistical distribution.
# Line 170 | Line 193 | Applying it to Eq.~\ref{introEq:SM10} one can obtain t
193   \langle A \rangle =
194   \frac{\int\limits_{\boldsymbol{\Gamma}} d\boldsymbol{\Gamma}
195          A( \boldsymbol{\Gamma} ) e^{-\beta E_{\gamma}}}
196 < {\int\limits_{\boldsymbol{\Gamma}} d\boldsymbol{\Gamma} e^{-\beta E_{\gamma}}}
196 > {\int\limits_{\boldsymbol{\Gamma}} d\boldsymbol{\Gamma} e^{-\beta E_{\gamma}}}.
197   \label{introEq:SM15}
198   \end{equation}
199  
200   \subsection{\label{introSec:ergodic}The Ergodic Hypothesis}
201  
202 < One last important consideration is that of ergodicity. Ergodicity is
203 < the assumption that given an infinite amount of time, a system will
204 < visit every available point in phase space.\cite{Frenkel1996} For most
205 < systems, this is a valid assumption, except in cases where the system
206 < may be trapped in a local feature (\emph{e.g.}~glasses). When valid,
207 < ergodicity allows the unification of a time averaged observation and
208 < an ensemble averaged one. If an observation is averaged over a
209 < sufficiently long time, the system is assumed to visit all
210 < appropriately available points in phase space, giving a properly
211 < weighted statistical average. This allows the researcher freedom of
212 < choice when deciding how best to measure a given observable.  When an
213 < ensemble averaged approach seems most logical, the Monte Carlo
214 < techniques described in Sec.~\ref{introSec:monteCarlo} can be utilized.
215 < Conversely, if a problem lends itself to a time averaging approach,
216 < the Molecular Dynamics techniques in Sec.~\ref{introSec:MD} can be
217 < employed.
202 > In the case of a Molecular Dynamics simulation, rather than
203 > calculating an ensemble average integral over phase space as in
204 > Eq.~\ref{introEq:SM15}, it becomes easier to caclulate the time
205 > average of an observable. Namely,
206 > \begin{equation}
207 > \langle A \rangle_t = \frac{1}{\tau}
208 >        \int_0^{\tau} A[\boldsymbol{\Gamma}(t)]\,dt,
209 > \label{introEq:SM16}
210 > \end{equation}
211 > where the value of an observable is averaged over the length of time,
212 > $\tau$, that the simulation is run. This type of measurement mirrors
213 > the experimental measurement of an observable. In an experiment, the
214 > instrument analyzing the system must average its observation over the
215 > finite time of the measurement. What is required then, is a principle
216 > to relate the time average to the ensemble average. This is the
217 > ergodic hypothesis.
218  
219 + Ergodicity is the assumption that given an infinite amount of time, a
220 + system will visit every available point in phase
221 + space.\cite{Frenkel1996} For most systems, this is a valid assumption,
222 + except in cases where the system may be trapped in a local feature
223 + (\emph{e.g.}~glasses). When valid, ergodicity allows the unification
224 + of a time averaged observation and an ensemble averaged one. If an
225 + observation is averaged over a sufficiently long time, the system is
226 + assumed to visit all energetically accessible points in phase space,
227 + giving a properly weighted statistical average. This allows the
228 + researcher freedom of choice when deciding how best to measure a given
229 + observable.  When an ensemble averaged approach seems most logical,
230 + the Monte Carlo techniques described in Sec.~\ref{introSec:monteCarlo}
231 + can be utilized.  Conversely, if a problem lends itself to a time
232 + averaging approach, the Molecular Dynamics techniques in
233 + Sec.~\ref{introSec:MD} can be employed.
234 +
235   \section{\label{introSec:monteCarlo}Monte Carlo Simulations}
236  
237   The Monte Carlo method was developed by Metropolis and Ulam for their
238   work in fissionable material.\cite{metropolis:1949} The method is so
239 < named, because it heavily uses random numbers in its
240 < solution.\cite{allen87:csl} The Monte Carlo method allows for the
241 < solution of integrals through the stochastic sampling of the values
242 < within the integral. In the simplest case, the evaluation of an
243 < integral would follow a brute force method of
244 < sampling.\cite{Frenkel1996} Consider the following single dimensional
206 < integral:
239 > named, because it uses random numbers extensively.\cite{allen87:csl}
240 > The Monte Carlo method allows for the solution of integrals through
241 > the stochastic sampling of the values within the integral. In the
242 > simplest case, the evaluation of an integral would follow a brute
243 > force method of sampling.\cite{Frenkel1996} Consider the following
244 > single dimensional integral:
245   \begin{equation}
246 < I = f(x)dx
246 > I = \int_a^b f(x)dx.
247   \label{eq:MCex1}
248   \end{equation}
249   The equation can be recast as:
250   \begin{equation}
251 < I = (b-a)\langle f(x) \rangle
251 > I = \int^b_a \frac{f(x)}{\rho(x)} \rho(x) dx,
252 > \label{introEq:Importance1}
253 > \end{equation}
254 > where $\rho(x)$ is an arbitrary probability distribution in $x$.  If
255 > one conducts $\tau$ trials selecting a random number, $\zeta_\tau$,
256 > from the distribution $\rho(x)$ on the interval $[a,b]$, then
257 > Eq.~\ref{introEq:Importance1} becomes
258 > \begin{equation}
259 > I= \lim_{\tau \rightarrow \infty}\biggl \langle \frac{f(x)}{\rho(x)} \biggr \rangle_{\text{trials}[a,b]}.
260 > \label{introEq:Importance2}
261 > \end{equation}
262 > If $\rho(x)$ is uniformly distributed over the interval $[a,b]$,
263 > \begin{equation}
264 > \rho(x) = \frac{1}{b-a},
265 > \label{introEq:importance2b}
266 > \end{equation}
267 > then the integral can be rewritten as
268 > \begin{equation}
269 > I = (b-a)\lim_{\tau \rightarrow \infty}
270 >        \langle f(x) \rangle_{\text{trials}[a,b]}.
271   \label{eq:MCex2}
272   \end{equation}
216 Where $\langle f(x) \rangle$ is the unweighted average over the interval
217 $[a,b]$. The calculation of the integral could then be solved by
218 randomly choosing points along the interval $[a,b]$ and calculating
219 the value of $f(x)$ at each point. The accumulated average would then
220 approach $I$ in the limit where the number of trials is infinitely
221 large.
273  
274   However, in Statistical Mechanics, one is typically interested in
275   integrals of the form:
276   \begin{equation}
277   \langle A \rangle = \frac{\int d^N \mathbf{r}~A(\mathbf{r}^N)%
278          e^{-\beta V(\mathbf{r}^N)}}%
279 <        {\int d^N \mathbf{r}~e^{-\beta V(\mathbf{r}^N)}}
279 >        {\int d^N \mathbf{r}~e^{-\beta V(\mathbf{r}^N)}},
280   \label{eq:mcEnsAvg}
281   \end{equation}
282 < Where $\mathbf{r}^N$ stands for the coordinates of all $N$ particles
282 > where $\mathbf{r}^N$ stands for the coordinates of all $N$ particles
283   and $A$ is some observable that is only dependent on position. This is
284   the ensemble average of $A$ as presented in
285   Sec.~\ref{introSec:statThermo}, except here $A$ is independent of
# Line 240 | Line 291 | Importance Sampling is a method where one selects a di
291   average. This is where importance sampling comes into
292   play.\cite{allen87:csl}
293  
294 < Importance Sampling is a method where one selects a distribution from
295 < which the random configurations are chosen in order to more
296 < efficiently calculate the integral.\cite{Frenkel1996} Consider again
297 < Eq.~\ref{eq:MCex1} rewritten to be:
247 < \begin{equation}
248 < I = \int^b_a \frac{f(x)}{\rho(x)} \rho(x) dx
249 < \label{introEq:Importance1}
250 < \end{equation}
251 < Where $\rho(x)$ is an arbitrary probability distribution in $x$.  If
252 < one conducts $\tau$ trials selecting a random number, $\zeta_\tau$,
253 < from the distribution $\rho(x)$ on the interval $[a,b]$, then
254 < Eq.~\ref{introEq:Importance1} becomes
255 < \begin{equation}
256 < I= \biggl \langle \frac{f(x)}{\rho(x)} \biggr \rangle_{\text{trials}}
257 < \label{introEq:Importance2}
258 < \end{equation}
259 < Looking at Eq.~\ref{eq:mcEnsAvg}, and realizing
294 > Importance sampling is a method where the distribution, from which the
295 > random configurations are chosen, is selected in such a way as to
296 > efficiently sample the integral in question.  Looking at
297 > Eq.~\ref{eq:mcEnsAvg}, and realizing
298   \begin {equation}
299   \rho_{kT}(\mathbf{r}^N) =
300          \frac{e^{-\beta V(\mathbf{r}^N)}}
301 <        {\int d^N \mathbf{r}~e^{-\beta V(\mathbf{r}^N)}}
301 >        {\int d^N \mathbf{r}~e^{-\beta V(\mathbf{r}^N)}},
302   \label{introEq:MCboltzman}
303   \end{equation}
304 < Where $\rho_{kT}$ is the Boltzmann distribution.  The ensemble average
304 > where $\rho_{kT}$ is the Boltzmann distribution.  The ensemble average
305   can be rewritten as
306   \begin{equation}
307   \langle A \rangle = \int d^N \mathbf{r}~A(\mathbf{r}^N)
308 <        \rho_{kT}(\mathbf{r}^N)
308 >        \rho_{kT}(\mathbf{r}^N).
309   \label{introEq:Importance3}
310   \end{equation}
311   Applying Eq.~\ref{introEq:Importance1} one obtains
312   \begin{equation}
313   \langle A \rangle = \biggl \langle
314          \frac{ A \rho_{kT}(\mathbf{r}^N) }
315 <        {\rho(\mathbf{r}^N)} \biggr \rangle_{\text{trials}}
315 >        {\rho(\mathbf{r}^N)} \biggr \rangle_{\text{trials}}.
316   \label{introEq:Importance4}
317   \end{equation}
318   By selecting $\rho(\mathbf{r}^N)$ to be $\rho_{kT}(\mathbf{r}^N)$
319   Eq.~\ref{introEq:Importance4} becomes
320   \begin{equation}
321 < \langle A \rangle = \langle A(\mathbf{r}^N) \rangle_{\text{trials}}
321 > \langle A \rangle = \langle A(\mathbf{r}^N) \rangle_{kT}.
322   \label{introEq:Importance5}
323   \end{equation}
324   The difficulty is selecting points $\mathbf{r}^N$ such that they are
325   sampled from the distribution $\rho_{kT}(\mathbf{r}^N)$.  A solution
326 < was proposed by Metropolis et al.\cite{metropolis:1953} which involved
326 > was proposed by Metropolis \emph{et al}.\cite{metropolis:1953} which involved
327   the use of a Markov chain whose limiting distribution was
328   $\rho_{kT}(\mathbf{r}^N)$.
329  
# Line 307 | Line 345 | The transition probability is given by the following:
345  
346   The transition probability is given by the following:
347   \begin{equation}
348 < \pi_{mn} = \alpha_{mn} \times \accMe(m \rightarrow n)
348 > \pi_{mn} = \alpha_{mn} \times \accMe(m \rightarrow n),
349   \label{introEq:MCpi}
350   \end{equation}
351 < Where $\alpha_{mn}$ is the probability of attempting the move $m
351 > where $\alpha_{mn}$ is the probability of attempting the move $m
352   \rightarrow n$, and $\accMe$ is the probability of accepting the move
353   $m \rightarrow n$.  Defining a probability vector,
354   $\boldsymbol{\rho}$, such that
355   \begin{equation}
356   \boldsymbol{\rho} = \{\rho_1, \rho_2, \ldots \rho_m, \rho_n,
357 <        \ldots \rho_N \}
357 >        \ldots \rho_N \},
358   \label{introEq:MCrhoVector}
359   \end{equation}
360   a transition matrix $\boldsymbol{\Pi}$ can be defined,
361   whose elements are $\pi_{mn}$, for each given transition.  The
362   limiting distribution of the Markov chain can then be found by
363   applying the transition matrix an infinite number of times to the
364 < distribution vector.
364 > distribution vector,
365   \begin{equation}
366   \boldsymbol{\rho}_{\text{limit}} =
367          \lim_{N \rightarrow \infty} \boldsymbol{\rho}_{\text{initial}}
368 <        \boldsymbol{\Pi}^N
368 >        \boldsymbol{\Pi}^N.
369   \label{introEq:MCmarkovLimit}
370   \end{equation}
371   The limiting distribution of the chain is independent of the starting
372   distribution, and successive applications of the transition matrix
373 < will only yield the limiting distribution again.
373 > will only yield the limiting distribution again,
374   \begin{equation}
375 < \boldsymbol{\rho}_{\text{limit}} = \boldsymbol{\rho}_{\text{initial}}
376 <        \boldsymbol{\Pi}
375 > \boldsymbol{\rho}_{\text{limit}} = \boldsymbol{\rho}_{\text{limit}}
376 >        \boldsymbol{\Pi}.
377   \label{introEq:MCmarkovEquil}
378   \end{equation}
379  
# Line 346 | Line 384 | distribution.  Meaning, that at equilibrium the probab
384   $\boldsymbol{\rho}_{\text{limit}}$ matches the Boltzmann distribution
385   of states.  The method accomplishes this by imposing the strong
386   condition of microscopic reversibility on the equilibrium
387 < distribution.  Meaning, that at equilibrium the probability of going
388 < from $m$ to $n$ is the same as going from $n$ to $m$.
387 > distribution.  This means that at equilibrium, the probability of going
388 > from $m$ to $n$ is the same as going from $n$ to $m$,
389   \begin{equation}
390 < \rho_m\pi_{mn} = \rho_n\pi_{nm}
390 > \rho_m\pi_{mn} = \rho_n\pi_{nm}.
391   \label{introEq:MCmicroReverse}
392   \end{equation}
393   Further, $\boldsymbol{\alpha}$ is chosen to be a symmetric matrix in
# Line 357 | Line 395 | Eq.~\ref{introEq:MCmicroReverse} becomes
395   Eq.~\ref{introEq:MCmicroReverse} becomes
396   \begin{equation}
397   \frac{\accMe(m \rightarrow n)}{\accMe(n \rightarrow m)} =
398 <        \frac{\rho_n}{\rho_m}
398 >        \frac{\rho_n}{\rho_m}.
399   \label{introEq:MCmicro2}
400   \end{equation}
401   For a Boltzmann limiting distribution,
402   \begin{equation}
403   \frac{\rho_n}{\rho_m} = e^{-\beta[\mathcal{U}(n) - \mathcal{U}(m)]}
404 <        = e^{-\beta \Delta \mathcal{U}}
404 >        = e^{-\beta \Delta \mathcal{U}}.
405   \label{introEq:MCmicro3}
406   \end{equation}
407   This allows for the following set of acceptance rules be defined:
# Line 381 | Line 419 | Metropolis method proceeds as follows
419   \begin{enumerate}
420   \item Generate an initial configuration $\mathbf{r}^N$ which has some finite probability in $\rho_{kT}$.
421   \item Modify $\mathbf{r}^N$, to generate configuration $\mathbf{r^{\prime}}^N$.
422 < \item If the new configuration lowers the energy of the system, accept the move with unity ($\mathbf{r}^N$ becomes $\mathbf{r^{\prime}}^N$).  Otherwise accept with probability $e^{-\beta \Delta \mathcal{U}}$.
422 > \item If the new configuration lowers the energy of the system, accept the move ($\mathbf{r}^N$ becomes $\mathbf{r^{\prime}}^N$).  Otherwise accept with probability $e^{-\beta \Delta \mathcal{U}}$.
423   \item Accumulate the average for the configurational observable of interest.
424   \item Repeat from step 2 until the average converges.
425   \end{enumerate}
# Line 394 | Line 432 | Molecular Dynamics is when the equations of motion for
432   \section{\label{introSec:MD}Molecular Dynamics Simulations}
433  
434   The main simulation tool used in this research is Molecular Dynamics.
435 < Molecular Dynamics is when the equations of motion for a system are
436 < integrated in order to obtain information about both the positions and
437 < momentum of a system, allowing the calculation of not only
438 < configurational observables, but momenta dependent ones as well:
439 < diffusion constants, velocity auto correlations, folding/unfolding
440 < events, etc.  Due to the principle of ergodicity,
441 < Sec.~\ref{introSec:ergodic}, the average of these observables over the
442 < time period of the simulation are taken to be the ensemble averages
443 < for the system.
435 > Molecular Dynamics is the numerical integration of the equations of
436 > motion for a system in order to obtain information about the dynamic
437 > changes in the positions and momentum of the constituent particles.
438 > This allows the calculation of not only configurational observables,
439 > but momentum dependent ones as well: diffusion constants, relaxation
440 > events, folding/unfolding events, etc. With the time dependent
441 > information gained from a Molecular Dynamics simulation, one can also
442 > calculate time correlation functions of the form\cite{Hansen86}
443 > \begin{equation}
444 > \langle A(t)\,A(0)\rangle = \lim_{\tau\rightarrow\infty} \frac{1}{\tau}
445 >        \int_0^{\tau} A(t+t^{\prime})\,A(t^{\prime})\,dt^{\prime}.
446 > \label{introEq:timeCorr}
447 > \end{equation}
448 > These correlations can be used to measure fundamental time constants
449 > of a system, such as diffusion constants from the velocity
450 > autocorrelation or dipole relaxation times from the dipole
451 > autocorrelation.  Due to the principle of ergodicity,
452 > Sec.~\ref{introSec:ergodic}, the average of static observables over
453 > the length of the simulation are taken to be the ensemble averages for
454 > the system.
455  
456   The choice of when to use molecular dynamics over Monte Carlo
457   techniques, is normally decided by the observables in which the
458 < researcher is interested.  If the observables depend on momenta in
459 < any fashion, then the only choice is molecular dynamics in some form.
458 > researcher is interested.  If the observables depend on time in any
459 > fashion, then the only choice is molecular dynamics in some form.
460   However, when the observable is dependent only on the configuration,
461 < then most of the time Monte Carlo techniques will be more efficient.
461 > then for most small systems, Monte Carlo techniques will be more
462 > efficient.  The focus of research in the second half of this
463 > dissertation is centered on the dynamic properties of phospholipid
464 > bilayers, making molecular dynamics key in the simulation of those
465 > properties.
466  
467 < The focus of research in the second half of this dissertation is
415 < centered around the dynamic properties of phospholipid bilayers,
416 < making molecular dynamics key in the simulation of those properties.
467 > \subsection{\label{introSec:mdAlgorithm}Molecular Dynamics Algorithms}
468  
418 \subsection{\label{introSec:mdAlgorithm}The Molecular Dynamics Algorithm}
419
469   To illustrate how the molecular dynamics technique is applied, the
470   following sections will describe the sequence involved in a
471   simulation.  Sec.~\ref{introSec:mdInit} deals with the initialization
# Line 432 | Line 481 | water, and in other cases structured the lipids into p
481   Ch.~\ref{chapt:lipid} deals with the formation and equilibrium dynamics of
482   phospholipid membranes.  Therefore in these simulations initial
483   positions were selected that in some cases dispersed the lipids in
484 < water, and in other cases structured the lipids into performed
484 > water, and in other cases structured the lipids into preformed
485   bilayers.  Important considerations at this stage of the simulation are:
486   \begin{itemize}
487 < \item There are no major overlaps of molecular or atomic orbitals
488 < \item Velocities are chosen in such a way as to not give the system a non=zero total momentum or angular momentum.
489 < \item It is also sometimes desirable to select the velocities to correctly sample the target temperature.
487 > \item There are no overlapping molecules or atoms.
488 > \item Velocities are chosen in such a way as to give the system zero total momentum and angular momentum.
489 > \item It is also sometimes desirable to select the velocities to correctly sample the ensemble at a particular target temperature.
490   \end{itemize}
491  
492 < The first point is important due to the amount of potential energy
493 < generated by having two particles too close together.  If overlap
494 < occurs, the first evaluation of forces will return numbers so large as
495 < to render the numerical integration of the motion meaningless.  The
496 < second consideration keeps the system from drifting or rotating as a
497 < whole.  This arises from the fact that most simulations are of systems
498 < in equilibrium in the absence of outside forces.  Therefore any net
492 > The first point is important due to the forces generated when two
493 > particles are too close together.  If overlap occurs, the first
494 > evaluation of forces will return numbers so large as to render the
495 > numerical integration of the equations motion meaningless.  The second
496 > consideration keeps the system from drifting or rotating as a whole.
497 > This arises from the fact that most simulations are of systems in
498 > equilibrium in the absence of outside forces.  Therefore any net
499   movement would be unphysical and an artifact of the simulation method
500   used.  The final point addresses the selection of the magnitude of the
501 < initial velocities.  For many simulations it is convenient to use
502 < this opportunity to scale the amount of kinetic energy to reflect the
501 > initial velocities.  For many simulations it is convenient to use this
502 > opportunity to scale the amount of kinetic energy to reflect the
503   desired thermal distribution of the system.  However, it must be noted
504 < that most systems will require further velocity rescaling after the
505 < first few initial simulation steps due to either loss or gain of
506 < kinetic energy from energy stored in potential degrees of freedom.
504 > that due to equipartition, most systems will require further velocity
505 > rescaling after the first few initial simulation steps due to either
506 > loss or gain of kinetic energy from energy stored as potential energy.
507  
508   \subsection{\label{introSec:mdForce}Force Evaluation}
509  
510   The evaluation of forces is the most computationally expensive portion
511 < of a given molecular dynamics simulation.  This is due entirely to the
512 < evaluation of long range forces in a simulation, typically pair-wise.
513 < These forces are most commonly the Van der Waals force, and sometimes
514 < Coulombic forces as well.  For a pair-wise force, there are $N(N-1)/ 2$
515 < pairs to be evaluated, where $N$ is the number of particles in the
516 < system.  This leads to the calculations scaling as $N^2$, making large
517 < simulations prohibitive in the absence of any computation saving
518 < techniques.
511 > of any molecular dynamics simulation.  This is due entirely to the
512 > evaluation of long range forces in a simulation, which are typically
513 > pair potentials.  These forces are most commonly the van der Waals
514 > force, and sometimes Coulombic forces as well.  For a pair-wise
515 > interaction, there are $N(N-1)/ 2$ pairs to be evaluated, where $N$ is
516 > the number of particles in the system.  This leads to calculations which
517 > scale as $N^2$, making large simulations prohibitive in the absence
518 > of any computation saving techniques.
519  
520 < Another consideration one must resolve, is that in a given simulation
520 > Another consideration one must resolve, is that in a given simulation,
521   a disproportionate number of the particles will feel the effects of
522   the surface.\cite{allen87:csl} For a cubic system of 1000 particles
523   arranged in a $10 \times 10 \times 10$ cube, 488 particles will be
524 < exposed to the surface.  Unless one is simulating an isolated particle
525 < group in a vacuum, the behavior of the system will be far from the
526 < desired bulk characteristics.  To offset this, simulations employ the
527 < use of periodic boundary images.\cite{born:1912}
524 > exposed to the surface.  Unless one is simulating an isolated cluster
525 > in a vacuum, the behavior of the system will be far from the desired
526 > bulk characteristics.  To offset this, simulations employ the use of
527 > periodic images.\cite{born:1912}
528  
529   The technique involves the use of an algorithm that replicates the
530 < simulation box on an infinite lattice in Cartesian space.  Any given
530 > simulation box on an infinite lattice in Cartesian space.  Any
531   particle leaving the simulation box on one side will have an image of
532 < itself enter on the opposite side (see Fig.~\ref{introFig:pbc}).  In
533 < addition, this sets that any given particle pair has an image, real or
534 < periodic, within $fix$ of each other.  A discussion of the method used
535 < to calculate the periodic image can be found in
536 < Sec.\ref{oopseSec:pbc}.
532 > itself enter on the opposite side (see
533 > Fig.~\ref{introFig:pbc}). Therefore, a pair of particles have an
534 > image of each other, real or periodic, within $\text{box}/2$.  A
535 > discussion of the method used to calculate the periodic image can be
536 > found in Sec.\ref{oopseSec:pbc}.
537  
538   \begin{figure}
539   \centering
# Line 493 | Line 542 | Returning to the topic of the computational scale of t
542   \label{introFig:pbc}
543   \end{figure}
544  
545 < Returning to the topic of the computational scale of the force
545 > Returning to the topic of the computational scaling of the force
546   evaluation, the use of periodic boundary conditions requires that a
547   cutoff radius be employed.  Using a cutoff radius improves the
548   efficiency of the force evaluation, as particles farther than a
549   predetermined distance, $r_{\text{cut}}$, are not included in the
550   calculation.\cite{Frenkel1996} In a simulation with periodic images,
551 < $r_{\text{cut}}$ has a maximum value of $\text{box}/2$.
552 < Fig.~\ref{introFig:rMax} illustrates how when using an
553 < $r_{\text{cut}}$ larger than this value, or in the extreme limit of no
554 < $r_{\text{cut}}$ at all, the corners of the simulation box are
555 < unequally weighted due to the lack of particle images in the $x$, $y$,
556 < or $z$ directions past a distance of $\text{box} / 2$.
551 > there are two methods to choose from, both with their own cutoff
552 > limits. In the minimum image convention, $r_{\text{cut}}$ has a
553 > maximum value of $\text{box}/2$. This is because each atom has only
554 > one image that is seen by other atoms. The image used is the one that
555 > minimizes the distance between the two atoms. A system of wrapped
556 > images about a central atom therefore has a maximum length scale of
557 > box on a side (Fig.~\ref{introFig:rMaxMin}). The second convention,
558 > multiple image convention, has a maximum $r_{\text{cut}}$ of the box
559 > length itself. Here multiple images of each atom are replicated in the
560 > periodic cells surrounding the central atom, this causes the atom to
561 > see multiple copies of several atoms. If the cutoff radius is larger
562 > than box, however, then the atom will see an image of itself, and
563 > attempt to calculate an unphysical self-self force interaction
564 > (Fig.~\ref{introFig:rMaxMult}). Due to the increased complexity and
565 > commputaional ineffeciency of the multiple image method, the minimum
566 > image method is the used throughout this research.
567  
568   \begin{figure}
569   \centering
570   \includegraphics[width=\linewidth]{rCutMaxFig.eps}
571 < \caption[An explanation of $r_{\text{cut}}$]{The yellow atom has all other images wrapped to itself as the center. If $r_{\text{cut}}=\text{box}/2$, then the distribution is uniform (blue atoms). However, when $r_{\text{cut}}>\text{box}/2$ the corners are disproportionately weighted (green atoms) vs the axial directions (shaded regions).}
572 < \label{introFig:rMax}
571 > \caption[An explanation of minimum image convention]{The yellow atom has all other images wrapped to itself as the center. If $r_{\text{cut}}=\text{box}/2$, then the distribution is uniform (blue atoms). However, when $r_{\text{cut}}>\text{box}/2$ the corners are disproportionately weighted (green atoms) vs the axial directions (shaded regions).}
572 > \label{introFig:rMaxMin}
573   \end{figure}
574  
575 < With the use of an $r_{\text{cut}}$, however, comes a discontinuity in
575 > \begin{figure}
576 > \centering
577 > \includegraphics[width=\linewidth]{rCutMaxMultFig.eps}
578 > \caption[An explanation of multiple image convention]{The yellow atom is the central wrapping point. The blue atoms are the minimum images of the system about the central atom. The boxes with the green atoms are multiple images of the central box. If $r_{\text{cut}} \geq \text{box}$ then the central atom sees multiple images of itself (red atom), creating a self-self force evaluation.}
579 > \label{introFig:rMaxMult}
580 > \end{figure}
581 >
582 > With the use of a cutoff radius, however, comes a discontinuity in
583   the potential energy curve (Fig.~\ref{introFig:shiftPot}). To fix this
584   discontinuity, one calculates the potential energy at the
585   $r_{\text{cut}}$, and adds that value to the potential, causing
# Line 529 | Line 595 | neighbor list. \cite{allen87:csl} In the Verlet method
595   \end{figure}
596  
597   The second main simplification used in this research is the Verlet
598 < neighbor list. \cite{allen87:csl} In the Verlet method, one generates
598 > neighbor list.\cite{allen87:csl} In the Verlet method, one generates
599   a list of all neighbor atoms, $j$, surrounding atom $i$ within some
600   cutoff $r_{\text{list}}$, where $r_{\text{list}}>r_{\text{cut}}$.
601   This list is created the first time forces are evaluated, then on
602   subsequent force evaluations, pair calculations are only calculated
603 < from the neighbor lists.  The lists are updated if any given particle
603 > from the neighbor lists.  The lists are updated if any particle
604   in the system moves farther than $r_{\text{list}}-r_{\text{cut}}$,
605 < giving rise to the possibility that a particle has left or joined a
605 > which indeicates the possibility that a particle has left or joined the
606   neighbor list.
607  
608 < \subsection{\label{introSec:mdIntegrate} Integration of the equations of motion}
608 > \subsection{\label{introSec:mdIntegrate} Integration of the Equations of Motion}
609  
610   A starting point for the discussion of molecular dynamics integrators
611   is the Verlet algorithm.\cite{Frenkel1996} It begins with a Taylor
612   expansion of position in time:
613   \begin{equation}
614   q(t+\Delta t)= q(t) + v(t)\Delta t + \frac{F(t)}{2m}\Delta t^2 +
615 <        \frac{\Delta t^3}{3!}\frac{\partial q(t)}{\partial t} +
616 <        \mathcal{O}(\Delta t^4)
615 >        \frac{\Delta t^3}{3!}\frac{\partial^3 q(t)}{\partial t^3} +
616 >        \mathcal{O}(\Delta t^4) .
617   \label{introEq:verletForward}
618   \end{equation}
619   As well as,
620   \begin{equation}
621   q(t-\Delta t)= q(t) - v(t)\Delta t + \frac{F(t)}{2m}\Delta t^2 -
622 <        \frac{\Delta t^3}{3!}\frac{\partial q(t)}{\partial t} +
623 <        \mathcal{O}(\Delta t^4)
622 >        \frac{\Delta t^3}{3!}\frac{\partial^3 q(t)}{\partial t^3} +
623 >        \mathcal{O}(\Delta t^4) ,
624   \label{introEq:verletBack}
625   \end{equation}
626 < Adding together Eq.~\ref{introEq:verletForward} and
626 > where $m$ is the mass of the particle, $q(t)$ is the position at time
627 > $t$, $v(t)$ the velocity, and $F(t)$ the force acting on the
628 > particle. Adding together Eq.~\ref{introEq:verletForward} and
629   Eq.~\ref{introEq:verletBack} results in,
630   \begin{equation}
631 < eq here
631 > q(t+\Delta t)+q(t-\Delta t) =
632 >        2q(t) + \frac{F(t)}{m}\Delta t^2 + \mathcal{O}(\Delta t^4) ,
633   \label{introEq:verletSum}
634   \end{equation}
635 < Or equivalently,
635 > or equivalently,
636   \begin{equation}
637 < eq here
637 > q(t+\Delta t) \approx
638 >        2q(t) - q(t-\Delta t) + \frac{F(t)}{m}\Delta t^2.
639   \label{introEq:verletFinal}
640   \end{equation}
641   Which contains an error in the estimate of the new positions on the
642   order of $\Delta t^4$.
643  
644   In practice, however, the simulations in this research were integrated
645 < with a velocity reformulation of the Verlet method.\cite{allen87:csl}
646 < \begin{equation}
647 < eq here
648 < \label{introEq:MDvelVerletPos}
649 < \end{equation}
650 < \begin{equation}
581 < eq here
645 > with the velocity reformulation of the Verlet method.\cite{allen87:csl}
646 > \begin{align}
647 > q(t+\Delta t) &= q(t) + v(t)\Delta t + \frac{F(t)}{2m}\Delta t^2 ,%
648 > \label{introEq:MDvelVerletPos} \\%
649 > %
650 > v(t+\Delta t) &= v(t) + \frac{\Delta t}{2m}[F(t) + F(t+\Delta t)] .%
651   \label{introEq:MDvelVerletVel}
652 < \end{equation}
652 > \end{align}
653   The original Verlet algorithm can be regained by substituting the
654   velocity back into Eq.~\ref{introEq:MDvelVerletPos}.  The Verlet
655   formulations are chosen in this research because the algorithms have
656 < very little long term drift in energy conservation.  Energy
657 < conservation in a molecular dynamics simulation is of extreme
658 < importance, as it is a measure of how closely one is following the
659 < ``true'' trajectory with the finite integration scheme.  An exact
660 < solution to the integration will conserve area in phase space, as well
656 > very little long time drift in energy.  Energy conservation in a
657 > molecular dynamics simulation is of extreme importance, as it is a
658 > measure of how closely one is following the ``true'' trajectory with
659 > the finite integration scheme.  An exact solution to the integration
660 > will conserve area in phase space (i.e.~will be symplectic), as well
661   as be reversible in time, that is, the trajectory integrated forward
662   or backwards will exactly match itself.  Having a finite algorithm
663   that both conserves area in phase space and is time reversible,
664 < therefore increases, but does not guarantee the ``correctness'' or the
665 < integrated trajectory.
664 > therefore increases the reliability, but does not guarantee the
665 > ``correctness'' of the integrated trajectory.
666  
667   It can be shown,\cite{Frenkel1996} that although the Verlet algorithm
668   does not rigorously preserve the actual Hamiltonian, it does preserve
# Line 602 | Line 671 | hypothesis (Sec.~\ref{introSec:StatThermo}), it is kno
671   reversible.  The fact that it shadows the true Hamiltonian in phase
672   space is acceptable in actual simulations as one is interested in the
673   ensemble average of the observable being measured.  From the ergodic
674 < hypothesis (Sec.~\ref{introSec:StatThermo}), it is known that the time
674 > hypothesis (Sec.~\ref{introSec:statThermo}), it is known that the time
675   average will match the ensemble average, therefore two similar
676   trajectories in phase space should give matching statistical averages.
677  
678   \subsection{\label{introSec:MDfurther}Further Considerations}
679 +
680   In the simulations presented in this research, a few additional
681 < parameters are needed to describe the motions.  The simulations
682 < involving water and phospholipids in Ch.~\ref{chaptLipids} are
681 > parameters are needed to describe the motions.  In the simulations
682 > involving water and phospholipids in Ch.~\ref{chapt:lipid}, we are
683   required to integrate the equations of motions for dipoles on atoms.
684   This involves an additional three parameters be specified for each
685   dipole atom: $\phi$, $\theta$, and $\psi$.  These three angles are
686   taken to be the Euler angles, where $\phi$ is a rotation about the
687   $z$-axis, and $\theta$ is a rotation about the new $x$-axis, and
688   $\psi$ is a final rotation about the new $z$-axis (see
689 < Fig.~\ref{introFig:euleerAngles}).  This sequence of rotations can be
690 < accumulated into a single $3 \times 3$ matrix $\mathbf{A}$
689 > Fig.~\ref{introFig:eulerAngles}).  This sequence of rotations can be
690 > accumulated into a single $3 \times 3$ matrix, $\mathsf{A}$,
691   defined as follows:
692   \begin{equation}
693 < eq here
693 > \mathsf{A} =
694 > \begin{bmatrix}
695 >        \cos\phi\cos\psi-\sin\phi\cos\theta\sin\psi &%
696 >        \sin\phi\cos\psi+\cos\phi\cos\theta\sin\psi &%
697 >        \sin\theta\sin\psi \\%
698 >        %
699 >        -\cos\phi\sin\psi-\sin\phi\cos\theta\cos\psi &%
700 >        -\sin\phi\sin\psi+\cos\phi\cos\theta\cos\psi &%
701 >        \sin\theta\cos\psi \\%
702 >        %
703 >        \sin\phi\sin\theta &%
704 >        -\cos\phi\sin\theta &%
705 >        \cos\theta
706 > \end{bmatrix}.
707   \label{introEq:EulerRotMat}
708   \end{equation}
709  
710 < The equations of motion for Euler angles can be written down as
711 < \cite{allen87:csl}
712 < \begin{equation}
713 < eq here
714 < \label{introEq:MDeuleeerPsi}
715 < \end{equation}
716 < Where $\omega^s_i$ is the angular velocity in the lab space frame
717 < along Cartesian coordinate $i$.  However, a difficulty arises when
710 > \begin{figure}
711 > \centering
712 > \includegraphics[width=\linewidth]{eulerRotFig.eps}
713 > \caption[Euler rotation of Cartesian coordinates]{The rotation scheme for Euler angles. First is a rotation of $\phi$ about the $z$ axis (blue rotation). Next is a rotation of $\theta$ about the new $x$ axis (green rotation). Lastly is a final rotation of $\psi$ about the new $z$ axis (red rotation).}
714 > \label{introFig:eulerAngles}
715 > \end{figure}
716 >
717 > The equations of motion for Euler angles can be written down
718 > as\cite{allen87:csl}
719 > \begin{align}
720 > \dot{\phi} &= -\omega^s_x \frac{\sin\phi\cos\theta}{\sin\theta} +
721 >        \omega^s_y \frac{\cos\phi\cos\theta}{\sin\theta} +
722 >        \omega^s_z,
723 > \label{introEq:MDeulerPhi} \\%
724 > %
725 > \dot{\theta} &= \omega^s_x \cos\phi + \omega^s_y \sin\phi,
726 > \label{introEq:MDeulerTheta} \\%
727 > %
728 > \dot{\psi} &= \omega^s_x \frac{\sin\phi}{\sin\theta} -
729 >        \omega^s_y \frac{\cos\phi}{\sin\theta},
730 > \label{introEq:MDeulerPsi}
731 > \end{align}
732 > where $\omega^s_{\alpha}$ is the angular velocity in the lab space frame
733 > along Cartesian coordinate $\alpha$.  However, a difficulty arises when
734   attempting to integrate Eq.~\ref{introEq:MDeulerPhi} and
735   Eq.~\ref{introEq:MDeulerPsi}. The $\frac{1}{\sin \theta}$ present in
736   both equations means there is a non-physical instability present when
737 < $\theta$ is 0 or $\pi$.
738 <
739 < To correct for this, the simulations integrate the rotation matrix,
740 < $\mathbf{A}$, directly, thus avoiding the instability.
642 < This method was proposed by Dullwebber
643 < \emph{et. al.}\cite{Dullwebber:1997}, and is presented in
737 > $\theta$ is 0 or $\pi$. To correct for this, the simulations integrate
738 > the rotation matrix, $\mathsf{A}$, directly, thus avoiding the
739 > instability.  This method was proposed by Dullweber
740 > \emph{et. al.}\cite{Dullweber1997}, and is presented in
741   Sec.~\ref{introSec:MDsymplecticRot}.
742  
743 < \subsubsection{\label{introSec:MDliouville}Liouville Propagator}
743 > \subsection{\label{introSec:MDliouville}Liouville Propagator}
744  
745   Before discussing the integration of the rotation matrix, it is
746   necessary to understand the construction of a ``good'' integration
747 < scheme.  It has been previously
748 < discussed(Sec.~\ref{introSec:MDintegrate}) how it is desirable for an
749 < integrator to be symplectic, or time reversible.  The following is an
747 > scheme.  It has been previously discussed
748 > (Sec.~\ref{introSec:mdIntegrate}) how it is desirable for an
749 > integrator to be symplectic and time reversible.  The following is an
750   outline of the Trotter factorization of the Liouville Propagator as a
751 < scheme for generating symplectic integrators. \cite{Tuckerman:1992}
751 > scheme for generating symplectic, time-reversible
752 > integrators.\cite{Tuckerman92}
753  
754   For a system with $f$ degrees of freedom the Liouville operator can be
755   defined as,
756   \begin{equation}
757 < eq here
757 > iL=\sum^f_{j=1} \biggl [\dot{q}_j\frac{\partial}{\partial q_j} +
758 >        F_j\frac{\partial}{\partial p_j} \biggr ].
759   \label{introEq:LiouvilleOperator}
760   \end{equation}
761 < Here, $r_j$ and $p_j$ are the position and conjugate momenta of a
762 < degree of freedom, and $f_j$ is the force on that degree of freedom.
761 > Here, $q_j$ and $p_j$ are the position and conjugate momenta of a
762 > degree of freedom, and $F_j$ is the force on that degree of freedom.
763   $\Gamma$ is defined as the set of all positions and conjugate momenta,
764 < $\{r_j,p_j\}$, and the propagator, $U(t)$, is defined
764 > $\{q_j,p_j\}$, and the propagator, $U(t)$, is defined
765   \begin {equation}
766 < eq here
766 > U(t) = e^{iLt}.
767   \label{introEq:Lpropagator}
768   \end{equation}
769   This allows the specification of $\Gamma$ at any time $t$ as
770   \begin{equation}
771 < eq here
771 > \Gamma(t) = U(t)\Gamma(0).
772   \label{introEq:Lp2}
773   \end{equation}
774   It is important to note, $U(t)$ is a unitary operator meaning
775   \begin{equation}
776 < U(-t)=U^{-1}(t)
776 > U(-t)=U^{-1}(t).
777   \label{introEq:Lp3}
778   \end{equation}
779  
780   Decomposing $L$ into two parts, $iL_1$ and $iL_2$, one can use the
781   Trotter theorem to yield
782 < \begin{equation}
783 < eq here
784 < \label{introEq:Lp4}
785 < \end{equation}
786 < Where $\Delta t = \frac{t}{P}$.
782 > \begin{align}
783 > e^{iLt} &= e^{i(L_1 + L_2)t}, \notag \\%
784 > %
785 >        &= \biggl [ e^{i(L_1 +L_2)\frac{t}{P}} \biggr]^P, \notag \\%
786 > %
787 >        &= \biggl [ e^{iL_1\frac{\Delta t}{2}}\, e^{iL_2\Delta t}\,
788 >        e^{iL_1\frac{\Delta t}{2}} \biggr ]^P +
789 >        \mathcal{O}\biggl (\frac{t^3}{P^2} \biggr ), \label{introEq:Lp4}
790 > \end{align}
791 > where $\Delta t = t/P$.
792   With this, a discrete time operator $G(\Delta t)$ can be defined:
793 < \begin{equation}
794 < eq here
793 > \begin{align}
794 > G(\Delta t) &= e^{iL_1\frac{\Delta t}{2}}\, e^{iL_2\Delta t}\,
795 >        e^{iL_1\frac{\Delta t}{2}}, \notag \\%
796 > %
797 >        &= U_1 \biggl ( \frac{\Delta t}{2} \biggr )\, U_2 ( \Delta t )\,
798 >        U_1 \biggl ( \frac{\Delta t}{2} \biggr ).
799   \label{introEq:Lp5}
800 < \end{equation}
801 < Because $U_1(t)$ and $U_2(t)$ are unitary, $G|\Delta t)$ is also
802 < unitary.  Meaning an integrator based on this factorization will be
800 > \end{align}
801 > Because $U_1(t)$ and $U_2(t)$ are unitary, $G(\Delta t)$ is also
802 > unitary.  This means that an integrator based on this factorization will be
803   reversible in time.
804  
805   As an example, consider the following decomposition of $L$:
806 + \begin{align}
807 + iL_1 &= \dot{q}\frac{\partial}{\partial q},%
808 + \label{introEq:Lp6a} \\%
809 + %
810 + iL_2 &= F(q)\frac{\partial}{\partial p}.%
811 + \label{introEq:Lp6b}
812 + \end{align}
813 + This leads to propagator $G( \Delta t )$ as,
814   \begin{equation}
815 < eq here
816 < \label{introEq:Lp6}
815 > G(\Delta t) =  e^{\frac{\Delta t}{2} F(q)\frac{\partial}{\partial p}} \,
816 >        e^{\Delta t\,\dot{q}\frac{\partial}{\partial q}} \,
817 >        e^{\frac{\Delta t}{2} F(q)\frac{\partial}{\partial p}}.
818 > \label{introEq:Lp7}
819   \end{equation}
820 < Operating $G(\Delta t)$ on $\Gamma)0)$, and utilizing the operator property
820 > Operating $G(\Delta t)$ on $\Gamma(0)$, and utilizing the operator property
821   \begin{equation}
822 < eq here
822 > e^{c\frac{\partial}{\partial x}}\, f(x) = f(x+c),
823   \label{introEq:Lp8}
824   \end{equation}
825 < Where $c$ is independent of $q$.  One obtains the following:  
826 < \begin{equation}
827 < eq here
828 < \label{introEq:Lp8}
829 < \end{equation}
825 > where $c$ is independent of $x$.  One obtains the following:  
826 > \begin{align}
827 > \dot{q}\biggl (\frac{\Delta t}{2}\biggr ) &=
828 >        \dot{q}(0) + \frac{\Delta t}{2m}\, F[q(0)], \label{introEq:Lp9a}\\%
829 > %
830 > q(\Delta t) &= q(0) + \Delta t\, \dot{q}\biggl (\frac{\Delta t}{2}\biggr ),%
831 >        \label{introEq:Lp9b}\\%
832 > %
833 > \dot{q}(\Delta t) &= \dot{q}\biggl (\frac{\Delta t}{2}\biggr ) +
834 >        \frac{\Delta t}{2m}\, F[q(0)]. \label{introEq:Lp9c}
835 > \end{align}
836   Or written another way,
837 < \begin{equation}
838 < eq here
839 < \label{intorEq:Lp9}
840 < \end{equation}
837 > \begin{align}
838 > q(t+\Delta t) &= q(0) + \dot{q}(0)\Delta t +
839 >        \frac{F[q(0)]}{m}\frac{\Delta t^2}{2}, %
840 > \label{introEq:Lp10a} \\%
841 > %
842 > \dot{q}(\Delta t) &= \dot{q}(0) + \frac{\Delta t}{2m}
843 >        \biggl [F[q(0)] + F[q(\Delta t)] \biggr]. %
844 > \label{introEq:Lp10b}
845 > \end{align}
846   This is the velocity Verlet formulation presented in
847 < Sec.~\ref{introSec:MDintegrate}.  Because this integration scheme is
847 > Sec.~\ref{introSec:mdIntegrate}.  Because this integration scheme is
848   comprised of unitary propagators, it is symplectic, and therefore area
849   preserving in phase space.  From the preceding factorization, one can
850   see that the integration of the equations of motion would follow:
# Line 729 | Line 858 | see that the integration of the equations of motion wo
858   \item Repeat from step 1 with the new position, velocities, and forces assuming the roles of the initial values.
859   \end{enumerate}
860  
861 < \subsubsection{\label{introSec:MDsymplecticRot} Symplectic Propagation of the Rotation Matrix}
861 > \subsection{\label{introSec:MDsymplecticRot} Symplectic Propagation of the Rotation Matrix}
862  
863   Based on the factorization from the previous section,
864 < Dullweber\emph{et al.}\cite{Dullweber:1997}~ proposed a scheme for the
865 < symplectic propagation of the rotation matrix, $\mathbf{A}$, as an
864 > Dullweber\emph{et al}.\cite{Dullweber1997}~ proposed a scheme for the
865 > symplectic propagation of the rotation matrix, $\mathsf{A}$, as an
866   alternative method for the integration of orientational degrees of
867   freedom. The method starts with a straightforward splitting of the
868   Liouville operator:
869 < \begin{equation}
870 < eq here
871 < \label{introEq:SR1}
872 < \end{equation}
873 < Where $\boldsymbol{\tau}(\mathbf{A})$ are the torques of the system
874 < due to the configuration, and $\boldsymbol{/pi}$ are the conjugate
869 > \begin{align}
870 > iL_{\text{pos}} &= \dot{q}\frac{\partial}{\partial q} +
871 >        \mathsf{\dot{A}}\frac{\partial}{\partial \mathsf{A}} ,
872 > \label{introEq:SR1a} \\%
873 > %
874 > iL_F &= F(q)\frac{\partial}{\partial p},
875 > \label{introEq:SR1b} \\%
876 > iL_{\tau} &= \tau(\mathsf{A})\frac{\partial}{\partial j},
877 > \label{introEq:SR1b} \\%
878 > \end{align}
879 > where $\tau(\mathsf{A})$ is the torque of the system
880 > due to the configuration, and $j$ is the conjugate
881   angular momenta of the system. The propagator, $G(\Delta t)$, becomes
882   \begin{equation}
883 < eq here
883 > G(\Delta t) = e^{\frac{\Delta t}{2} F(q)\frac{\partial}{\partial p}} \,
884 >        e^{\frac{\Delta t}{2} \tau(\mathsf{A})\frac{\partial}{\partial j}} \,
885 >        e^{\Delta t\,iL_{\text{pos}}} \,
886 >        e^{\frac{\Delta t}{2} \tau(\mathsf{A})\frac{\partial}{\partial j}} \,
887 >        e^{\frac{\Delta t}{2} F(q)\frac{\partial}{\partial p}}.
888   \label{introEq:SR2}
889   \end{equation}
890   Propagation of the linear and angular momenta follows as in the Verlet
891   scheme.  The propagation of positions also follows the Verlet scheme
892   with the addition of a further symplectic splitting of the rotation
893 < matrix propagation, $\mathcal{G}_{\text{rot}}(\Delta t)$.
893 > matrix propagation, $\mathcal{U}_{\text{rot}}(\Delta t)$, within
894 > $U_{\text{pos}}(\Delta t)$.
895   \begin{equation}
896 < eq here
896 > \mathcal{U}_{\text{rot}}(\Delta t) =
897 >        \mathcal{U}_x \biggl(\frac{\Delta t}{2}\biggr)\,
898 >        \mathcal{U}_y \biggl(\frac{\Delta t}{2}\biggr)\,
899 >        \mathcal{U}_z (\Delta t)\,
900 >        \mathcal{U}_y \biggl(\frac{\Delta t}{2}\biggr)\,
901 >        \mathcal{U}_x \biggl(\frac{\Delta t}{2}\biggr),
902   \label{introEq:SR3}
903   \end{equation}
904 < Where $\mathcal{G}_j$ is a unitary rotation of $\mathbf{A}$ and
905 < $\boldsymbol{\pi}$ about each axis $j$.  As all propagations are now
904 > where $\mathcal{U}_{\alpha}$ is a unitary rotation of $\mathsf{A}$ and
905 > $j$ about each axis $\alpha$.  As all propagations are now
906   unitary and symplectic, the entire integration scheme is also
907   symplectic and time reversible.
908  
909   \section{\label{introSec:layout}Dissertation Layout}
910  
911 < This dissertation is divided as follows:Chapt.~\ref{chapt:RSA}
911 > This dissertation is divided as follows: Ch.~\ref{chapt:RSA}
912   presents the random sequential adsorption simulations of related
913 < pthalocyanines on a gold (111) surface. Ch.~\ref{chapt:OOPSE}
914 < is about the writing of the molecular dynamics simulation package
915 < {\sc oopse}, Ch.~\ref{chapt:lipid} regards the simulations of
916 < phospholipid bilayers using a mesoscale model, and lastly,
913 > pthalocyanines on a gold (111) surface. Ch.~\ref{chapt:oopse}
914 > is about our molecular dynamics simulation package
915 > {\sc oopse}. Ch.~\ref{chapt:lipid} regards the simulations of
916 > phospholipid bilayers using a mesoscale model. And lastly,
917   Ch.~\ref{chapt:conclusion} concludes this dissertation with a
918   summary of all results. The chapters are arranged in chronological
919   order, and reflect the progression of techniques I employed during my
920   research.  
921  
922 < The chapter concerning random sequential adsorption
923 < simulations is a study in applying the principles of theoretical
924 < research in order to obtain a simple model capable of explaining the
925 < results.  My advisor, Dr. Gezelter, and I were approached by a
926 < colleague, Dr. Lieberman, about possible explanations for partial
927 < coverage of a gold surface by a particular compound of hers. We
922 > The chapter concerning random sequential adsorption simulations is a
923 > study in applying Statistical Mechanics simulation techniques in order
924 > to obtain a simple model capable of explaining the results.  My
925 > advisor, Dr. Gezelter, and I were approached by a colleague,
926 > Dr. Lieberman, about possible explanations for the partial coverage of
927 > a gold surface by a particular compound synthesized in her group. We
928   suggested it might be due to the statistical packing fraction of disks
929 < on a plane, and set about to simulate this system.  As the events in
929 > on a plane, and set about simulating this system.  As the events in
930   our model were not dynamic in nature, a Monte Carlo method was
931   employed.  Here, if a molecule landed on the surface without
932 < overlapping another, then its landing was accepted.  However, if there
932 > overlapping with another, then its landing was accepted.  However, if there
933   was overlap, the landing we rejected and a new random landing location
934   was chosen.  This defined our acceptance rules and allowed us to
935   construct a Markov chain whose limiting distribution was the surface
# Line 792 | Line 937 | written in order to study phospholipid bilayer.  Altho
937  
938   The following chapter, about the simulation package {\sc oopse},
939   describes in detail the large body of scientific code that had to be
940 < written in order to study phospholipid bilayer.  Although there are
940 > written in order to study phospholipid bilayers.  Although there are
941   pre-existing molecular dynamic simulation packages available, none
942 < were capable of implementing the models we were developing.{\sc oopse}
943 < is a unique package capable of not only integrating the equations of
944 < motion in Cartesian space, but is also able to integrate the
945 < rotational motion of rigid bodies and dipoles.  Add to this the
946 < ability to perform calculations across parallel processors and a
947 < flexible script syntax for creating systems, and {\sc oopse} becomes a
948 < very powerful scientific instrument for the exploration of our model.
942 > were capable of implementing the models we were developing. {\sc
943 > oopse} is a unique package capable of not only integrating the
944 > equations of motion in Cartesian space, but is also able to integrate
945 > the rotational motion of rigid bodies and dipoles.  It also has the
946 > ability to perform calculations across distributed parallel processors
947 > and contains a flexible script syntax for creating systems.  {\sc
948 > oopse} has become a very powerful scientific instrument for the
949 > exploration of our model.
950  
951 < Bringing us to Ch.~\ref{chapt:lipid}. Using {\sc oopse}, I have been
951 > In Ch.~\ref{chapt:lipid}, utilizing {\sc oopse}, I have been
952   able to parameterize a mesoscale model for phospholipid simulations.
953 < This model retains information about solvent ordering about the
953 > This model retains information about solvent ordering around the
954   bilayer, as well as information regarding the interaction of the
955 < phospholipid head groups' dipole with each other and the surrounding
955 > phospholipid head groups' dipoles with each other and the surrounding
956   solvent.  These simulations give us insight into the dynamic events
957   that lead to the formation of phospholipid bilayers, as well as
958   provide the foundation for future exploration of bilayer phase
959   behavior with this model.  
960  
961 < Which leads into the last chapter, where I discuss future directions
962 < for both{\sc oopse} and this mesoscale model.  Additionally, I will
963 < give a summary of results for this dissertation.
818 <
819 <
961 > In the last chapter, I discuss future directions
962 > for both {\sc oopse} and this mesoscale model.  Additionally, I will
963 > give a summary of the results found in this dissertation.

Diff Legend

Removed lines
+ Added lines
< Changed lines
> Changed lines