--- trunk/mattDisertation/Introduction.tex 2004/03/05 22:16:34 1087 +++ trunk/mattDisertation/Introduction.tex 2004/03/07 04:01:29 1088 @@ -5,32 +5,33 @@ Carlo. Molecular Dynamics simulations integrate the eq The techniques used in the course of this research fall under the two main classes of molecular simulation: Molecular Dynamics and Monte -Carlo. Molecular Dynamics simulations integrate the equations of -motion for a given system of particles, allowing the researcher to -gain insight into the time dependent evolution of a system. Diffusion -phenomena are readily studied with this simulation technique, making -Molecular Dynamics the main simulation technique used in this -research. Other aspects of the research fall under the Monte Carlo -class of simulations. In Monte Carlo, the configuration space -available to the collection of particles is sampled stochastically, or -randomly. Each configuration is chosen with a given probability based -on the Maxwell Boltzmann distribution. These types of simulations are -best used to probe properties of a system that are dependent only on -the state of the system. Structural information about a system is most -readily obtained through these types of methods. +Carlo. Molecular Dynamics simulations are carried out by integrating +the equations of motion for a given system of particles, allowing the +researcher to gain insight into the time dependent evolution of a +system. Transport phenomena are readily studied with this simulation +technique, making Molecular Dynamics the main simulation technique +used in this research. Other aspects of the research fall in the +Monte Carlo class of simulations. In Monte Carlo, the configuration +space available to the collection of particles is sampled +stochastically, or randomly. Each configuration is chosen with a given +probability based on the Boltzmann distribution. These types +of simulations are best used to probe properties of a system that are +dependent only on the state of the system. Structural information +about a system is most readily obtained through these types of +methods. Although the two techniques employed seem dissimilar, they are both linked by the overarching principles of Statistical -Mechanics. Statistical Meachanics governs the behavior of +Mechanics. Statistical Mechanics governs the behavior of both classes of simulations and dictates what each method can and -cannot do. When investigating a system, one most first analyze what -thermodynamic properties of the system are being probed, then chose +cannot do. When investigating a system, one must first analyze what +thermodynamic properties of the system are being probed, then choose which method best suits that objective. \section{\label{introSec:statThermo}Statistical Mechanics} The following section serves as a brief introduction to some of the -Statistical Mechanics concepts present in this dissertation. What +Statistical Mechanics concepts presented in this dissertation. What follows is a brief derivation of Boltzmann weighted statistics and an explanation of how one can use the information to calculate an observable for a system. This section then concludes with a brief @@ -124,18 +125,20 @@ Showing that the partitioning of energy between the tw \beta( E_{\gamma} ) = \beta( E_{\text{bath}} ) \label{introEq:SM8} \end{equation} -Showing that the partitioning of energy between the two systems is -actually a process of thermal equilibration.\cite{Frenkel1996} +Eq.~\ref{introEq:SM8} shows that the partitioning of energy between +the two systems is actually a process of thermal +equilibration.\cite{Frenkel1996} -An application of these results is to formulate the form of an -expectation value of an observable, $A$, in the canonical ensemble. In -the canonical ensemble, the number of particles, $N$, the volume, $V$, -and the temperature, $T$, are all held constant while the energy, $E$, -is allowed to fluctuate. Returning to the previous example, the bath +An application of these results is to formulate the expectation value +of an observable, $A$, in the canonical ensemble. In the canonical +ensemble, the number of particles, $N$, the volume, $V$, and the +temperature, $T$, are all held constant while the energy, $E$, is +allowed to fluctuate. Returning to the previous example, the bath system is now an infinitely large thermal bath, whose exchange of energy with the system $\gamma$ holds the temperature constant. The -partitioning of energy in the bath system is then related to the total -energy of both systems and the fluctuations in $E_{\gamma}$: +partitioning of energy between the bath and the system is then related +to the total energy of both systems and the fluctuations in +$E_{\gamma}$: \begin{equation} \Omega( E_{\gamma} ) = \Omega( E - E_{\gamma} ) \label{introEq:SM9} @@ -148,9 +151,9 @@ an integration over all accessible phase space, $P_{\g \label{introEq:SM10} \end{equation} Where $\int\limits_{\boldsymbol{\Gamma}} d\boldsymbol{\Gamma}$ denotes -an integration over all accessible phase space, $P_{\gamma}$ is the -probability of being in a given phase state and -$A(\boldsymbol{\Gamma})$ is some observable that is a function of the +an integration over all accessible points in phase space, $P_{\gamma}$ +is the probability of being in a given phase state and +$A(\boldsymbol{\Gamma})$ is an observable that is a function of the phase state. Because entropy seeks to maximize the number of degenerate states at a @@ -206,9 +209,9 @@ Where the value of an observable is averaged over the \int_0^{\tau} A[\boldsymbol{\Gamma}(t)]\,dt \label{introEq:SM16} \end{equation} -Where the value of an observable is averaged over the length of time -that the simulation is run. This type of measurement mirrors the -experimental measurement of an observable. In an experiment, the +Where the value of an observable is averaged over the length of time, +$\tau$, that the simulation is run. This type of measurement mirrors +the experimental measurement of an observable. In an experiment, the instrument analyzing the system must average its observation over the finite time of the measurement. What is required then, is a principle to relate the time average to the ensemble average. This is the @@ -221,7 +224,7 @@ assumed to visit all appropriately available points in (\emph{e.g.}~glasses). When valid, ergodicity allows the unification of a time averaged observation and an ensemble averaged one. If an observation is averaged over a sufficiently long time, the system is -assumed to visit all appropriately available points in phase space, +assumed to visit all energetically accessible points in phase space, giving a properly weighted statistical average. This allows the researcher freedom of choice when deciding how best to measure a given observable. When an ensemble averaged approach seems most logical, @@ -234,15 +237,14 @@ named, because it heavily uses random numbers in its The Monte Carlo method was developed by Metropolis and Ulam for their work in fissionable material.\cite{metropolis:1949} The method is so -named, because it heavily uses random numbers in its -solution.\cite{allen87:csl} The Monte Carlo method allows for the -solution of integrals through the stochastic sampling of the values -within the integral. In the simplest case, the evaluation of an -integral would follow a brute force method of -sampling.\cite{Frenkel1996} Consider the following single dimensional -integral: +named, because it uses random numbers extensively.\cite{allen87:csl} +The Monte Carlo method allows for the solution of integrals through +the stochastic sampling of the values within the integral. In the +simplest case, the evaluation of an integral would follow a brute +force method of sampling.\cite{Frenkel1996} Consider the following +single dimensional integral: \begin{equation} -I = f(x)dx +I = \int_a^b f(x)dx \label{eq:MCex1} \end{equation} The equation can be recast as: @@ -371,7 +373,7 @@ will only yield the limiting distribution again. distribution, and successive applications of the transition matrix will only yield the limiting distribution again. \begin{equation} -\boldsymbol{\rho}_{\text{limit}} = \boldsymbol{\rho}_{\text{initial}} +\boldsymbol{\rho}_{\text{limit}} = \boldsymbol{\rho}_{\text{limit}} \boldsymbol{\Pi} \label{introEq:MCmarkovEquil} \end{equation} @@ -383,7 +385,7 @@ distribution. Meaning, that at equilibrium the probab $\boldsymbol{\rho}_{\text{limit}}$ matches the Boltzmann distribution of states. The method accomplishes this by imposing the strong condition of microscopic reversibility on the equilibrium -distribution. Meaning, that at equilibrium the probability of going +distribution. This means that at equilibrium, the probability of going from $m$ to $n$ is the same as going from $n$ to $m$. \begin{equation} \rho_m\pi_{mn} = \rho_n\pi_{nm} @@ -418,7 +420,7 @@ Metropolis method proceeds as follows \begin{enumerate} \item Generate an initial configuration $\mathbf{r}^N$ which has some finite probability in $\rho_{kT}$. \item Modify $\mathbf{r}^N$, to generate configuration $\mathbf{r^{\prime}}^N$. -\item If the new configuration lowers the energy of the system, accept the move with unity ($\mathbf{r}^N$ becomes $\mathbf{r^{\prime}}^N$). Otherwise accept with probability $e^{-\beta \Delta \mathcal{U}}$. +\item If the new configuration lowers the energy of the system, accept the move ($\mathbf{r}^N$ becomes $\mathbf{r^{\prime}}^N$). Otherwise accept with probability $e^{-\beta \Delta \mathcal{U}}$. \item Accumulate the average for the configurational observable of interest. \item Repeat from step 2 until the average converges. \end{enumerate} @@ -431,14 +433,14 @@ Molecular Dynamics is when the equations of motion for \section{\label{introSec:MD}Molecular Dynamics Simulations} The main simulation tool used in this research is Molecular Dynamics. -Molecular Dynamics is when the equations of motion for a system are -integrated in order to obtain information about both the positions and -momentum of a system, allowing the calculation of not only -configurational observables, but momenta dependent ones as well: -diffusion constants, relaxation events, folding/unfolding -events, etc. With the time dependent information gained from a -Molecular Dynamics simulation, one can also calculate time correlation -functions of the form\cite{Hansen86} +Molecular Dynamics is the numerical integration of the equations of +motion for a system in order to obtain information about the dynamic +changes in the positions and momentum of the constituent particles. +This allows the calculation of not only configurational observables, +but momentum dependent ones as well: diffusion constants, relaxation +events, folding/unfolding events, etc. With the time dependent +information gained from a Molecular Dynamics simulation, one can also +calculate time correlation functions of the form\cite{Hansen86} \begin{equation} \langle A(t)\,A(0)\rangle = \lim_{\tau\rightarrow\infty} \frac{1}{\tau} \int_0^{\tau} A(t+t^{\prime})\,A(t^{\prime})\,dt^{\prime} @@ -448,23 +450,23 @@ Sec.~\ref{introSec:ergodic}, the average of these obse of a system, such as diffusion constants from the velocity autocorrelation or dipole relaxation times from the dipole autocorrelation. Due to the principle of ergodicity, -Sec.~\ref{introSec:ergodic}, the average of these observables over the -time period of the simulation are taken to be the ensemble averages -for the system. +Sec.~\ref{introSec:ergodic}, the average of static observables over +the length of the simulation are taken to be the ensemble averages for +the system. The choice of when to use molecular dynamics over Monte Carlo techniques, is normally decided by the observables in which the -researcher is interested. If the observables depend on time in -any fashion, then the only choice is molecular dynamics in some form. +researcher is interested. If the observables depend on time in any +fashion, then the only choice is molecular dynamics in some form. However, when the observable is dependent only on the configuration, -then for most small systems, Monte Carlo techniques will be more efficient. +then for most small systems, Monte Carlo techniques will be more +efficient. The focus of research in the second half of this +dissertation is centered on the dynamic properties of phospholipid +bilayers, making molecular dynamics key in the simulation of those +properties. -The focus of research in the second half of this dissertation is -centered around the dynamic properties of phospholipid bilayers, -making molecular dynamics key in the simulation of those properties. +\subsection{\label{introSec:mdAlgorithm}Molecular Dynamics Algorithms} -\subsection{\label{introSec:mdAlgorithm}The Molecular Dynamics Algorithm} - To illustrate how the molecular dynamics technique is applied, the following sections will describe the sequence involved in a simulation. Sec.~\ref{introSec:mdInit} deals with the initialization @@ -480,59 +482,59 @@ water, and in other cases structured the lipids into p Ch.~\ref{chapt:lipid} deals with the formation and equilibrium dynamics of phospholipid membranes. Therefore in these simulations initial positions were selected that in some cases dispersed the lipids in -water, and in other cases structured the lipids into performed +water, and in other cases structured the lipids into preformed bilayers. Important considerations at this stage of the simulation are: \begin{itemize} -\item There are no major overlaps of molecular or atomic orbitals -\item Velocities are chosen in such a way as to not give the system a non=zero total momentum or angular momentum. -\item It is also sometimes desirable to select the velocities to correctly sample the target temperature. +\item There are no overlapping molecules or atoms. +\item Velocities are chosen in such a way as to give the system zero total momentum and angular momentum. +\item It is also sometimes desirable to select the velocities to correctly sample the ensemble at a particular target temperature. \end{itemize} -The first point is important due to the amount of potential energy -generated by having two particles too close together. If overlap -occurs, the first evaluation of forces will return numbers so large as -to render the numerical integration of the motion meaningless. The -second consideration keeps the system from drifting or rotating as a -whole. This arises from the fact that most simulations are of systems -in equilibrium in the absence of outside forces. Therefore any net +The first point is important due to the forces generated when two +particles are too close together. If overlap occurs, the first +evaluation of forces will return numbers so large as to render the +numerical integration of the equations motion meaningless. The second +consideration keeps the system from drifting or rotating as a whole. +This arises from the fact that most simulations are of systems in +equilibrium in the absence of outside forces. Therefore any net movement would be unphysical and an artifact of the simulation method used. The final point addresses the selection of the magnitude of the -initial velocities. For many simulations it is convenient to use -this opportunity to scale the amount of kinetic energy to reflect the +initial velocities. For many simulations it is convenient to use this +opportunity to scale the amount of kinetic energy to reflect the desired thermal distribution of the system. However, it must be noted -that most systems will require further velocity rescaling after the -first few initial simulation steps due to either loss or gain of -kinetic energy from energy stored in potential degrees of freedom. +that due to equipartition, most systems will require further velocity +rescaling after the first few initial simulation steps due to either +loss or gain of kinetic energy from energy stored as potential energy. \subsection{\label{introSec:mdForce}Force Evaluation} The evaluation of forces is the most computationally expensive portion -of a given molecular dynamics simulation. This is due entirely to the -evaluation of long range forces in a simulation, typically pair-wise. -These forces are most commonly the Van der Waals force, and sometimes -Coulombic forces as well. For a pair-wise force, there are $N(N-1)/ 2$ -pairs to be evaluated, where $N$ is the number of particles in the -system. This leads to the calculations scaling as $N^2$, making large -simulations prohibitive in the absence of any computation saving -techniques. +of any molecular dynamics simulation. This is due entirely to the +evaluation of long range forces in a simulation, which are typically +pair potentials. These forces are most commonly the van der Waals +force, and sometimes Coulombic forces as well. For a pair-wise +interaction, there are $N(N-1)/ 2$ pairs to be evaluated, where $N$ is +the number of particles in the system. This leads to calculations which +scale as $N^2$, making large simulations prohibitive in the absence +of any computation saving techniques. -Another consideration one must resolve, is that in a given simulation +Another consideration one must resolve, is that in a given simulation, a disproportionate number of the particles will feel the effects of the surface.\cite{allen87:csl} For a cubic system of 1000 particles arranged in a $10 \times 10 \times 10$ cube, 488 particles will be -exposed to the surface. Unless one is simulating an isolated particle -group in a vacuum, the behavior of the system will be far from the -desired bulk characteristics. To offset this, simulations employ the -use of periodic boundary images.\cite{born:1912} +exposed to the surface. Unless one is simulating an isolated cluster +in a vacuum, the behavior of the system will be far from the desired +bulk characteristics. To offset this, simulations employ the use of +periodic images.\cite{born:1912} The technique involves the use of an algorithm that replicates the -simulation box on an infinite lattice in Cartesian space. Any given +simulation box on an infinite lattice in Cartesian space. Any particle leaving the simulation box on one side will have an image of -itself enter on the opposite side (see Fig.~\ref{introFig:pbc}). In -addition, this sets that any two particles have an image, real or -periodic, within $\text{box}/2$ of each other. A discussion of the -method used to calculate the periodic image can be found in -Sec.\ref{oopseSec:pbc}. +itself enter on the opposite side (see +Fig.~\ref{introFig:pbc}). Therefore, a pair of particles have an +image of each other, real or periodic, within $\text{box}/2$. A +discussion of the method used to calculate the periodic image can be +found in Sec.\ref{oopseSec:pbc}. \begin{figure} \centering @@ -541,7 +543,7 @@ Returning to the topic of the computational scale of t \label{introFig:pbc} \end{figure} -Returning to the topic of the computational scale of the force +Returning to the topic of the computational scaling of the force evaluation, the use of periodic boundary conditions requires that a cutoff radius be employed. Using a cutoff radius improves the efficiency of the force evaluation, as particles farther than a @@ -550,19 +552,19 @@ one image that is seen by another atom, and further th there are two methods to choose from, both with their own cutoff limits. In the minimum image convention, $r_{\text{cut}}$ has a maximum value of $\text{box}/2$. This is because each atom has only -one image that is seen by another atom, and further the image used is -the one that minimizes the distance between the two atoms. A system of -wrapped images about a central atom therefore has a maximum length -scale of box on a side (Fig.~\ref{introFig:rMaxMin}). The second -convention, multiple image convention, has a maximum $r_{\text{cut}}$ -of box. Here multiple images of each atom are replicated in the +one image that is seen by other atoms. The image used is the one that +minimizes the distance between the two atoms. A system of wrapped +images about a central atom therefore has a maximum length scale of +box on a side (Fig.~\ref{introFig:rMaxMin}). The second convention, +multiple image convention, has a maximum $r_{\text{cut}}$ of the box +length itself. Here multiple images of each atom are replicated in the periodic cells surrounding the central atom, this causes the atom to see multiple copies of several atoms. If the cutoff radius is larger than box, however, then the atom will see an image of itself, and attempt to calculate an unphysical self-self force interaction (Fig.~\ref{introFig:rMaxMult}). Due to the increased complexity and commputaional ineffeciency of the multiple image method, the minimum -image method is the periodic method used throughout this research. +image method is the used throughout this research. \begin{figure} \centering @@ -574,11 +576,11 @@ image method is the periodic method used throughout th \begin{figure} \centering \includegraphics[width=\linewidth]{rCutMaxMultFig.eps} -\caption[An explanation of multiple image convention]{The yellow atom is the central wrapping point. The blue atoms are the minimum images of the system about the central atom. The boxes with the green atoms are multiple images of the central box. If $r_{\text{cut}} \geq \{text{box}$ then the central atom sees multiple images of itself (red atom), creating a self-self force evaluation.} +\caption[An explanation of multiple image convention]{The yellow atom is the central wrapping point. The blue atoms are the minimum images of the system about the central atom. The boxes with the green atoms are multiple images of the central box. If $r_{\text{cut}} \geq \text{box}$ then the central atom sees multiple images of itself (red atom), creating a self-self force evaluation.} \label{introFig:rMaxMult} \end{figure} -With the use of an $r_{\text{cut}}$, however, comes a discontinuity in +With the use of a cutoff radius, however, comes a discontinuity in the potential energy curve (Fig.~\ref{introFig:shiftPot}). To fix this discontinuity, one calculates the potential energy at the $r_{\text{cut}}$, and adds that value to the potential, causing @@ -594,31 +596,31 @@ neighbor list. \cite{allen87:csl} In the Verlet method \end{figure} The second main simplification used in this research is the Verlet -neighbor list. \cite{allen87:csl} In the Verlet method, one generates +neighbor list.\cite{allen87:csl} In the Verlet method, one generates a list of all neighbor atoms, $j$, surrounding atom $i$ within some cutoff $r_{\text{list}}$, where $r_{\text{list}}>r_{\text{cut}}$. This list is created the first time forces are evaluated, then on subsequent force evaluations, pair calculations are only calculated -from the neighbor lists. The lists are updated if any given particle +from the neighbor lists. The lists are updated if any particle in the system moves farther than $r_{\text{list}}-r_{\text{cut}}$, -giving rise to the possibility that a particle has left or joined a +which indeicates the possibility that a particle has left or joined the neighbor list. -\subsection{\label{introSec:mdIntegrate} Integration of the equations of motion} +\subsection{\label{introSec:mdIntegrate} Integration of the Equations of Motion} A starting point for the discussion of molecular dynamics integrators is the Verlet algorithm.\cite{Frenkel1996} It begins with a Taylor expansion of position in time: \begin{equation} q(t+\Delta t)= q(t) + v(t)\Delta t + \frac{F(t)}{2m}\Delta t^2 + - \frac{\Delta t^3}{3!}\frac{\partial q(t)}{\partial t} + + \frac{\Delta t^3}{3!}\frac{\partial^3 q(t)}{\partial t^3} + \mathcal{O}(\Delta t^4) \label{introEq:verletForward} \end{equation} As well as, \begin{equation} q(t-\Delta t)= q(t) - v(t)\Delta t + \frac{F(t)}{2m}\Delta t^2 - - \frac{\Delta t^3}{3!}\frac{\partial q(t)}{\partial t} + + \frac{\Delta t^3}{3!}\frac{\partial^3 q(t)}{\partial t^3} + \mathcal{O}(\Delta t^4) \label{introEq:verletBack} \end{equation} @@ -641,7 +643,7 @@ with a velocity reformulation of the Verlet method.\ci order of $\Delta t^4$. In practice, however, the simulations in this research were integrated -with a velocity reformulation of the Verlet method.\cite{allen87:csl} +with the velocity reformulation of the Verlet method.\cite{allen87:csl} \begin{align} q(t+\Delta t) &= q(t) + v(t)\Delta t + \frac{F(t)}{2m}\Delta t^2 % \label{introEq:MDvelVerletPos} \\% @@ -652,16 +654,16 @@ very little long term drift in energy conservation. E The original Verlet algorithm can be regained by substituting the velocity back into Eq.~\ref{introEq:MDvelVerletPos}. The Verlet formulations are chosen in this research because the algorithms have -very little long term drift in energy conservation. Energy -conservation in a molecular dynamics simulation is of extreme -importance, as it is a measure of how closely one is following the -``true'' trajectory with the finite integration scheme. An exact -solution to the integration will conserve area in phase space, as well +very little long time drift in energy. Energy conservation in a +molecular dynamics simulation is of extreme importance, as it is a +measure of how closely one is following the ``true'' trajectory with +the finite integration scheme. An exact solution to the integration +will conserve area in phase space (i.e.~will be symplectic), as well as be reversible in time, that is, the trajectory integrated forward or backwards will exactly match itself. Having a finite algorithm that both conserves area in phase space and is time reversible, -therefore increases, but does not guarantee the ``correctness'' or the -integrated trajectory. +therefore increases the reliability, but does not guarantee the +``correctness'' of the integrated trajectory. It can be shown,\cite{Frenkel1996} that although the Verlet algorithm does not rigorously preserve the actual Hamiltonian, it does preserve @@ -677,8 +679,8 @@ parameters are needed to describe the motions. The si \subsection{\label{introSec:MDfurther}Further Considerations} In the simulations presented in this research, a few additional -parameters are needed to describe the motions. The simulations -involving water and phospholipids in Ch.~\ref{chapt:lipid} are +parameters are needed to describe the motions. In the simulations +involving water and phospholipids in Ch.~\ref{chapt:lipid}, we are required to integrate the equations of motions for dipoles on atoms. This involves an additional three parameters be specified for each dipole atom: $\phi$, $\theta$, and $\psi$. These three angles are @@ -686,10 +688,10 @@ accumulated into a single $3 \times 3$ matrix, $\mathb $z$-axis, and $\theta$ is a rotation about the new $x$-axis, and $\psi$ is a final rotation about the new $z$-axis (see Fig.~\ref{introFig:eulerAngles}). This sequence of rotations can be -accumulated into a single $3 \times 3$ matrix, $\mathbf{A}$, +accumulated into a single $3 \times 3$ matrix, $\mathsf{A}$, defined as follows: \begin{equation} -\mathbf{A} = +\mathsf{A} = \begin{bmatrix} \cos\phi\cos\psi-\sin\phi\cos\theta\sin\psi &% \sin\phi\cos\psi+\cos\phi\cos\theta\sin\psi &% @@ -728,13 +730,13 @@ Where $\omega^s_i$ is the angular velocity in the lab \omega^s_y \frac{\cos\phi}{\sin\theta} \label{introEq:MDeulerPsi} \end{align} -Where $\omega^s_i$ is the angular velocity in the lab space frame -along Cartesian coordinate $i$. However, a difficulty arises when +Where $\omega^s_{\alpha}$ is the angular velocity in the lab space frame +along Cartesian coordinate $\alpha$. However, a difficulty arises when attempting to integrate Eq.~\ref{introEq:MDeulerPhi} and Eq.~\ref{introEq:MDeulerPsi}. The $\frac{1}{\sin \theta}$ present in both equations means there is a non-physical instability present when $\theta$ is 0 or $\pi$. To correct for this, the simulations integrate -the rotation matrix, $\mathbf{A}$, directly, thus avoiding the +the rotation matrix, $\mathsf{A}$, directly, thus avoiding the instability. This method was proposed by Dullweber \emph{et. al.}\cite{Dullweber1997}, and is presented in Sec.~\ref{introSec:MDsymplecticRot}. @@ -743,11 +745,12 @@ scheme. It has been previously Before discussing the integration of the rotation matrix, it is necessary to understand the construction of a ``good'' integration -scheme. It has been previously -discussed(Sec.~\ref{introSec:mdIntegrate}) how it is desirable for an -integrator to be symplectic, or time reversible. The following is an +scheme. It has been previously discussed +(Sec.~\ref{introSec:mdIntegrate}) how it is desirable for an +integrator to be symplectic and time reversible. The following is an outline of the Trotter factorization of the Liouville Propagator as a -scheme for generating symplectic integrators.\cite{Tuckerman92} +scheme for generating symplectic, time-reversible +integrators.\cite{Tuckerman92} For a system with $f$ degrees of freedom the Liouville operator can be defined as, @@ -797,7 +800,7 @@ unitary. Meaning an integrator based on this factoriz \label{introEq:Lp5} \end{align} Because $U_1(t)$ and $U_2(t)$ are unitary, $G(\Delta t)$ is also -unitary. Meaning an integrator based on this factorization will be +unitary. This means that an integrator based on this factorization will be reversible in time. As an example, consider the following decomposition of $L$: @@ -860,28 +863,28 @@ symplectic propagation of the rotation matrix, $\mathb Based on the factorization from the previous section, Dullweber\emph{et al}.\cite{Dullweber1997}~ proposed a scheme for the -symplectic propagation of the rotation matrix, $\mathbf{A}$, as an +symplectic propagation of the rotation matrix, $\mathsf{A}$, as an alternative method for the integration of orientational degrees of freedom. The method starts with a straightforward splitting of the Liouville operator: \begin{align} iL_{\text{pos}} &= \dot{q}\frac{\partial}{\partial q} + - \mathbf{\dot{A}}\frac{\partial}{\partial \mathbf{A}} + \mathsf{\dot{A}}\frac{\partial}{\partial \mathsf{A}} \label{introEq:SR1a} \\% % iL_F &= F(q)\frac{\partial}{\partial p} \label{introEq:SR1b} \\% -iL_{\tau} &= \tau(\mathbf{A})\frac{\partial}{\partial \pi} +iL_{\tau} &= \tau(\mathsf{A})\frac{\partial}{\partial j} \label{introEq:SR1b} \\% \end{align} -Where $\tau(\mathbf{A})$ is the torque of the system -due to the configuration, and $\pi$ is the conjugate +Where $\tau(\mathsf{A})$ is the torque of the system +due to the configuration, and $j$ is the conjugate angular momenta of the system. The propagator, $G(\Delta t)$, becomes \begin{equation} G(\Delta t) = e^{\frac{\Delta t}{2} F(q)\frac{\partial}{\partial p}} \, - e^{\frac{\Delta t}{2} \tau(\mathbf{A})\frac{\partial}{\partial \pi}} \, + e^{\frac{\Delta t}{2} \tau(\mathsf{A})\frac{\partial}{\partial j}} \, e^{\Delta t\,iL_{\text{pos}}} \, - e^{\frac{\Delta t}{2} \tau(\mathbf{A})\frac{\partial}{\partial \pi}} \, + e^{\frac{\Delta t}{2} \tau(\mathsf{A})\frac{\partial}{\partial j}} \, e^{\frac{\Delta t}{2} F(q)\frac{\partial}{\partial p}} \label{introEq:SR2} \end{equation} @@ -899,17 +902,17 @@ Where $\mathcal{U}_j$ is a unitary rotation of $\mathb \mathcal{U}_x \biggl(\frac{\Delta t}{2}\biggr)\, \label{introEq:SR3} \end{equation} -Where $\mathcal{U}_j$ is a unitary rotation of $\mathbf{A}$ and -$\pi$ about each axis $j$. As all propagations are now +Where $\mathcal{U}_{\alpha}$ is a unitary rotation of $\mathsf{A}$ and +$j$ about each axis $\alpha$. As all propagations are now unitary and symplectic, the entire integration scheme is also symplectic and time reversible. \section{\label{introSec:layout}Dissertation Layout} -This dissertation is divided as follows:Ch.~\ref{chapt:RSA} +This dissertation is divided as follows: Ch.~\ref{chapt:RSA} presents the random sequential adsorption simulations of related pthalocyanines on a gold (111) surface. Ch.~\ref{chapt:oopse} -is about the writing of the molecular dynamics simulation package +is about our molecular dynamics simulation package {\sc oopse}. Ch.~\ref{chapt:lipid} regards the simulations of phospholipid bilayers using a mesoscale model. And lastly, Ch.~\ref{chapt:conclusion} concludes this dissertation with a @@ -921,31 +924,32 @@ Dr. Lieberman, about possible explanations for the pa study in applying Statistical Mechanics simulation techniques in order to obtain a simple model capable of explaining the results. My advisor, Dr. Gezelter, and I were approached by a colleague, -Dr. Lieberman, about possible explanations for the partial coverage of a -gold surface by a particular compound of hers. We suggested it might -be due to the statistical packing fraction of disks on a plane, and -set about to simulate this system. As the events in our model were -not dynamic in nature, a Monte Carlo method was employed. Here, if a -molecule landed on the surface without overlapping another, then its -landing was accepted. However, if there was overlap, the landing we -rejected and a new random landing location was chosen. This defined -our acceptance rules and allowed us to construct a Markov chain whose -limiting distribution was the surface coverage in which we were -interested. +Dr. Lieberman, about possible explanations for the partial coverage of +a gold surface by a particular compound synthesized in her group. We +suggested it might be due to the statistical packing fraction of disks +on a plane, and set about simulating this system. As the events in +our model were not dynamic in nature, a Monte Carlo method was +employed. Here, if a molecule landed on the surface without +overlapping with another, then its landing was accepted. However, if there +was overlap, the landing we rejected and a new random landing location +was chosen. This defined our acceptance rules and allowed us to +construct a Markov chain whose limiting distribution was the surface +coverage in which we were interested. The following chapter, about the simulation package {\sc oopse}, describes in detail the large body of scientific code that had to be written in order to study phospholipid bilayers. Although there are pre-existing molecular dynamic simulation packages available, none -were capable of implementing the models we were developing.{\sc oopse} -is a unique package capable of not only integrating the equations of -motion in Cartesian space, but is also able to integrate the -rotational motion of rigid bodies and dipoles. Add to this the -ability to perform calculations across parallel processors and a -flexible script syntax for creating systems, and {\sc oopse} becomes a -very powerful scientific instrument for the exploration of our model. +were capable of implementing the models we were developing. {\sc +oopse} is a unique package capable of not only integrating the +equations of motion in Cartesian space, but is also able to integrate +the rotational motion of rigid bodies and dipoles. It also has the +ability to perform calculations across distributed parallel processors +and contains a flexible script syntax for creating systems. {\sc +oopse} has become a very powerful scientific instrument for the +exploration of our model. -Bringing us to Ch.~\ref{chapt:lipid}. Using {\sc oopse}, I have been +In Ch.~\ref{chapt:lipid}, utilizing {\sc oopse}, I have been able to parameterize a mesoscale model for phospholipid simulations. This model retains information about solvent ordering around the bilayer, as well as information regarding the interaction of the @@ -955,8 +959,8 @@ Which leads into the last chapter, where I discuss fut provide the foundation for future exploration of bilayer phase behavior with this model. -Which leads into the last chapter, where I discuss future directions -for both{\sc oopse} and this mesoscale model. Additionally, I will -give a summary of results for this dissertation. +In the last chapter, I discuss future directions +for both {\sc oopse} and this mesoscale model. Additionally, I will +give a summary of the results found in this dissertation.