The distributed-computing project Folding@home uses scientific computer programs, referred to as "cores" or "fahcores", to perform calculations.[1][2] Folding@home's cores are based on modified and optimized versions of molecular simulation programs for calculation, including TINKER, GROMACS, AMBER, CPMD, SHARPEN, ProtoMol and Desmond.[1][3][4] These variants are each given an arbitrary identifier (Core xx). While the same core can be used by various versions of the client, separating the core from the client enables the scientific methods to be updated automatically as needed without a client update.[1]
These cores listed below are currently used by the project.[1]
GROMACS
- Core a7
- Core a8
- Available for Windows, Linux, macOS and ARM, uses Gromacs 2020.5 [6]
GPU
Cores for the Graphics Processing Unit use the graphics chip of modern video cards to do molecular dynamics. The GPU Gromacs core is not a true port of Gromacs, but rather key elements from Gromacs were taken and enhanced for GPU capabilities.[7]
GPU3
These are the third generation GPU cores, and are based on OpenMM, Pande Group's own open library for molecular simulation. Although based on the GPU2 code, this adds stability and new capabilities.[8]
- core 22 (last core to use old style numbering convention)
- v0.0.16 Available to Windows and Linux for AMD and NVIDIA GPUs using OpenCL and CUDA, if available. It uses OpenMM 7.5.1
- v0.0.17 Available to Windows and Linux for AMD and NVIDIA GPUs using OpenCL and CUDA, if available. It uses OpenMM 7.5.1
- v0.0.18 Available to Windows and Linux for AMD and NVIDIA GPUs using OpenCL and CUDA, if available. It uses OpenMM 7.6.0 [9]
- v0.0.20 Available to Windows and Linux for AMD and NVIDIA GPUs using OpenCL and CUDA, if available. It uses OpenMM 7.7.0, which provides performance improvements and many new science features [10]
- core 23
- v8.0.3 Available to Windows and Linux for AMD and NVIDIA GPUs using OpenCL and CUDA, if available. It uses OpenMM 8.0.0, which provides performance improvements, particularly to CUDA, and many new science features [11]
- core 24
- v8.1.3 Available to Windows and Linux for AMD and NVIDIA GPUs using OpenCL and CUDA, if available. It uses OpenMM 8.1.1, which includes some major bug fixes. Awaiting announcement.
These cores are not currently used by the project, as they are either retired due to becoming obsolete, or are not yet ready for general release.[1]
TINKER
TINKER is a computer software application for molecular dynamics simulation with a complete and general package for molecular mechanics and molecular dynamics, with some special features for biopolymers.[12]
- Tinker core (Core 65)
- An unoptimized uniprocessor core, this was officially retired as the AMBER and Gromacs cores perform the same tasks much faster. This core was available for Windows, Linux, and Macs.[13]
GROMACS
- GroGPU (Core 10)
- Available for ATI series 1xxx GPUs running under Windows.[14][15] Although mostly Gromacs based, parts of the core were rewritten.[14] This core was retired as of June 6, 2008 due to a move to the second generation of the GPU clients.[14]
- Gro-SMP (Core a1)
- GroCVS (Core a2)
- Available only to x86 Macs and x86/64 Linux, this core is very similar to Core a1, as it uses much of the same core base, including use of MPI. However, this core utilizes more recent Gromacs code, and supports more features such as extra-large work units.[19][20] Officially retired due to move to a threads-based SMP2 client.
- Gro-PS3
- Also known as the SCEARD core, this variant was for the PlayStation 3 game system,[21][22] which supported a Folding@Home client until it was retired in November 2012. This core performed implicit solvation calculations like the GPU cores, but was also capable of running explicit solvent calculations like the CPU cores, and took the middle ground between the inflexible high-speed GPU cores and flexible low-speed CPU cores.[23] This core used SPE cores for optimization, but did not support SIMD.
- Gromacs (Core 78)
- This is the original Gromacs core,[16] and is currently available for uniprocessor clients only, supporting Windows, Linux, and macOS.[24]
- Gromacs 33 (Core a0)
- Available to Windows, Linux, and macOS uniprocessor clients only, this core uses the Gromacs 3.3 codebase, which allowing a broader range of simulations to be run.[16][25]
- Gromacs SREM (Core 80)
- This core uses the Serial Replica Exchange Method, which is also known as REMD (Replica Exchange Molecular Dynamics) or GroST (Gromacs Serial replica exchange with Temperatures) in its simulations, and is available for Windows and Linux uniprocessor clients only.[16][26][27]
- GroSimT (Core 81)
- This core performs simulated tempering, of which the basic idea is to enhance sampling by periodically raising and lowering temperature. This may allow Folding@home to more efficiently sample the transitions between folded and unfolded conformations of proteins.[16] Available for Windows and Linux uniprocessor clients only.[28]
- DGromacs (Core 79)
- Available for uniprocessor clients, this core uses SSE2 processor optimization where supported and is capable of running on Windows, Linux, and macOS.[16][29]
- DGromacsB (Core 7b)
- Distinct from Core 79 in that it has several scientific additions.[16] Initially released only to the Linux platform in August 2007, it will eventually be available for all platforms.[30]
- DGromacsC (Core 7c)
- Very similar to Core 79, and initially released for Linux and Windows in April 2008 for Windows, Linux, and macOS uniprocessor clients.[31]
- GB Gromacs (Core 7a)
- Available solely for all uniprocessor clients on Windows, Linux, and macOS.[1][16][32]
- GB Gromacs (Core a4)
- Available for Windows, Linux,[33] and macOS,[34] this core was originally released in early October 2010,[35] and as of February 2010 uses the latest version of Gromacs, v4.5.3.[33]
- SMP2 (Core a3)
- The next generation of the SMP cores, this core uses threads instead of MPI for inter-process communication, and is available for Windows, Linux, and macOS.[36][37]
- SMP2 bigadv (Core a5)
- Similar to a3, but this core is specifically designed to run larger-than-normal simulations.[38][39]
- SMP2 bigadv (Core a6)
- A newer version of the a5 core.
CPMD
Short for Car–Parrinello Molecular Dynamics, this core performs ab-initio quantum mechanical molecular dynamics. Unlike classical molecular dynamics calculations which use a force field approach, CPMD includes the motion of electrons in the calculations of energy, forces and motion.[40][41]
Quantum chemical calculations have the possibility to yield a very reliable potential energy surface, and can naturally incorporate multi-body interactions.[41]
- QMD (Core 96)
- This is a double-precision[41] variant for Windows and Linux uniprocessor clients.[42] This core is currently "on hold" due to the main QMD developer, Young Min Rhee, graduating in 2006.[41] This core can use a substantial amount of memory, and was only available to machines that chose to "opt in".[41] SSE2 optimization on Intel CPUs is supported.[41] Due to licensing issues involving Intel libraries and SSE2, QMD Work Units were not assigned to AMD CPUs.[41][43]
SHARPEN
- SHARPEN Core[44][45]
- In early 2010 Vijay Pande said "We've put SHARPEN on hold for now. No ETA to give, sorry. Pushing it further depends a lot on the scientific needs at the time."[46] This core uses different format to standard F@H cores, in that there is more than one "Work Unit" (using the normal definition) in each work packet sent to clients.
Desmond
The software for this core was developed at D. E. Shaw Research. Desmond performs high-speed molecular dynamics simulations of biological systems on conventional computer clusters.[47][48][49][50]
The code uses novel parallel algorithms[51]
and numerical techniques[52]
to achieve high performance on platforms containing a large number of processors,[53]
but may also be executed on a single computer. Desmond and its source code are available without cost for non-commercial use by universities and other not-for-profit research institutions.
- Desmond Core
- Possible available for Windows x86 and Linux x86/64,[54] this core is currently in development.[8]
AMBER
Short for Assisted Model Building with Energy Refinement, AMBER is a family of force fields for molecular dynamics, as well as the name for the software package that simulates these force fields.[55] AMBER was originally developed by Peter Kollman at the University of California, San Francisco, and is currently maintained by professors at various universities.[56] The double-precision AMBER core is not currently optimized with SSE nor SSE2,[57][58]
but AMBER is significantly faster than Tinker cores and adds some functionality which cannot be performed using Gromacs cores.[58]
- PMD (Core 82)
- Available for Windows and Linux uniprocessor clients only.[57]
ProtoMol
ProtoMol is an object-oriented, component based, framework for molecular dynamics (MD) simulations. ProtoMol offers high flexibility, easy extendibility and maintenance, and high performance demands, including parallelization.[59] In 2009, the Pande Group was working on a complementary new technique called Normal Mode Langevin Dynamics which had the possibility to greatly speed simulations while maintaining the same accuracy.[8][60]
- ProtoMol Core (Core b4)
- Available to Linux x86/64 and x86 Windows.[61]
GPU
GPU2
These are the second generation GPU cores. Unlike the retired GPU1 cores, these variants are for ATI CAL-enabled 2xxx/3xxx or later series and NVIDIA CUDA-enabled NVIDIA 8xxx or later series GPUs.[62]
- GPU2 (Core 11)
- Available for x86 Windows clients only.[62] Supported until approximately September 1, 2011 due to AMD/ATI dropping support for the utilized Brook programming language and moving to OpenCL. This forced F@h to rewrite its ATI GPU core code in OpenCL, the result of which is Core 16.[63]
- GPU2 (Core 12)
- Available for x86 Windows clients only.[62]
- GPU2 (Core 13)
- Available for x86 Windows clients only.[62]
- GPU2 (Core 14)
- Available for x86 Windows clients only,[62] this core was officially released Mar 02, 2009.[64]
GPU3
These are the third generation GPU cores, and are based on OpenMM, Pande Group's own open library for molecular simulation. Although based on the GPU2 code, this adds stability and new capabilities.[8]
- GPU3 (core 15)
- Available to x86 Windows only.[65]
- GPU3 (core 16)
- Available to x86 Windows only.[65] Released alongside the new v7 client, this is a rewrite of Core 11 in OpenCL.[63]
- GPU3 (core 17)
- Available to Windows and Linux for AMD and NVIDIA GPUs using OpenCL. Much better performance because of OpenMM 5.1[66]
- GPU3 (core 18)
- Available to Windows for AMD and NVIDIA GPUs using OpenCL. This core was developed to address some critical scientific issues in Core17 [67] and uses the latest technology from OpenMM[68] 6.0.1. There are currently issues regarding the stability and performance of this core on some AMD and NVIDIA Maxwell GPUs. This is why assignment of work units running on this core has been temporarily stopped for some GPUs.[69]
- GPU3 (core 21)
- Available to Windows and Linux for AMD and NVIDIA GPUs using OpenCL. It uses OpenMM 6.2 and fixes the Core 18 AMD/NVIDIA performance issues.[70]
Vijay Pande (2005-10-16). "Folding@home with QMD core FAQ" (FAQ). Stanford University. Retrieved 2006-12-03. The site indicates that Folding@home uses a modification of CPMD allowing it to run on the supercluster environment.
"Amber". 2011. Retrieved 2011-08-23.