wal.sgml 44.4 KB
Newer Older
1
<!-- doc/src/sgml/wal.sgml -->
2

3 4 5 6
<chapter id="wal">
 <title>Reliability and the Write-Ahead Log</title>

 <para>
7
  This chapter explains how the Write-Ahead Log is used to obtain
8 9 10 11 12
  efficient, reliable operation.
 </para>

 <sect1 id="wal-reliability">
  <title>Reliability</title>
13

14
  <para>
15
   Reliability is an important property of any serious database
16
   system, and <productname>PostgreSQL</productname> does everything possible to
17 18 19 20 21 22 23 24 25 26
   guarantee reliable operation. One aspect of reliable operation is
   that all data recorded by a committed transaction should be stored
   in a nonvolatile area that is safe from power loss, operating
   system failure, and hardware failure (except failure of the
   nonvolatile area itself, of course).  Successfully writing the data
   to the computer's permanent storage (disk drive or equivalent)
   ordinarily meets this requirement.  In fact, even if a computer is
   fatally damaged, if the disk drives survive they can be moved to
   another computer with similar hardware and all committed
   transactions will remain intact.
27
  </para>
Peter Eisentraut's avatar
Peter Eisentraut committed
28

29
  <para>
30
   While forcing data to the disk platters periodically might seem like
31 32
   a simple operation, it is not. Because disk drives are dramatically
   slower than main memory and CPUs, several layers of caching exist
33 34 35
   between the computer's main memory and the disk platters.
   First, there is the operating system's buffer cache, which caches
   frequently requested disk blocks and combines disk writes. Fortunately,
36
   all operating systems give applications a way to force writes from
37
   the buffer cache to disk, and <productname>PostgreSQL</productname> uses those
38
   features.  (See the <xref linkend="guc-wal-sync-method"/> parameter
39
   to adjust how this is done.)
40
  </para>
41

42
  <para>
43
   Next, there might be a cache in the disk drive controller; this is
44 45
   particularly common on <acronym>RAID</acronym> controller cards. Some of
   these caches are <firstterm>write-through</firstterm>, meaning writes are sent
46
   to the drive as soon as they arrive. Others are
47
   <firstterm>write-back</firstterm>, meaning data is sent to the drive at
48 49 50
   some later time. Such caches can be a reliability hazard because the
   memory in the disk controller cache is volatile, and will lose its
   contents in a power failure.  Better controller cards have
51
   <firstterm>battery-backup units</firstterm> (<acronym>BBU</acronym>s), meaning
52
   the card has a battery that
53 54
   maintains power to the cache in case of system power loss.  After power
   is restored the data will be written to the disk drives.
55
  </para>
Peter Eisentraut's avatar
Peter Eisentraut committed
56

57
  <para>
58
   And finally, most disk drives have caches. Some are write-through
59
   while some are write-back, and the same concerns about data loss
60
   exist for write-back drive caches as for disk controller
61
   caches.  Consumer-grade IDE and SATA drives are particularly likely
62 63
   to have write-back caches that will not survive a power failure.  Many
   solid-state drives (SSD) also have volatile write-back caches.
64 65 66
  </para>

  <para>
67 68
   These caches can typically be disabled; however, the method for doing
   this varies by operating system and drive type:
69 70
  </para>

71 72 73
  <itemizedlist>
    <listitem>
      <para>
74
        On <productname>Linux</productname>, IDE and SATA drives can be queried using
75
        <command>hdparm -I</command>; write caching is enabled if there is
76
        a <literal>*</literal> next to <literal>Write cache</literal>.  <command>hdparm -W 0</command>
77
        can be used to turn off write caching.  SCSI drives can be queried
78
        using <ulink url="http://sg.danny.cz/sg/sdparm.html"><application>sdparm</application></ulink>.
79
        Use <command>sdparm --get=WCE</command> to check
80
        whether the write cache is enabled and <command>sdparm --clear=WCE</command>
81 82 83 84 85 86
        to disable it.
      </para>
    </listitem>

    <listitem>
      <para>
87
        On <productname>FreeBSD</productname>, IDE drives can be queried using
88
        <command>atacontrol</command> and write caching turned off using
89
        <literal>hw.ata.wc=0</literal> in <filename>/boot/loader.conf</filename>;
90 91 92
        SCSI drives can be queried using <command>camcontrol identify</command>,
        and the write cache both queried and changed using
        <command>sdparm</command> when available.
93 94 95 96
      </para>
    </listitem>

    <listitem>
Peter Eisentraut's avatar
Peter Eisentraut committed
97
      <para>
98 99 100
        On <productname>Solaris</productname>, the disk write cache is controlled by
        <command>format -e</command>.
        (The Solaris <acronym>ZFS</acronym> file system is safe with disk write-cache
101 102 103 104 105 106
        enabled because it issues its own disk cache flush commands.)
      </para>
    </listitem>

    <listitem>
      <para>
107 108 109
        On <productname>Windows</productname>, if <varname>wal_sync_method</varname> is
        <literal>open_datasync</literal> (the default), write caching can be disabled
        by unchecking <literal>My Computer\Open\<replaceable>disk drive</replaceable>\Properties\Hardware\Properties\Policies\Enable write caching on the disk</literal>.
110
        Alternatively, set <varname>wal_sync_method</varname> to
111
        <literal>fsync</literal> or <literal>fsync_writethrough</literal>, which prevent
112
        write caching.
113 114 115 116 117
      </para>
    </listitem>

    <listitem>
      <para>
118
        On <productname>macOS</productname>, write caching can be prevented by
119
        setting <varname>wal_sync_method</varname> to <literal>fsync_writethrough</literal>.
120 121 122 123
      </para>
    </listitem>
  </itemizedlist>

124
  <para>
125 126
   Recent SATA drives (those following <acronym>ATAPI-6</acronym> or later)
   offer a drive cache flush command (<command>FLUSH CACHE EXT</command>),
127
   while SCSI drives have long supported a similar command
128 129 130
   <command>SYNCHRONIZE CACHE</command>.  These commands are not directly
   accessible to <productname>PostgreSQL</productname>, but some file systems
   (e.g., <acronym>ZFS</acronym>, <acronym>ext4</acronym>) can use them to flush
131 132
   data to the platters on write-back-enabled drives.  Unfortunately, such
   file systems behave suboptimally when combined with battery-backup unit
133
   (<acronym>BBU</acronym>) disk controllers.  In such setups, the synchronize
134
   command forces all data from the controller cache to the disks,
135
   eliminating much of the benefit of the BBU.  You can run the
136
   <xref linkend="pgtestfsync"/> program to see
137
   if you are affected.  If you are affected, the performance benefits
138
   of the BBU can be regained by turning off write barriers in
139 140
   the file system or reconfiguring the disk controller, if that is
   an option.  If write barriers are turned off, make sure the battery
141
   remains functional; a faulty battery can potentially lead to data loss.
142 143 144 145
   Hopefully file system and disk controller designers will eventually
   address this suboptimal behavior.
  </para>

146
  <para>
147
   When the operating system sends a write request to the storage hardware,
148 149
   there is little it can do to make sure the data has arrived at a truly
   non-volatile storage area. Rather, it is the
150
   administrator's responsibility to make certain that all storage components
151
   ensure integrity for both data and file-system metadata.
152 153
   Avoid disk controllers that have non-battery-backed write caches.
   At the drive level, disable write-back caching if the
154
   drive cannot guarantee the data will be written before shutdown.
155 156
   If you use SSDs, be aware that many of these do not honor cache flush
   commands by default.
157
   You can test for reliable I/O subsystem behavior using <ulink
158
   url="https://brad.livejournal.com/2116715.html"><filename>diskchecker.pl</filename></ulink>.
159
  </para>
160

161
  <para>
162 163 164 165
   Another risk of data loss is posed by the disk platter write
   operations themselves. Disk platters are divided into sectors,
   commonly 512 bytes each.  Every physical read or write operation
   processes a whole sector.
166
   When a write request arrives at the drive, it might be for some multiple
167
   of 512 bytes (<productname>PostgreSQL</productname> typically writes 8192 bytes, or
168
   16 sectors, at a time), and the process of writing could fail due
169
   to power loss at any time, meaning some of the 512-byte sectors were
170
   written while others were not.  To guard against such failures,
171 172 173
   <productname>PostgreSQL</productname> periodically writes full page images to
   permanent WAL storage <emphasis>before</emphasis> modifying the actual page on
   disk. By doing this, during crash recovery <productname>PostgreSQL</productname> can
174 175 176
   restore partially-written pages from WAL.  If you have file-system software
   that prevents partial page writes (e.g., ZFS),  you can turn off
   this page imaging by turning off the <xref
177
   linkend="guc-full-page-writes"/> parameter. Battery-Backed Unit
178 179
   (BBU) disk controllers do not prevent partial page writes unless
   they guarantee that data is written to the BBU as full (8kB) pages.
180
  </para>
181
  <para>
182
   <productname>PostgreSQL</productname> also protects against some kinds of data corruption
183 184 185 186 187 188 189 190 191 192 193 194 195
   on storage devices that may occur because of hardware errors or media failure over time,
   such as reading/writing garbage data.
   <itemizedlist>
    <listitem>
     <para>
      Each individual record in a WAL file is protected by a CRC-32 (32-bit) check
      that allows us to tell if record contents are correct. The CRC value
      is set when we write each WAL record and checked during crash recovery,
      archive recovery and replication.
     </para>
    </listitem>
    <listitem>
     <para>
196
      Data pages are not currently checksummed by default, though full page images
197
      recorded in WAL records will be protected; see <link
198
      linkend="app-initdb-data-checksums"><application>initdb</application></link>
199
      for details about enabling data checksums.
200 201 202 203
     </para>
    </listitem>
    <listitem>
     <para>
204
      Internal data structures such as <filename>pg_xact</filename>, <filename>pg_subtrans</filename>, <filename>pg_multixact</filename>,
205
      <filename>pg_serial</filename>, <filename>pg_notify</filename>, <filename>pg_stat</filename>, <filename>pg_snapshots</filename> are not directly
206 207 208 209 210 211
      checksummed, nor are pages protected by full page writes. However, where
      such data structures are persistent, WAL records are written that allow
      recent changes to be accurately rebuilt at crash recovery and those
      WAL records are protected as discussed above.
     </para>
    </listitem>
212 213
    <listitem>
     <para>
214
      Individual state files in <filename>pg_twophase</filename> are protected by CRC-32.
215 216
     </para>
    </listitem>
217 218 219 220 221 222 223 224 225 226
    <listitem>
     <para>
      Temporary data files used in larger SQL queries for sorts,
      materializations and intermediate results are not currently checksummed,
      nor will WAL records be written for changes to those files.
     </para>
    </listitem>
   </itemizedlist>
  </para>
  <para>
227
   <productname>PostgreSQL</productname> does not protect against correctable memory errors
228 229 230
   and it is assumed you will operate using RAM that uses industry standard
   Error Correcting Codes (ECC) or better protection.
  </para>
231
 </sect1>
232

233 234 235 236 237 238 239
 <sect1 id="checksums">
  <title>Data Checksums</title>
  <indexterm>
   <primary>checksums</primary>
  </indexterm>

  <para>
240 241 242
   By default, data pages are not protected by checksums, but this can
   optionally be enabled for a cluster. When enabled, each data page includes
   a checksum that is updated when the page is written and verified each time
243
   the page is read. Only data pages are protected by checksums; internal data
244 245 246 247
   structures and temporary files are not.
  </para>

  <para>
248
   Checksums are normally enabled when the cluster is initialized using <link
249 250 251
   linkend="app-initdb-data-checksums"><application>initdb</application></link>.
   They can also be enabled or disabled at a later time as an offline
   operation. Data checksums are enabled or disabled at the full cluster
252
   level, and cannot be specified individually for databases or tables.
253 254 255 256 257 258 259 260 261 262
  </para>

  <para>
   The current state of checksums in the cluster can be verified by viewing the
   value of the read-only configuration variable <xref
   linkend="guc-data-checksums" /> by issuing the command <command>SHOW
   data_checksums</command>.
  </para>

  <para>
263 264 265
   When attempting to recover from page corruptions, it may be necessary to
   bypass the checksum protection. To do this, temporarily set the
   configuration parameter <xref linkend="guc-ignore-checksum-failure" />.
266 267 268 269 270 271 272
  </para>

  <sect2 id="checksums-offline-enable-disable">
   <title>Off-line Enabling of Checksums</title>

   <para>
    The <link linkend="app-pgchecksums"><application>pg_checksums</application></link>
273
    application can be used to enable or disable data checksums, as well as
274 275 276 277 278 279
    verify checksums, on an offline cluster.
   </para>

  </sect2>
 </sect1>

280
  <sect1 id="wal-intro">
281 282 283 284 285 286 287 288 289 290 291 292 293
   <title>Write-Ahead Logging (<acronym>WAL</acronym>)</title>

   <indexterm zone="wal">
    <primary>WAL</primary>
   </indexterm>

   <indexterm>
    <primary>transaction log</primary>
    <see>WAL</see>
   </indexterm>

   <para>
    <firstterm>Write-Ahead Logging</firstterm> (<acronym>WAL</acronym>)
294
    is a standard method for ensuring data integrity.  A detailed
295
    description can be found in most (if not all) books about
296 297 298
    transaction processing. Briefly, <acronym>WAL</acronym>'s central
    concept is that changes to data files (where tables and indexes
    reside) must be written only after those changes have been logged,
299
    that is, after log records describing the changes have been flushed
300 301 302 303 304 305 306
    to permanent storage. If we follow this procedure, we do not need
    to flush data pages to disk on every transaction commit, because we
    know that in the event of a crash we will be able to recover the
    database using the log: any changes that have not been applied to
    the data pages can be redone from the log records.  (This is
    roll-forward recovery, also known as REDO.)
   </para>
307

308 309 310
   <tip>
    <para>
     Because <acronym>WAL</acronym> restores database file
311
     contents after a crash, journaled file systems are not necessary for
312
     reliable storage of the data files or WAL files.  In fact, journaling
313 314 315
     overhead can reduce performance, especially if journaling
     causes file system <emphasis>data</emphasis> to be flushed
     to disk.  Fortunately, data flushing during journaling can
316
     often be disabled with a file system mount option, e.g.,
317
     <literal>data=writeback</literal> on a Linux ext3 file system.
318
     Journaled file systems do improve boot speed after a crash.
319 320 321 322
    </para>
   </tip>


323
   <para>
324
    Using <acronym>WAL</acronym> results in a
325
    significantly reduced number of disk writes, because only the log
326 327 328
    file needs to be flushed to disk to guarantee that a transaction is
    committed, rather than every data file changed by the transaction.
    The log file is written sequentially,
329
    and so the cost of syncing the log is much less than the cost of
330
    flushing the data pages.  This is especially true for servers
331
    handling many small transactions touching different parts of the data
332 333 334
    store.  Furthermore, when the server is processing many small concurrent
    transactions, one <function>fsync</function> of the log file may
    suffice to commit many transactions.
335 336 337
   </para>

   <para>
338
    <acronym>WAL</acronym> also makes it possible to support on-line
339
    backup and point-in-time recovery, as described in <xref
340
    linkend="continuous-archiving"/>.  By archiving the WAL data we can support
341 342 343 344
    reverting to any time instant covered by the available WAL data:
    we simply install a prior physical backup of the database, and
    replay the WAL log just as far as the desired time.  What's more,
    the physical backup doesn't have to be an instantaneous snapshot
345
    of the database state &mdash; if it is made over some period of time,
346 347
    then replaying the WAL log for that period will fix any internal
    inconsistencies.
348
   </para>
349
  </sect1>
350

351 352 353 354 355 356 357 358 359 360 361 362
 <sect1 id="wal-async-commit">
  <title>Asynchronous Commit</title>

   <indexterm>
    <primary>synchronous commit</primary>
   </indexterm>

   <indexterm>
    <primary>asynchronous commit</primary>
   </indexterm>

  <para>
363
   <firstterm>Asynchronous commit</firstterm> is an option that allows transactions
364 365
   to complete more quickly, at the cost that the most recent transactions may
   be lost if the database should crash.  In many applications this is an
Peter Eisentraut's avatar
Peter Eisentraut committed
366
   acceptable trade-off.
367 368 369 370
  </para>

  <para>
   As described in the previous section, transaction commit is normally
371
   <firstterm>synchronous</firstterm>: the server waits for the transaction's
372 373 374 375 376 377 378 379 380 381 382 383 384 385 386 387 388 389 390 391 392 393 394 395 396 397 398 399 400 401 402 403 404 405 406 407 408 409 410 411 412 413
   <acronym>WAL</acronym> records to be flushed to permanent storage
   before returning a success indication to the client.  The client is
   therefore guaranteed that a transaction reported to be committed will
   be preserved, even in the event of a server crash immediately after.
   However, for short transactions this delay is a major component of the
   total transaction time.  Selecting asynchronous commit mode means that
   the server returns success as soon as the transaction is logically
   completed, before the <acronym>WAL</acronym> records it generated have
   actually made their way to disk.  This can provide a significant boost
   in throughput for small transactions.
  </para>

  <para>
   Asynchronous commit introduces the risk of data loss. There is a short
   time window between the report of transaction completion to the client
   and the time that the transaction is truly committed (that is, it is
   guaranteed not to be lost if the server crashes).  Thus asynchronous
   commit should not be used if the client will take external actions
   relying on the assumption that the transaction will be remembered.
   As an example, a bank would certainly not use asynchronous commit for
   a transaction recording an ATM's dispensing of cash.  But in many
   scenarios, such as event logging, there is no need for a strong
   guarantee of this kind.
  </para>

  <para>
   The risk that is taken by using asynchronous commit is of data loss,
   not data corruption.  If the database should crash, it will recover
   by replaying <acronym>WAL</acronym> up to the last record that was
   flushed.  The database will therefore be restored to a self-consistent
   state, but any transactions that were not yet flushed to disk will
   not be reflected in that state.  The net effect is therefore loss of
   the last few transactions.  Because the transactions are replayed in
   commit order, no inconsistency can be introduced &mdash; for example,
   if transaction B made changes relying on the effects of a previous
   transaction A, it is not possible for A's effects to be lost while B's
   effects are preserved.
  </para>

  <para>
   The user can select the commit mode of each transaction, so that
   it is possible to have both synchronous and asynchronous commit
Peter Eisentraut's avatar
Peter Eisentraut committed
414
   transactions running concurrently.  This allows flexible trade-offs
415 416
   between performance and certainty of transaction durability.
   The commit mode is controlled by the user-settable parameter
417
   <xref linkend="guc-synchronous-commit"/>, which can be changed in any of
418 419 420 421 422 423
   the ways that a configuration parameter can be set.  The mode used for
   any one transaction depends on the value of
   <varname>synchronous_commit</varname> when transaction commit begins.
  </para>

  <para>
424
   Certain utility commands, for instance <command>DROP TABLE</command>, are
425 426
   forced to commit synchronously regardless of the setting of
   <varname>synchronous_commit</varname>.  This is to ensure consistency
Peter Eisentraut's avatar
Peter Eisentraut committed
427
   between the server's file system and the logical state of the database.
428
   The commands supporting two-phase commit, such as <command>PREPARE
429
   TRANSACTION</command>, are also always synchronous.
430 431 432 433 434 435
  </para>

  <para>
   If the database crashes during the risk window between an
   asynchronous commit and the writing of the transaction's
   <acronym>WAL</acronym> records,
436
   then changes made during that transaction <emphasis>will</emphasis> be lost.
437
   The duration of the
Peter Eisentraut's avatar
Peter Eisentraut committed
438
   risk window is limited because a background process (the <quote>WAL
439
   writer</quote>) flushes unwritten <acronym>WAL</acronym> records to disk
440
   every <xref linkend="guc-wal-writer-delay"/> milliseconds.
441
   The actual maximum duration of the risk window is three times
Peter Eisentraut's avatar
Peter Eisentraut committed
442
   <varname>wal_writer_delay</varname> because the WAL writer is
443 444 445 446 447 448 449 450 451 452 453 454
   designed to favor writing whole pages at a time during busy periods.
  </para>

  <caution>
   <para>
    An immediate-mode shutdown is equivalent to a server crash, and will
    therefore cause loss of any unflushed asynchronous commits.
   </para>
  </caution>

  <para>
   Asynchronous commit provides behavior different from setting
455
   <xref linkend="guc-fsync"/> = off.
456 457
   <varname>fsync</varname> is a server-wide
   setting that will alter the behavior of all transactions.  It disables
458
   all logic within <productname>PostgreSQL</productname> that attempts to synchronize
459 460
   writes to different portions of the database, and therefore a system
   crash (that is, a hardware or operating system crash, not a failure of
461
   <productname>PostgreSQL</productname> itself) could result in arbitrarily bad
462 463 464 465 466 467 468
   corruption of the database state.  In many scenarios, asynchronous
   commit provides most of the performance improvement that could be
   obtained by turning off <varname>fsync</varname>, but without the risk
   of data corruption.
  </para>

  <para>
469
   <xref linkend="guc-commit-delay"/> also sounds very similar to
470 471 472
   asynchronous commit, but it is actually a synchronous commit method
   (in fact, <varname>commit_delay</varname> is ignored during an
   asynchronous commit).  <varname>commit_delay</varname> causes a delay
473 474 475 476 477 478
   just before a transaction flushes <acronym>WAL</acronym> to disk, in
   the hope that a single flush executed by one such transaction can also
   serve other transactions committing at about the same time.  The
   setting can be thought of as a way of increasing the time window in
   which transactions can join a group about to participate in a single
   flush, to amortize the cost of the flush among multiple transactions.
479 480 481 482
  </para>

 </sect1>

483 484 485 486
 <sect1 id="wal-configuration">
  <title><acronym>WAL</acronym> Configuration</title>

  <para>
487
   There are several <acronym>WAL</acronym>-related configuration parameters that
488
   affect database performance. This section explains their use.
489
   Consult <xref linkend="runtime-config"/> for general information about
490
   setting server configuration parameters.
491 492 493
  </para>

  <para>
494
   <firstterm>Checkpoints</firstterm><indexterm><primary>checkpoint</primary></indexterm>
495
   are points in the sequence of transactions at which it is guaranteed
496 497 498 499 500
   that the heap and index data files have been updated with all
   information written before that checkpoint.  At checkpoint time, all
   dirty data pages are flushed to disk and a special checkpoint record is
   written to the log file.  (The change records were previously flushed
   to the <acronym>WAL</acronym> files.)
501 502 503
   In the event of a crash, the crash recovery procedure looks at the latest
   checkpoint record to determine the point in the log (known as the redo
   record) from which it should start the REDO operation.  Any changes made to
504 505
   data files before that point are guaranteed to be already on disk.
   Hence, after a checkpoint, log segments preceding the one containing
506 507 508
   the redo record are no longer needed and can be recycled or removed. (When
   <acronym>WAL</acronym> archiving is being done, the log segments must be
   archived before being recycled or removed.)
509 510
  </para>

511 512 513
  <para>
   The checkpoint requirement of flushing all dirty data pages to disk
   can cause a significant I/O load.  For this reason, checkpoint
514 515
   activity is throttled so that I/O begins at checkpoint start and completes
   before the next checkpoint is due to start; this minimizes performance
516 517 518
   degradation during checkpoints.
  </para>

519
  <para>
520
   The server's checkpointer process automatically performs
521
   a checkpoint every so often.  A checkpoint is begun every <xref
522 523
   linkend="guc-checkpoint-timeout"/> seconds, or if
   <xref linkend="guc-max-wal-size"/> is about to be exceeded,
524
   whichever comes first.
525
   The default settings are 5 minutes and 1 GB, respectively.
526
   If no WAL has been written since the previous checkpoint, new checkpoints
527
   will be skipped even if <varname>checkpoint_timeout</varname> has passed.
528 529
   (If WAL archiving is being used and you want to put a lower limit on how
   often files are archived in order to bound potential data loss, you should
530
   adjust the <xref linkend="guc-archive-timeout"/> parameter rather than the
531 532
   checkpoint parameters.)
   It is also possible to force a checkpoint by using the SQL
533
   command <command>CHECKPOINT</command>.
534 535
  </para>

536
  <para>
537 538
   Reducing <varname>checkpoint_timeout</varname> and/or
   <varname>max_wal_size</varname> causes checkpoints to occur
539 540
   more often. This allows faster after-crash recovery, since less work
   will need to be redone. However, one must balance this against the
541
   increased cost of flushing dirty data pages more often. If
542
   <xref linkend="guc-full-page-writes"/> is set (as is the default), there is
543 544
   another factor to consider. To ensure data page consistency,
   the first modification of a data page after each checkpoint results in
545 546
   logging the entire page content. In that case,
   a smaller checkpoint interval increases the volume of output to the WAL log,
547
   partially negating the goal of using a smaller interval,
548
   and in any case causing more disk I/O.
549 550 551
  </para>

  <para>
552 553 554
   Checkpoints are fairly expensive, first because they require writing
   out all currently dirty buffers, and second because they result in
   extra subsequent WAL traffic as discussed above.  It is therefore
555
   wise to set the checkpointing parameters high enough so that checkpoints
556
   don't happen too often.  As a simple sanity check on your checkpointing
557
   parameters, you can set the <xref linkend="guc-checkpoint-warning"/>
558
   parameter.  If checkpoints happen closer together than
559
   <varname>checkpoint_warning</varname> seconds,
560
   a message will be output to the server log recommending increasing
561
   <varname>max_wal_size</varname>.  Occasional appearance of such
562
   a message is not cause for alarm, but if it appears often then the
563
   checkpoint control parameters should be increased. Bulk operations such
564 565
   as large <command>COPY</command> transfers might cause a number of such warnings
   to appear if you have not set <varname>max_wal_size</varname> high
566
   enough.
567 568 569
  </para>

  <para>
570 571 572
   To avoid flooding the I/O system with a burst of page writes,
   writing dirty buffers during a checkpoint is spread over a period of time.
   That period is controlled by
573
   <xref linkend="guc-checkpoint-completion-target"/>, which is
574 575
   given as a fraction of the checkpoint interval (configured by using
   <varname>checkpoint_timeout</varname>).
576
   The I/O rate is adjusted so that the checkpoint finishes when the
577 578 579
   given fraction of
   <varname>checkpoint_timeout</varname> seconds have elapsed, or before
   <varname>max_wal_size</varname> is exceeded, whichever is sooner.
580
   With the default value of 0.9,
581
   <productname>PostgreSQL</productname> can be expected to complete each checkpoint
582 583 584 585 586 587 588 589 590 591 592 593 594 595 596
   a bit before the next scheduled checkpoint (at around 90% of the last checkpoint's
   duration).  This spreads out the I/O as much as possible so that the checkpoint
   I/O load is consistent throughout the checkpoint interval.  The disadvantage of
   this is that prolonging checkpoints affects recovery time, because more WAL
   segments will need to be kept around for possible use in recovery.  A user
   concerned about the amount of time required to recover might wish to reduce
   <varname>checkpoint_timeout</varname> so that checkpoints occur more frequently
   but still spread the I/O across the checkpoint interval.  Alternatively,
   <varname>checkpoint_completion_target</varname> could be reduced, but this would
   result in times of more intense I/O (during the checkpoint) and times of less I/O
   (after the checkpoint completed but before the next scheduled checkpoint) and
   therefore is not recommended.
   Although <varname>checkpoint_completion_target</varname> could be set as high as
   1.0, it is typically recommended to set it to no higher than 0.9 (the default)
   since checkpoints include some other activities besides writing dirty buffers.
597 598 599 600 601
   A setting of 1.0 is quite likely to result in checkpoints not being
   completed on time, which would result in performance loss due to
   unexpected variation in the number of WAL segments needed.
  </para>

602
  <para>
603
   On Linux and POSIX platforms <xref linkend="guc-checkpoint-flush-after"/>
604 605 606
   allows to force the OS that pages written by the checkpoint should be
   flushed to disk after a configurable number of bytes.  Otherwise, these
   pages may be kept in the OS's page cache, inducing a stall when
607
   <literal>fsync</literal> is issued at the end of a checkpoint.  This setting will
608 609
   often help to reduce transaction latency, but it also can have an adverse
   effect on performance; particularly for workloads that are bigger than
610
   <xref linkend="guc-shared-buffers"/>, but smaller than the OS's page cache.
611 612
  </para>

613
  <para>
614 615
   The number of WAL segment files in <filename>pg_wal</filename> directory depends on
   <varname>min_wal_size</varname>, <varname>max_wal_size</varname> and
616 617 618
   the amount of WAL generated in previous checkpoint cycles. When old log
   segment files are no longer needed, they are removed or recycled (that is,
   renamed to become future segments in the numbered sequence). If, due to a
619
   short-term peak of log output rate, <varname>max_wal_size</varname> is
620 621 622 623 624 625
   exceeded, the unneeded segment files will be removed until the system
   gets back under this limit. Below that limit, the system recycles enough
   WAL files to cover the estimated need until the next checkpoint, and
   removes the rest. The estimate is based on a moving average of the number
   of WAL files used in previous checkpoint cycles. The moving average
   is increased immediately if the actual usage exceeds the estimate, so it
Magnus Hagander's avatar
Magnus Hagander committed
626
   accommodates peak usage rather than average usage to some extent.
627
   <varname>min_wal_size</varname> puts a minimum on the amount of WAL files
628 629 630 631 632 633 634
   recycled for future usage; that much WAL is always recycled for future use,
   even if the system is idle and the WAL usage estimate suggests that little
   WAL is needed.
  </para>

  <para>
   Independently of <varname>max_wal_size</varname>,
635 636
   the most recent <xref linkend="guc-wal-keep-size"/> megabytes of
   WAL files plus one additional WAL file are
637
   kept at all times. Also, if WAL archiving is used, old segments cannot be
638 639
   removed or recycled until they are archived. If WAL archiving cannot keep up
   with the pace that WAL is generated, or if <varname>archive_command</varname>
640
   fails repeatedly, old WAL files will accumulate in <filename>pg_wal</filename>
641 642
   until the situation is resolved. A slow or failed standby server that
   uses a replication slot will have the same effect (see
643
   <xref linkend="streaming-replication-slots"/>).
644 645
  </para>

646 647
  <para>
   In archive recovery or standby mode, the server periodically performs
648
   <firstterm>restartpoints</firstterm>,<indexterm><primary>restartpoint</primary></indexterm>
649
   which are similar to checkpoints in normal operation: the server forces
650
   all its state to disk, updates the <filename>pg_control</filename> file to
651
   indicate that the already-processed WAL data need not be scanned again,
652
   and then recycles any old log segment files in the <filename>pg_wal</filename>
653
   directory.
654 655
   Restartpoints can't be performed more frequently than checkpoints on the
   primary because restartpoints can only be performed at checkpoint records.
656
   A restartpoint is triggered when a checkpoint record is reached if at
657
   least <varname>checkpoint_timeout</varname> seconds have passed since the last
658
   restartpoint, or if WAL size is about to exceed
659 660
   <varname>max_wal_size</varname>. However, because of limitations on when a
   restartpoint can be performed, <varname>max_wal_size</varname> is often exceeded
661
   during recovery, by up to one checkpoint cycle's worth of WAL.
662
   (<varname>max_wal_size</varname> is never a hard limit anyway, so you should
663
   always leave plenty of headroom to avoid running out of disk space.)
664 665
  </para>

666
  <para>
667
   There are two commonly used internal <acronym>WAL</acronym> functions:
668 669
   <function>XLogInsertRecord</function> and <function>XLogFlush</function>.
   <function>XLogInsertRecord</function> is used to place a new record into
670
   the <acronym>WAL</acronym> buffers in shared memory. If there is no
671
   space for the new record, <function>XLogInsertRecord</function> will have
672
   to write (move to kernel cache) a few filled <acronym>WAL</acronym>
673
   buffers. This is undesirable because <function>XLogInsertRecord</function>
674 675 676
   is used on every database low level modification (for example, row
   insertion) at a time when an exclusive lock is held on affected
   data pages, so the operation needs to be as fast as possible.  What
677
   is worse, writing <acronym>WAL</acronym> buffers might also force the
678
   creation of a new log segment, which takes even more
679
   time. Normally, <acronym>WAL</acronym> buffers should be written
680
   and flushed by an <function>XLogFlush</function> request, which is
681 682
   made, for the most part, at transaction commit time to ensure that
   transaction records are flushed to permanent storage. On systems
683
   with high log output, <function>XLogFlush</function> requests might
684
   not occur often enough to prevent <function>XLogInsertRecord</function>
685
   from having to do writes.  On such systems
686
   one should increase the number of <acronym>WAL</acronym> buffers by
687 688
   modifying the <xref linkend="guc-wal-buffers"/> parameter.  When
   <xref linkend="guc-full-page-writes"/> is set and the system is very busy,
689
   setting <varname>wal_buffers</varname> higher will help smooth response times
690
   during the period immediately following each checkpoint.
691 692
  </para>

693
  <para>
694
   The <xref linkend="guc-commit-delay"/> parameter defines for how many
695 696 697 698 699
   microseconds a group commit leader process will sleep after acquiring a
   lock within <function>XLogFlush</function>, while group commit
   followers queue up behind the leader.  This delay allows other server
   processes to add their commit records to the WAL buffers so that all of
   them will be flushed by the leader's eventual sync operation.  No sleep
700 701
   will occur if <xref linkend="guc-fsync"/> is not enabled, or if fewer
   than <xref linkend="guc-commit-siblings"/> other sessions are currently
702 703 704 705 706 707 708 709 710 711 712 713 714 715 716 717 718
   in active transactions; this avoids sleeping when it's unlikely that
   any other session will commit soon.  Note that on some platforms, the
   resolution of a sleep request is ten milliseconds, so that any nonzero
   <varname>commit_delay</varname> setting between 1 and 10000
   microseconds would have the same effect.  Note also that on some
   platforms, sleep operations may take slightly longer than requested by
   the parameter.
  </para>

  <para>
   Since the purpose of <varname>commit_delay</varname> is to allow the
   cost of each flush operation to be amortized across concurrently
   committing transactions (potentially at the expense of transaction
   latency), it is necessary to quantify that cost before the setting can
   be chosen intelligently.  The higher that cost is, the more effective
   <varname>commit_delay</varname> is expected to be in increasing
   transaction throughput, up to a point.  The <xref
719
   linkend="pgtestfsync"/> program can be used to measure the average time
720 721 722 723 724 725 726 727 728 729 730 731 732 733 734 735 736 737 738 739 740 741 742 743
   in microseconds that a single WAL flush operation takes.  A value of
   half of the average time the program reports it takes to flush after a
   single 8kB write operation is often the most effective setting for
   <varname>commit_delay</varname>, so this value is recommended as the
   starting point to use when optimizing for a particular workload.  While
   tuning <varname>commit_delay</varname> is particularly useful when the
   WAL log is stored on high-latency rotating disks, benefits can be
   significant even on storage media with very fast sync times, such as
   solid-state drives or RAID arrays with a battery-backed write cache;
   but this should definitely be tested against a representative workload.
   Higher values of <varname>commit_siblings</varname> should be used in
   such cases, whereas smaller <varname>commit_siblings</varname> values
   are often helpful on higher latency media.  Note that it is quite
   possible that a setting of <varname>commit_delay</varname> that is too
   high can increase transaction latency by so much that total transaction
   throughput suffers.
  </para>

  <para>
   When <varname>commit_delay</varname> is set to zero (the default), it
   is still possible for a form of group commit to occur, but each group
   will consist only of sessions that reach the point where they need to
   flush their commit records during the window in which the previous
   flush operation (if any) is occurring.  At higher client counts a
744
   <quote>gangway effect</quote> tends to occur, so that the effects of group
745 746 747 748 749 750 751 752
   commit become significant even when <varname>commit_delay</varname> is
   zero, and thus explicitly setting <varname>commit_delay</varname> tends
   to help less.  Setting <varname>commit_delay</varname> can only help
   when (1) there are some concurrently committing transactions, and (2)
   throughput is limited to some degree by commit rate; but with high
   rotational latency this setting can be effective in increasing
   transaction throughput with as few as two clients (that is, a single
   committing client with one sibling transaction).
753
  </para>
754 755

  <para>
756
   The <xref linkend="guc-wal-sync-method"/> parameter determines how
757
   <productname>PostgreSQL</productname> will ask the kernel to force
758 759
   <acronym>WAL</acronym> updates out to disk.
   All the options should be the same in terms of reliability, with
760
   the exception of <literal>fsync_writethrough</literal>, which can sometimes
761
   force a flush of the disk cache even when other options do not do so.
762 763
   However, it's quite platform-specific which one will be the fastest.
   You can test the speeds of different options using the <xref
764
   linkend="pgtestfsync"/> program.
765
   Note that this parameter is irrelevant if <varname>fsync</varname>
766 767 768 769
   has been turned off.
  </para>

  <para>
770
   Enabling the <xref linkend="guc-wal-debug"/> configuration parameter
771 772
   (provided that <productname>PostgreSQL</productname> has been
   compiled with support for it) will result in each
773
   <function>XLogInsertRecord</function> and <function>XLogFlush</function>
774
   <acronym>WAL</acronym> call being logged to the server log. This
775
   option might be replaced by a more general mechanism in the future.
776
  </para>
777 778 779 780 781 782 783 784 785

  <para>
   There are two internal functions to write WAL data to disk:
   <function>XLogWrite</function> and <function>issue_xlog_fsync</function>.
   When <xref linkend="guc-track-wal-io-timing"/> is enabled, the total
   amounts of time <function>XLogWrite</function> writes and
   <function>issue_xlog_fsync</function> syncs WAL data to disk are counted as
   <literal>wal_write_time</literal> and <literal>wal_sync_time</literal> in
   <xref linkend="pg-stat-wal-view"/>, respectively.
786
   <function>XLogWrite</function> is normally called by
787 788 789 790 791 792 793 794 795 796 797 798 799
   <function>XLogInsertRecord</function> (when there is no space for the new
   record in WAL buffers), <function>XLogFlush</function> and the WAL writer,
   to write WAL buffers to disk and call <function>issue_xlog_fsync</function>.
   <function>issue_xlog_fsync</function> is normally called by
   <function>XLogWrite</function> to sync WAL files to disk.
   If <varname>wal_sync_method</varname> is either
   <literal>open_datasync</literal> or <literal>open_sync</literal>,
   a write operation in <function>XLogWrite</function> guarantees to sync written
   WAL data to disk and <function>issue_xlog_fsync</function> does nothing.
   If <varname>wal_sync_method</varname> is either <literal>fdatasync</literal>,
   <literal>fsync</literal>, or <literal>fsync_writethrough</literal>,
   the write operation moves WAL buffers to kernel cache and
   <function>issue_xlog_fsync</function> syncs them to disk. Regardless
800
   of the setting of <varname>track_wal_io_timing</varname>, the number
801 802 803 804 805
   of times <function>XLogWrite</function> writes and
   <function>issue_xlog_fsync</function> syncs WAL data to disk are also
   counted as <literal>wal_write</literal> and <literal>wal_sync</literal>
   in <structname>pg_stat_wal</structname>, respectively.
  </para>
806
 </sect1>
807 808

 <sect1 id="wal-internals">
809
  <title>WAL Internals</title>
810

811 812 813 814
  <indexterm zone="wal-internals">
   <primary>LSN</primary>
  </indexterm>

815 816
  <para>
   <acronym>WAL</acronym> is automatically enabled; no action is
817 818
   required from the administrator except ensuring that the
   disk-space requirements for the <acronym>WAL</acronym> logs are met,
819
   and that any necessary tuning is done (see <xref
820
   linkend="wal-configuration"/>).
821 822
  </para>

823 824 825 826 827 828
  <para>
   <acronym>WAL</acronym> records are appended to the <acronym>WAL</acronym>
   logs as each new record is written. The insert position is described by
   a Log Sequence Number (<acronym>LSN</acronym>) that is a byte offset into
   the logs, increasing monotonically with each new record.
   <acronym>LSN</acronym> values are returned as the datatype
829
   <link linkend="datatype-pg-lsn"><type>pg_lsn</type></link>. Values can be
830 831 832 833 834
   compared to calculate the volume of <acronym>WAL</acronym> data that
   separates them, so they are used to measure the progress of replication
   and recovery.
  </para>

835 836
  <para>
   <acronym>WAL</acronym> logs are stored in the directory
837
   <filename>pg_wal</filename> under the data directory, as a set of
838
   segment files, normally each 16 MB in size (but the size can be changed
Peter Eisentraut's avatar
Peter Eisentraut committed
839
   by altering the <option>--wal-segsize</option> <application>initdb</application> option).  Each segment is
840
   divided into pages, normally 8 kB each (this size can be changed via the
841
   <option>--with-wal-blocksize</option> configure option).  The log record headers
842 843 844
   are described in <filename>access/xlogrecord.h</filename>; the record
   content is dependent on the type of event that is being logged.  Segment
   files are given ever-increasing numbers as names, starting at
845
   <filename>000000010000000000000001</filename>.  The numbers do not wrap,
846
   but it will take a very, very long time to exhaust the
847 848 849 850
   available stock of numbers.
  </para>

  <para>
851 852
   It is advantageous if the log is located on a different disk from the
   main database files.  This can be achieved by moving the
853
   <filename>pg_wal</filename> directory to another location (while the server
854 855 856 857 858
   is shut down, of course) and creating a symbolic link from the
   original location in the main data directory to the new location.
  </para>

  <para>
859 860
   The aim of <acronym>WAL</acronym> is to ensure that the log is
   written before database records are altered, but this can be subverted by
861
   disk drives<indexterm><primary>disk drive</primary></indexterm> that falsely report a
862
   successful write to the kernel,
863
   when in fact they have only cached the data and not yet stored it
864
   on the disk.  A power failure in such a situation might lead to
865 866 867
   irrecoverable data corruption.  Administrators should try to ensure
   that disks holding <productname>PostgreSQL</productname>'s
   <acronym>WAL</acronym> log files do not make such false reports.
868
   (See <xref linkend="wal-reliability"/>.)
869 870 871 872 873
  </para>

  <para>
   After a checkpoint has been made and the log flushed, the
   checkpoint's position is saved in the file
874 875
   <filename>pg_control</filename>. Therefore, at the start of recovery,
   the server first reads <filename>pg_control</filename> and
876
   then the checkpoint record; then it performs the REDO operation by
877
   scanning forward from the log location indicated in the checkpoint
878
   record.  Because the entire content of data pages is saved in the
879
   log on the first page modification after a checkpoint (assuming
880
   <xref linkend="guc-full-page-writes"/> is not disabled), all pages
881 882 883 884 885
   changed since the checkpoint will be restored to a consistent
   state.
  </para>

  <para>
886
   To deal with the case where <filename>pg_control</filename> is
887
   corrupt, we should support the possibility of scanning existing log
888
   segments in reverse order &mdash; newest to oldest &mdash; in order to find the
889 890 891
   latest checkpoint.  This has not been implemented yet.
   <filename>pg_control</filename> is small enough (less than one disk page)
   that it is not subject to partial-write problems, and as of this writing
892
   there have been no reports of database failures due solely to the inability
893 894 895
   to read <filename>pg_control</filename> itself.  So while it is
   theoretically a weak spot, <filename>pg_control</filename> does not
   seem to be a problem in practice.
896 897
  </para>
 </sect1>
898
</chapter>