1. 12 Apr, 2022 3 commits
  2. 06 Apr, 2022 2 commits
  3. 05 Apr, 2022 1 commit
  4. 02 Apr, 2022 2 commits
  5. 01 Apr, 2022 1 commit
  6. 31 Mar, 2022 3 commits
  7. 30 Mar, 2022 1 commit
  8. 29 Mar, 2022 1 commit
  9. 28 Mar, 2022 3 commits
  10. 27 Mar, 2022 2 commits
    • Tom Lane's avatar
      Fix breakage of get_ps_display() in the PS_USE_NONE case. · 3f7a59c5
      Tom Lane authored
      Commit 8c6d30f2 caused this function to fail to set *displen
      in the PS_USE_NONE code path.  If the variable's previous value
      had been negative, that'd lead to a memory clobber at some call
      sites.  We'd managed not to notice due to very thin test coverage
      of such configurations, but this appears to explain buildfarm member
      lorikeet's recent struggles.
      
      Credit to Andrew Dunstan for spotting the problem.  Back-patch
      to v13 where the bug was introduced.
      
      Discussion: https://postgr.es/m/136102.1648320427@sss.pgh.pa.us
      3f7a59c5
    • Michael Paquier's avatar
      pageinspect: Add more sanity checks to prevent out-of-bound reads · 27d38444
      Michael Paquier authored
      A couple of code paths use the special area on the page passed by the
      function caller, expecting to find some data in it.  However, feeding
      an incorrect page can lead to out-of-bound reads when trying to access
      the page special area (like a heap page that has no special area,
      leading PageGetSpecialPointer() to grab a pointer outside the allocated
      page).
      
      The functions used for hash and btree indexes have some protection
      already against that, while some other functions using a relation OID
      as argument would make sure that the access method involved is correct,
      but functions taking in input a raw page without knowing the relation
      the page is attached to would run into problems.
      
      This commit improves the set of checks used in the code paths of BRIN,
      btree (including one check if a leaf page is found with a non-zero
      level), GIN and GiST to verify that the page given in input has a
      special area size that fits with each access method, which is done
      though PageGetSpecialSize(), becore calling PageGetSpecialPointer().
      
      The scope of the checks done is limited to work with pages that one
      would pass after getting a block with get_raw_page(), as it is possible
      to craft byteas that could bypass existing code paths.  Having too many
      checks would also impact the usability of pageinspect, as the existing
      code is very useful to look at the content details in a corrupted page,
      so the focus is really to avoid out-of-bound reads as this is never a
      good thing even with functions whose execution is limited to
      superusers.
      
      The safest approach could be to rework the functions so as these fetch a
      block using a relation OID and a block number, but there are also cases
      where using a raw page is useful.
      
      Tests are added to cover all the code paths that needed such checks, and
      an error message for hash indexes is reworded to fit better with what
      this commit adds.
      
      Reported-By: Alexander Lakhin
      Author: Julien Rouhaud, Michael Paquier
      Discussion: https://postgr.es/m/16527-ef7606186f0610a1@postgresql.org
      Discussion: https://postgr.es/m/561e187b-3549-c8d5-03f5-525c14e65bd0@postgrespro.ru
      Backpatch-through: 10
      27d38444
  11. 26 Mar, 2022 1 commit
    • Tom Lane's avatar
      Suppress compiler warning in relptr_store(). · 0144c9c7
      Tom Lane authored
      clang 13 with -Wextra warns that "performing pointer subtraction with
      a null pointer has undefined behavior" in the places where freepage.c
      tries to set a relptr variable to constant NULL.  This appears to be
      a compiler bug, but it's unlikely to get fixed instantly.  Fortunately,
      we can work around it by introducing an inline support function, which
      seems like a good change anyway because it removes the macro's existing
      double-evaluation hazard.
      
      Backpatch to v10 where this code was introduced.
      
      Patch by me, based on an idea of Andres Freund's.
      
      Discussion: https://postgr.es/m/48826.1648310694@sss.pgh.pa.us
      0144c9c7
  12. 25 Mar, 2022 2 commits
    • Tom Lane's avatar
      Harden TAP tests that intentionally corrupt page checksums. · 579cef5f
      Tom Lane authored
      The previous method for doing that was to write zeroes into a
      predetermined set of page locations.  However, there's a roughly
      1-in-64K chance that the existing checksum will match by chance,
      and yesterday several buildfarm animals started to reproducibly
      see that, resulting in test failures because no checksum mismatch
      was reported.
      
      Since the checksum includes the page LSN, test success depends on
      the length of the installation's WAL history, which is affected by
      (at least) the initial catalog contents, the set of locales installed
      on the system, and the length of the pathname of the test directory.
      Sooner or later we were going to hit a chance match, and today is
      that day.
      
      Harden these tests by specifically inverting the checksum field and
      leaving all else alone, thereby guaranteeing that the checksum is
      incorrect.
      
      In passing, fix places that were using seek() to set up for syswrite(),
      a combination that the Perl docs very explicitly warn against.  We've
      probably escaped problems because no regular buffered I/O is done on
      these filehandles; but if it ever breaks, we wouldn't deserve or get
      much sympathy.
      
      Although we've only seen problems in HEAD, now that we recognize the
      environmental dependencies it seems like it might be just a matter
      of time until someone manages to hit this in back-branch testing.
      Hence, back-patch to v11 where we started doing this kind of test.
      
      Discussion: https://postgr.es/m/3192026.1648185780@sss.pgh.pa.us
      579cef5f
    • Alvaro Herrera's avatar
      Fix replay of create database records on standby · ffd28516
      Alvaro Herrera authored
      Crash recovery on standby may encounter missing directories when
      replaying create database WAL records.  Prior to this patch, the standby
      would fail to recover in such a case.  However, the directories could be
      legitimately missing.  Consider a sequence of WAL records as follows:
      
          CREATE DATABASE
          DROP DATABASE
          DROP TABLESPACE
      
      If, after replaying the last WAL record and removing the tablespace
      directory, the standby crashes and has to replay the create database
      record again, the crash recovery must be able to move on.
      
      This patch adds a mechanism similar to invalid-page tracking, to keep a
      tally of missing directories during crash recovery.  If all the missing
      directory references are matched with corresponding drop records at the
      end of crash recovery, the standby can safely continue following the
      primary.
      
      Backpatch to 13, at least for now.  The bug is older, but fixing it in
      older branches requires more careful study of the interactions with
      commit e6d80695, which appeared in 13.
      
      A new TAP test file is added to verify the condition.  However, because
      it depends on commit d6d317dbf615, it can only be added to branch
      master.  I (Álvaro) manually verified that the code behaves as expected
      in branch 14.  It's a bit nervous-making to leave the code uncovered by
      tests in older branches, but leaving the bug unfixed is even worse.
      Also, the main reason this fix took so long is precisely that we
      couldn't agree on a good strategy to approach testing for the bug, so
      perhaps this is the best we can do.
      Diagnosed-by: default avatarPaul Guo <paulguo@gmail.com>
      Author: Paul Guo <paulguo@gmail.com>
      Author: Kyotaro Horiguchi <horikyota.ntt@gmail.com>
      Author: Asim R Praveen <apraveen@pivotal.io>
      Discussion: https://postgr.es/m/CAEET0ZGx9AvioViLf7nbR_8tH9-=27DN5xWJ2P9-ROH16e4JUA@mail.gmail.com
      ffd28516
  13. 24 Mar, 2022 1 commit
  14. 23 Mar, 2022 6 commits
  15. 22 Mar, 2022 2 commits
    • Andres Freund's avatar
      Add missing dependency of pg_dumpall to WIN32RES. · 2d608c96
      Andres Freund authored
      When cross-building to windows, or building with mingw on windows, the build
      could fail with
        x86_64-w64-mingw32-gcc: error: win32ver.o: No such file or director
      because pg_dumpall didn't depend on WIN32RES, but it's recipe references
      it. The build nevertheless succeeded most of the time, due to
      pg_dump/pg_restore having the required dependency, causing win32ver.o to be
      built.
      Reported-By: default avatarThomas Munro <thomas.munro@gmail.com>
      Discussion: https://postgr.es/m/CA+hUKGJeekpUPWW6yCVdf9=oBAcCp86RrBivo4Y4cwazAzGPng@mail.gmail.com
      Backpatch: 10-, omission present on all live branches
      2d608c96
    • Michael Paquier's avatar
      Fix failures in SSL tests caused by out-of-tree keys and certificates · fdb1be49
      Michael Paquier authored
      This issue is environment-sensitive, where the SSL tests could fail in
      various way by feeding on defaults provided by sslcert, sslkey,
      sslrootkey, sslrootcert, sslcrl and sslcrldir coming from a local setup,
      as of ~/.postgresql/ by default.  Horiguchi-san has reported two
      failures, but more advanced testing from me (aka inclusion of garbage
      SSL configuration in ~/.postgresql/ for all the configuration
      parameters) has showed dozens of failures that can be triggered in the
      whole test suite.
      
      History has showed that we are not good when it comes to address such
      issues, fixing them locally like in dd877998, and such problems keep
      appearing.  This commit strengthens the entire test suite to put an end
      to this set of problems by embedding invalid default values in all the
      connection strings used in the tests.  The invalid values are prefixed
      in each connection string, relying on the follow-up values passed in the
      connection string to enforce any invalid value previously set.  Note
      that two tests related to CRLs are required to fail with certain pre-set
      configurations, but we can rely on enforcing an empty value instead
      after the invalid set of values.
      
      Reported-by: Kyotaro Horiguchi
      Reviewed-by: Andrew Dunstan, Daniel Gustafsson, Kyotaro Horiguchi
      Discussion: https://postgr.es/m/20220316.163658.1122740600489097632.horikyota.ntt@gmail.com
      backpatch-through: 10
      fdb1be49
  16. 21 Mar, 2022 2 commits
    • Tom Lane's avatar
      Fix assorted missing logic for GroupingFunc nodes. · 48b6035f
      Tom Lane authored
      The planner needs to treat GroupingFunc like Aggref for many purposes,
      in particular with respect to processing of the argument expressions,
      which are not to be evaluated at runtime.  A few places hadn't gotten
      that memo, notably including subselect.c's processing of outer-level
      aggregates.  This resulted in assertion failures or wrong plans for
      cases in which a GROUPING() construct references an outer aggregation
      level.
      
      Also fix missing special cases for GroupingFunc in cost_qual_eval
      (resulting in wrong cost estimates for GROUPING(), although it's
      not clear that that would affect plan shapes in practice) and in
      ruleutils.c (resulting in excess parentheses in pretty-print mode).
      
      Per bug #17088 from Yaoguang Chen.  Back-patch to all supported
      branches.
      
      Richard Guo, Tom Lane
      
      Discussion: https://postgr.es/m/17088-e33882b387de7f5c@postgresql.org
      48b6035f
    • Tom Lane's avatar
      Fix risk of deadlock failure while dropping a partitioned index. · 05ccf974
      Tom Lane authored
      DROP INDEX needs to lock the index's table before the index itself,
      else it will deadlock against ordinary queries that acquire the
      relation locks in that order.  This is correctly mechanized for
      plain indexes by RangeVarCallbackForDropRelation; but in the case of
      a partitioned index, we neglected to lock the child tables in advance
      of locking the child indexes.  We can fix that by traversing the
      inheritance tree and acquiring the needed locks in RemoveRelations,
      after we have acquired our locks on the parent partitioned table and
      index.
      
      While at it, do some refactoring to eliminate confusion between
      the actual and expected relkind in RangeVarCallbackForDropRelation.
      We can save a couple of syscache lookups too, by having that function
      pass back info that RemoveRelations will need.
      
      Back-patch to v11 where partitioned indexes were added.
      
      Jimmy Yih, Gaurab Dey, Tom Lane
      
      Discussion: https://postgr.es/m/BYAPR05MB645402330042E17D91A70C12BD5F9@BYAPR05MB6454.namprd05.prod.outlook.com
      05ccf974
  17. 20 Mar, 2022 1 commit
  18. 19 Mar, 2022 1 commit
  19. 18 Mar, 2022 1 commit
  20. 17 Mar, 2022 1 commit
    • Tom Lane's avatar
      Revert applying column aliases to the output of whole-row Vars. · 1d072bd2
      Tom Lane authored
      In commit bf7ca158, I had the bright idea that we could make the
      result of a whole-row Var (that is, foo.*) track any column aliases
      that had been applied to the FROM entry the Var refers to.  However,
      that's not terribly logically consistent, because now the output of
      the Var is no longer of the named composite type that the Var claims
      to emit.  bf7ca158 tried to handle that by changing the output
      tuple values to be labeled with a blessed RECORD type, but that's
      really pretty disastrous: we can wind up storing such tuples onto
      disk, whereupon they're not readable by other sessions.
      
      The only practical fix I can see is to give up on what bf7ca158
      tried to do, and say that the column names of tuples produced by
      a whole-row Var are always those of the underlying named composite
      type, query aliases or no.  While this introduces some inconsistencies,
      it removes others, so it's not that awful in the abstract.  What *is*
      kind of awful is to make such a behavioral change in a back-patched
      bug fix.  But corrupt data is worse, so back-patched it will be.
      
      (A workaround available to anyone who's unhappy about this is to
      introduce an extra level of sub-SELECT, so that the whole-row Var is
      referring to the sub-SELECT's output and not to a named table type.
      Then the Var is of type RECORD to begin with and there's no issue.)
      
      Per report from Miles Delahunty.  The faulty commit dates to 9.5,
      so back-patch to all supported branches.
      
      Discussion: https://postgr.es/m/2950001.1638729947@sss.pgh.pa.us
      1d072bd2
  21. 16 Mar, 2022 3 commits
    • Tomas Vondra's avatar
      Fix publish_as_relid with multiple publications · 677a1dc0
      Tomas Vondra authored
      Commit 83fd4532 allowed publishing of changes via ancestors, for
      publications defined with publish_via_partition_root. But the way
      the ancestor was determined in get_rel_sync_entry() was incorrect,
      simply updating the same variable. So with multiple publications,
      replicating different ancestors, the outcome depended on the order
      of publications in the list - the value from the last loop was used,
      even if it wasn't the top-most ancestor.
      
      This is a probably rare situation, as in most cases publications do
      not overlap, so each partition has exactly one candidate ancestor
      to replicate as and there's no ambiguity.
      
      Fixed by tracking the "ancestor level" for each publication, and
      picking the top-most ancestor. Adds a test case, verifying the
      correct ancestor is used for publishing the changes and that this
      does not depend on order of publications in the list.
      
      Older releases have another bug in this loop - once all actions are
      replicated, the loop is terminated, on the assumption that inspecting
      additional publications is unecessary. But that misses the fact that
      those additional applications may replicate different ancestors.
      
      Fixed by removal of this break condition. We might still terminate the
      loop in some cases (e.g. when replicating all actions and the ancestor
      is the partition root).
      
      Backpatch to 13, where publish_via_partition_root was introduced.
      
      Initial report and fix by me, test added by Hou zj. Reviews and
      improvements by Amit Kapila.
      
      Author: Tomas Vondra, Hou zj, Amit Kapila
      Reviewed-by: Amit Kapila, Hou zj
      Discussion: https://postgr.es/m/d26d24dd-2fab-3c48-0162-2b7f84a9c893%40enterprisedb.com
      677a1dc0
    • Alexander Korotkov's avatar
      Fix default signature length for gist_ltree_ops · 7d30f59d
      Alexander Korotkov authored
      911e7020 implemented operator class parameters including the signature length
      in ltree.  Previously, the signature length for gist_ltree_ops was 8.  Because
      of bug 911e7020 the default signature length for gist_ltree_ops became 28 for
      ltree 1.1 (where options method is NOT provided) and 8 for ltree 1.2 (where
      options method is provided).  This commit changes the default signature length
      for ltree 1.1 to 8.
      
      Existing gist_ltree_ops indexes might be corrupted in various scenarios.
      Thus, we have to recommend reindexing all the gist_ltree_ops indexes after
      the upgrade.
      
      Reported-by: Victor Yegorov
      Reviewed-by: Tomas Vondra, Tom Lane, Andres Freund, Nikita Glukhov
      Reviewed-by: Andrew Dunstan
      Author: Tomas Vondra, Alexander Korotkov
      Discussion: https://postgr.es/m/17406-71e02820ae79bb40%40postgresql.org
      Discussion: https://postgr.es/m/d80e0a55-6c3e-5b26-53e3-3c4f973f737c%40enterprisedb.com
      7d30f59d
    • Thomas Munro's avatar
      Fix race between DROP TABLESPACE and checkpointing. · 26e00793
      Thomas Munro authored
      Commands like ALTER TABLE SET TABLESPACE may leave files for the next
      checkpoint to clean up.  If such files are not removed by the time DROP
      TABLESPACE is called, we request a checkpoint so that they are deleted.
      However, there is presently a window before checkpoint start where new
      unlink requests won't be scheduled until the following checkpoint.  This
      means that the checkpoint forced by DROP TABLESPACE might not remove the
      files we expect it to remove, and the following ERROR will be emitted:
      
      	ERROR:  tablespace "mytblspc" is not empty
      
      To fix, add a call to AbsorbSyncRequests() just before advancing the
      unlink cycle counter.  This ensures that any unlink requests forwarded
      prior to checkpoint start (i.e., when ckpt_started is incremented) will
      be processed by the current checkpoint.  Since AbsorbSyncRequests()
      performs memory allocations, it cannot be called within a critical
      section, so we also need to move SyncPreCheckpoint() to before
      CreateCheckPoint()'s critical section.
      
      This is an old bug, so back-patch to all supported versions.
      
      Author: Nathan Bossart <nathandbossart@gmail.com>
      Reported-by: default avatarNathan Bossart <nathandbossart@gmail.com>
      Reviewed-by: default avatarThomas Munro <thomas.munro@gmail.com>
      Reviewed-by: default avatarAndres Freund <andres@anarazel.de>
      Discussion: https://postgr.es/m/20220215235845.GA2665318%40nathanxps13
      26e00793