From 719b793cf61b8bdacc1b2bf5d44dad4b24b0f36c Mon Sep 17 00:00:00 2001 From: David Christensen Date: Fri, 21 Oct 2022 18:36:12 -0400 Subject: [PATCH 6/6] Add 64-bit checksum page feature --- src/backend/access/transam/xlog.c | 3 +- src/backend/backup/basebackup.c | 9 +- src/backend/storage/page/bufpage.c | 21 +- src/backend/utils/misc/guc_tables.c | 11 + src/bin/initdb/initdb.c | 30 +- src/bin/pg_controldata/pg_controldata.c | 4 +- src/common/pagefeat.c | 3 + src/include/common/komihash.h | 565 ++++++++++++++++++++++++ src/include/common/pagefeat.h | 2 + src/include/storage/checksum.h | 1 + src/include/storage/checksum_impl.h | 62 +++ 11 files changed, 695 insertions(+), 16 deletions(-) create mode 100644 src/include/common/komihash.h diff --git a/src/backend/access/transam/xlog.c b/src/backend/access/transam/xlog.c index a1379efd49..f184f2a257 100644 --- a/src/backend/access/transam/xlog.c +++ b/src/backend/access/transam/xlog.c @@ -4234,7 +4234,8 @@ DataChecksumsEnabled(void) { Assert(ControlFile != NULL); return (ControlFile->data_checksum_version > 0) || \ - PageFeatureSetHasFeature(ControlFile->page_features, PF_PAGE_CHECKSUMS32); + PageFeatureSetHasFeature(ControlFile->page_features, PF_PAGE_CHECKSUMS32) || \ + PageFeatureSetHasFeature(ControlFile->page_features, PF_PAGE_CHECKSUMS64); } diff --git a/src/backend/backup/basebackup.c b/src/backend/backup/basebackup.c index 90cd97b938..f74350f9de 100644 --- a/src/backend/backup/basebackup.c +++ b/src/backend/backup/basebackup.c @@ -1493,7 +1493,7 @@ sendFile(bbsink *sink, const char *readfilename, const char *tarfilename, int fd; BlockNumber blkno = 0; bool block_retry = false; - uint32 checksum, page_checksum; + uint64 checksum, page_checksum; int checksum_failures = 0; off_t cnt; int i; @@ -1612,7 +1612,12 @@ sendFile(bbsink *sink, const char *readfilename, const char *tarfilename, char *extended_checksum_loc = NULL; /* are we using extended checksums? */ - if ((extended_checksum_loc = ClusterPageFeatureOffset(page, PF_PAGE_CHECKSUMS32))) + if ((extended_checksum_loc = ClusterPageFeatureOffset(page, PF_PAGE_CHECKSUMS64))) + { + page_checksum = *(uint64*)extended_checksum_loc; + checksum = pg_checksum64_page(page, blkno + segmentno * RELSEG_SIZE, (uint64*)extended_checksum_loc); + } + else if ((extended_checksum_loc = ClusterPageFeatureOffset(page, PF_PAGE_CHECKSUMS32))) { page_checksum = *(uint16*)extended_checksum_loc; page_checksum <<= 16; diff --git a/src/backend/storage/page/bufpage.c b/src/backend/storage/page/bufpage.c index b50b4e76a0..afe8707266 100644 --- a/src/backend/storage/page/bufpage.c +++ b/src/backend/storage/page/bufpage.c @@ -93,8 +93,8 @@ PageIsVerifiedExtended(Page page, BlockNumber blkno, int flags) bool checksum_failure = false; bool header_sane = false; bool all_zeroes = false; - uint32 checksum = 0; - uint32 page_checksum = 0; + uint64 checksum = 0; + uint64 page_checksum = 0; char *extended_checksum_loc = NULL; /* * Don't verify page data unless the page passes basic non-zero test @@ -104,7 +104,12 @@ PageIsVerifiedExtended(Page page, BlockNumber blkno, int flags) if (DataChecksumsEnabled()) { /* are we using extended checksums? */ - if ((extended_checksum_loc = ClusterPageFeatureOffset(page, PF_PAGE_CHECKSUMS32))) + if ((extended_checksum_loc = ClusterPageFeatureOffset(page, PF_PAGE_CHECKSUMS64))) + { + page_checksum = *((uint64*)extended_checksum_loc); + checksum = pg_checksum64_page(page, blkno, (uint64*)extended_checksum_loc); + } + else if ((extended_checksum_loc = ClusterPageFeatureOffset(page, PF_PAGE_CHECKSUMS32))) { /* high bits of the existing checksum are stored as a uint16 at extended_checksum_loc, low bits in pd_checksum */ page_checksum = *((uint16*)extended_checksum_loc); @@ -163,7 +168,7 @@ PageIsVerifiedExtended(Page page, BlockNumber blkno, int flags) if ((flags & PIV_LOG_WARNING) != 0) ereport(WARNING, (errcode(ERRCODE_DATA_CORRUPTED), - errmsg("page verification failed, calculated checksum %u but expected %u", + errmsg("page verification failed, calculated checksum %lu but expected %lu", checksum, page_checksum))); if ((flags & PIV_REPORT_STAT) != 0) @@ -1541,7 +1546,9 @@ PageSetChecksumCopy(Page page, BlockNumber blkno) memcpy(pageCopy, (char *) page, BLCKSZ); - if ((extended_checksum_loc = ClusterPageFeatureOffset(pageCopy, PF_PAGE_CHECKSUMS32))) + if ((extended_checksum_loc = ClusterPageFeatureOffset(pageCopy, PF_PAGE_CHECKSUMS64))) + *(uint64*)extended_checksum_loc = pg_checksum64_page(pageCopy, blkno, (uint64*)extended_checksum_loc); + else if ((extended_checksum_loc = ClusterPageFeatureOffset(pageCopy, PF_PAGE_CHECKSUMS32))) { /* 32-bit checksums split storage between pd_checksum and page feature offset */ uint32 checksum = pg_checksum32_page((char*)pageCopy, blkno, extended_checksum_loc); @@ -1571,7 +1578,9 @@ PageSetChecksumInplace(Page page, BlockNumber blkno) return; /* are we using extended checksums? */ - if ((extended_checksum_loc = ClusterPageFeatureOffset(page, PF_PAGE_CHECKSUMS32))) + if ((extended_checksum_loc = ClusterPageFeatureOffset(page, PF_PAGE_CHECKSUMS64))) + *(uint64*)extended_checksum_loc = pg_checksum64_page(page, blkno, (uint64*)extended_checksum_loc); + else if ((extended_checksum_loc = ClusterPageFeatureOffset(page, PF_PAGE_CHECKSUMS32))) { /* 32-bit checksums split storage between pd_checksum and page feature offset */ uint32 checksum = pg_checksum32_page((char*)page, blkno, extended_checksum_loc); diff --git a/src/backend/utils/misc/guc_tables.c b/src/backend/utils/misc/guc_tables.c index dd7d80feef..cf194696ac 100644 --- a/src/backend/utils/misc/guc_tables.c +++ b/src/backend/utils/misc/guc_tables.c @@ -1809,6 +1809,17 @@ struct config_bool ConfigureNamesBool[] = NULL, NULL, NULL }, + { + {"page_checksums64", PGC_INTERNAL, PRESET_OPTIONS, + gettext_noop("Shows whether 64-bit extended checksums are turned on for this cluster."), + NULL, + GUC_NOT_IN_SAMPLE | GUC_DISALLOW_IN_FILE | GUC_RUNTIME_COMPUTED + }, + &page_feature_page_checksums64, + false, + NULL, NULL, NULL + }, + { {"wasted_space", PGC_INTERNAL, PRESET_OPTIONS, gettext_noop("Waste some space in the page. Not even a fill factor. Just testing multiple page features."), diff --git a/src/bin/initdb/initdb.c b/src/bin/initdb/initdb.c index 0622983c06..c9267973b8 100644 --- a/src/bin/initdb/initdb.c +++ b/src/bin/initdb/initdb.c @@ -151,6 +151,7 @@ static bool sync_only = false; static bool show_setting = false; static bool data_checksums = false; static bool page_checksums32 = false; +static bool page_checksums64 = false; static bool waste_space = false; static char *xlog_dir = NULL; static char *str_wal_segment_size_mb = NULL; @@ -1324,11 +1325,12 @@ bootstrap_template1(void) unsetenv("PGCLIENTENCODING"); snprintf(cmd, sizeof(cmd), - "\"%s\" --boot -X %d %s %s %s %s %s %s", + "\"%s\" --boot -X %d %s %s %s %s %s %s %s", backend_exec, wal_segment_size_mb * (1024 * 1024), data_checksums ? "-k" : "", page_checksums32 ? "-e page_checksums32" : "", + page_checksums64 ? "-e page_checksums64" : "", waste_space ? "-e wasted_space" : "", boot_options, extra_options, debug ? "-d 5" : ""); @@ -2152,7 +2154,8 @@ usage(const char *progname) printf(_(" -g, --allow-group-access allow group read/execute on data directory\n")); printf(_(" --icu-locale=LOCALE set ICU locale ID for new databases\n")); printf(_(" -k, --data-checksums use data page checksums\n")); - printf(_(" -K, --extended-checksums use extended data page checksums\n")); + printf(_(" -K, --extended-checksums={32|64}\n" + " use extended data page checksums of the given bitsize\n")); printf(_(" --locale=LOCALE set default locale for new databases\n")); printf(_(" --lc-collate=, --lc-ctype=, --lc-messages=LOCALE\n" " --lc-monetary=, --lc-numeric=, --lc-time=LOCALE\n" @@ -2808,7 +2811,7 @@ main(int argc, char *argv[]) {"waldir", required_argument, NULL, 'X'}, {"wal-segsize", required_argument, NULL, 12}, {"data-checksums", no_argument, NULL, 'k'}, - {"extended-checksums", no_argument, NULL, 'K'}, + {"extended-checksums", required_argument, NULL, 'K'}, {"waste-space", no_argument, NULL, 'w'}, {"allow-group-access", no_argument, NULL, 'g'}, {"discard-caches", no_argument, NULL, 14}, @@ -2855,7 +2858,7 @@ main(int argc, char *argv[]) /* process command-line options */ - while ((c = getopt_long(argc, argv, "A:dD:E:gkKL:nNsST:U:WwX:", long_options, &option_index)) != -1) + while ((c = getopt_long(argc, argv, "A:dD:E:gkK:L:nNsST:U:WwX:", long_options, &option_index)) != -1) { switch (c) { @@ -2908,7 +2911,22 @@ main(int argc, char *argv[]) data_checksums = true; break; case 'K': - page_checksums32 = true; + { + if (!strcmp(optarg,"32")) + { + page_checksums32 = true; + page_checksums64 = false; + } + else if (!strcmp(optarg,"64")) + { + page_checksums32 = false; + page_checksums64 = true; + } + else + { + pg_fatal("Must provide \"32\" or \"64\" as extended checksum size"); + } + } break; case 'L': share_path = pg_strdup(optarg); @@ -3028,7 +3046,7 @@ main(int argc, char *argv[]) if (pwprompt && pwfilename) pg_fatal("password prompt and password file cannot be specified together"); - if (data_checksums && page_checksums32) + if (data_checksums && (page_checksums32 || page_checksums64) ) pg_fatal("data checksums and extended data checksums cannot be specified together"); check_authmethod_unspecified(&authmethodlocal); diff --git a/src/bin/pg_controldata/pg_controldata.c b/src/bin/pg_controldata/pg_controldata.c index 56eb8d88a2..8f77ff1be2 100644 --- a/src/bin/pg_controldata/pg_controldata.c +++ b/src/bin/pg_controldata/pg_controldata.c @@ -332,7 +332,9 @@ main(int argc, char *argv[]) printf(_("Reserved page size for features: %d\n"), CalculateReservedPageSize(ControlFile->page_features)); printf(_("Using extended checksums: %s\n"), + PageFeatureSetHasFeature(ControlFile->page_features, PF_PAGE_CHECKSUMS64) \ + ? _("64-bit") : PageFeatureSetHasFeature(ControlFile->page_features, PF_PAGE_CHECKSUMS32) \ - ? _("yes") : _("no")); + ? _("32-bit") : _("no")); return 0; } diff --git a/src/common/pagefeat.c b/src/common/pagefeat.c index 1af22b9876..2a42f5e57b 100644 --- a/src/common/pagefeat.c +++ b/src/common/pagefeat.c @@ -22,6 +22,7 @@ PageFeatureSet cluster_page_features; /* status GUCs, display only. set by XLog startup */ bool page_feature_page_checksums32; +bool page_feature_page_checksums64; bool page_feature_wasted_space; /* @@ -50,6 +51,8 @@ typedef struct PageFeatureDesc static PageFeatureDesc feature_descs[PF_MAX_FEATURE] = { /* PF_PAGE_CHECKSUMS32 */ { 2, "page_checksums32" }, + /* PF_PAGE_CHECKSUMS64 */ + { 8, "page_checksums64" }, /* PF_WASTED_SPACE */ { 40, "wasted_space" }, }; diff --git a/src/include/common/komihash.h b/src/include/common/komihash.h new file mode 100644 index 0000000000..898a8d48a1 --- /dev/null +++ b/src/include/common/komihash.h @@ -0,0 +1,565 @@ +/** + * komihash.h version 4.3.1 + * + * The inclusion file for the "komihash" hash function. + * + * Description is available at https://github.com/avaneev/komihash + * + * License + * + * Copyright (c) 2021-2022 Aleksey Vaneev + * + * Permission is hereby granted, free of charge, to any person obtaining a + * copy of this software and associated documentation files (the "Software"), + * to deal in the Software without restriction, including without limitation + * the rights to use, copy, modify, merge, publish, distribute, sublicense, + * and/or sell copies of the Software, and to permit persons to whom the + * Software is furnished to do so, subject to the following conditions: + * + * The above copyright notice and this permission notice shall be included in + * all copies or substantial portions of the Software. + * + * THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR + * IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY, + * FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT. IN NO EVENT SHALL THE + * AUTHORS OR COPYRIGHT HOLDERS BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER + * LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING + * FROM, OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER + * DEALINGS IN THE SOFTWARE. + */ + +#ifndef KOMIHASH_INCLUDED +#define KOMIHASH_INCLUDED + +#include +#include + +// Macros that apply byte-swapping. + +#if defined( __GNUC__ ) || defined( __clang__ ) + + #define KOMIHASH_BYTESW32( v ) __builtin_bswap32( v ) + #define KOMIHASH_BYTESW64( v ) __builtin_bswap64( v ) + +#elif defined( _MSC_VER ) + + #define KOMIHASH_BYTESW32( v ) _byteswap_ulong( v ) + #define KOMIHASH_BYTESW64( v ) _byteswap_uint64( v ) + +#else // defined( _MSC_VER ) + + #define KOMIHASH_BYTESW32( v ) ( \ + ( v & 0xFF000000 ) >> 24 | \ + ( v & 0x00FF0000 ) >> 8 | \ + ( v & 0x0000FF00 ) << 8 | \ + ( v & 0x000000FF ) << 24 ) + + #define KOMIHASH_BYTESW64( v ) ( \ + ( v & 0xFF00000000000000 ) >> 56 | \ + ( v & 0x00FF000000000000 ) >> 40 | \ + ( v & 0x0000FF0000000000 ) >> 24 | \ + ( v & 0x000000FF00000000 ) >> 8 | \ + ( v & 0x00000000FF000000 ) << 8 | \ + ( v & 0x0000000000FF0000 ) << 24 | \ + ( v & 0x000000000000FF00 ) << 40 | \ + ( v & 0x00000000000000FF ) << 56 ) + +#endif // defined( _MSC_VER ) + +// Endianness-definition macro, can be defined externally (e.g. =1, if +// endianness-correction is unnecessary in any case, to reduce its associated +// overhead). + +#if !defined( KOMIHASH_LITTLE_ENDIAN ) + #if defined( _WIN32 ) || defined( __LITTLE_ENDIAN__ ) || \ + ( defined( __BYTE_ORDER__ ) && __BYTE_ORDER__ == __ORDER_LITTLE_ENDIAN__ ) + + #define KOMIHASH_LITTLE_ENDIAN 1 + + #elif defined( __BIG_ENDIAN__ ) || \ + ( defined( __BYTE_ORDER__ ) && __BYTE_ORDER__ == __ORDER_BIG_ENDIAN__ ) + + #define KOMIHASH_LITTLE_ENDIAN 0 + + #else // defined( __BIG_ENDIAN__ ) + + #warning KOMIHASH: cannot determine endianness, assuming little-endian. + + #define KOMIHASH_LITTLE_ENDIAN 1 + + #endif // defined( __BIG_ENDIAN__ ) +#endif // !defined( KOMIHASH_LITTLE_ENDIAN ) + +// Macros that apply byte-swapping, used for endianness-correction. + +#if KOMIHASH_LITTLE_ENDIAN + + #define KOMIHASH_EC32( v ) ( v ) + #define KOMIHASH_EC64( v ) ( v ) + +#else // KOMIHASH_LITTLE_ENDIAN + + #define KOMIHASH_EC32( v ) KOMIHASH_BYTESW32( v ) + #define KOMIHASH_EC64( v ) KOMIHASH_BYTESW64( v ) + +#endif // KOMIHASH_LITTLE_ENDIAN + +// Likelihood macros that are used for manually-guided micro-optimization. + +#if defined( __GNUC__ ) || defined( __clang__ ) + + #define KOMIHASH_LIKELY( x ) __builtin_expect( x, 1 ) + #define KOMIHASH_UNLIKELY( x ) __builtin_expect( x, 0 ) + +#else // likelihood macros + + #define KOMIHASH_LIKELY( x ) ( x ) + #define KOMIHASH_UNLIKELY( x ) ( x ) + +#endif // likelihood macros + +// In-memory data prefetch macro (temporal locality=1, in case a collision +// resolution would be necessary). + +#if defined( __GNUC__ ) || defined( __clang__ ) + + #define KOMIHASH_PREFETCH( addr ) __builtin_prefetch( addr, 0, 1 ) + +#else // prefetch macro + + #define KOMIHASH_PREFETCH( addr ) + +#endif // prefetch macro + +/** + * An auxiliary function that returns an unsigned 32-bit value created out of + * a sequence of bytes in memory. This function is used to convert endianness + * of in-memory 32-bit unsigned values, and to avoid unaligned memory + * accesses. + * + * @param p Pointer to 4 bytes in memory. Alignment is unimportant. + */ + +static inline uint32_t kh_lu32ec( const uint8_t* const p ) +{ + uint32_t v; + memcpy( &v, p, 4 ); + + return( KOMIHASH_EC32( v )); +} + +/** + * An auxiliary function that returns an unsigned 64-bit value created out of + * a sequence of bytes in memory. This function is used to convert endianness + * of in-memory 64-bit unsigned values, and to avoid unaligned memory + * accesses. + * + * @param p Pointer to 8 bytes in memory. Alignment is unimportant. + */ + +static inline uint64_t kh_lu64ec( const uint8_t* const p ) +{ + uint64_t v; + memcpy( &v, p, 8 ); + + return( KOMIHASH_EC64( v )); +} + +/** + * Function builds an unsigned 64-bit value out of remaining bytes in a + * message, and pads it with the "final byte". This function can only be + * called if less than 8 bytes are left to read. The message should be "long", + * permitting Msg[ -3 ] reads. + * + * @param Msg Message pointer, alignment is unimportant. + * @param MsgLen Message's remaining length, in bytes; can be 0. + * @param fb Final byte used for padding. + */ + +static inline uint64_t kh_lpu64ec_l3( const uint8_t* const Msg, + const size_t MsgLen, uint64_t fb ) +{ + if( MsgLen < 4 ) + { + const uint8_t* const Msg3 = Msg + MsgLen - 3; + const int ml8 = (int) ( MsgLen << 3 ); + const uint64_t m = (uint64_t) Msg3[ 0 ] | (uint64_t) Msg3[ 1 ] << 8 | + (uint64_t) Msg3[ 2 ] << 16; + + return( fb << ml8 | m >> ( 24 - ml8 )); + } + + const int ml8 = (int) ( MsgLen << 3 ); + const uint64_t mh = kh_lu32ec( Msg + MsgLen - 4 ); + const uint64_t ml = kh_lu32ec( Msg ); + + return( fb << ml8 | ml | ( mh >> ( 64 - ml8 )) << 32 ); +} + +/** + * Function builds an unsigned 64-bit value out of remaining bytes in a + * message, and pads it with the "final byte". This function can only be + * called if less than 8 bytes are left to read. Can be used on "short" + * messages, but MsgLen should be greater than 0. + * + * @param Msg Message pointer, alignment is unimportant. + * @param MsgLen Message's remaining length, in bytes; cannot be 0. + * @param fb Final byte used for padding. + */ + +static inline uint64_t kh_lpu64ec_nz( const uint8_t* const Msg, + const size_t MsgLen, uint64_t fb ) +{ + if( MsgLen < 4 ) + { + fb <<= ( MsgLen << 3 ); + uint64_t m = Msg[ 0 ]; + + if( MsgLen > 1 ) + { + m |= (uint64_t) Msg[ 1 ] << 8; + + if( MsgLen > 2 ) + { + m |= (uint64_t) Msg[ 2 ] << 16; + } + } + + return( fb | m ); + } + + const int ml8 = (int) ( MsgLen << 3 ); + const uint64_t mh = kh_lu32ec( Msg + MsgLen - 4 ); + const uint64_t ml = kh_lu32ec( Msg ); + + return( fb << ml8 | ml | ( mh >> ( 64 - ml8 )) << 32 ); +} + +/** + * Function builds an unsigned 64-bit value out of remaining bytes in a + * message, and pads it with the "final byte". This function can only be + * called if less than 8 bytes are left to read. The message should be "long", + * permitting Msg[ -4 ] reads. + * + * @param Msg Message pointer, alignment is unimportant. + * @param MsgLen Message's remaining length, in bytes; can be 0. + * @param fb Final byte used for padding. + */ + +static inline uint64_t kh_lpu64ec_l4( const uint8_t* const Msg, + const size_t MsgLen, uint64_t fb ) +{ + if( MsgLen < 5 ) + { + const int ml8 = (int) ( MsgLen << 3 ); + + return( fb << ml8 | + (uint64_t) kh_lu32ec( Msg + MsgLen - 4 ) >> ( 32 - ml8 )); + } + else + { + const int ml8 = (int) ( MsgLen << 3 ); + + return( fb << ml8 | kh_lu64ec( Msg + MsgLen - 8 ) >> ( 64 - ml8 )); + } +} + +#if defined( __SIZEOF_INT128__ ) + + /** + * 64-bit by 64-bit unsigned multiplication. + * + * @param m1 Multiplier 1. + * @param m2 Multiplier 2. + * @param[out] rl The lower half of the 128-bit result. + * @param[out] rh The higher half of the 128-bit result. + */ + + static inline void kh_m128( const uint64_t m1, const uint64_t m2, + uint64_t* const rl, uint64_t* const rh ) + { + const __uint128_t r = (__uint128_t) m1 * m2; + + *rl = (uint64_t) r; + *rh = (uint64_t) ( r >> 64 ); + } + +#elif defined( _MSC_VER ) && defined( _M_X64 ) + + #include + + static inline void kh_m128( const uint64_t m1, const uint64_t m2, + uint64_t* const rl, uint64_t* const rh ) + { + *rl = _umul128( m1, m2, rh ); + } + +#else // defined( _MSC_VER ) + + // _umul128() code for 32-bit systems, adapted from mullu(), + // from https://go.dev/src/runtime/softfloat64.go + // Licensed under BSD-style license. + + static inline uint64_t kh__emulu( const uint32_t x, const uint32_t y ) + { + return( x * (uint64_t) y ); + } + + static inline void kh_m128( const uint64_t u, const uint64_t v, + uint64_t* const rl, uint64_t* const rh ) + { + *rl = u * v; + + const uint32_t u0 = (uint32_t) u; + const uint32_t v0 = (uint32_t) v; + const uint64_t w0 = kh__emulu( u0, v0 ); + const uint32_t u1 = (uint32_t) ( u >> 32 ); + const uint32_t v1 = (uint32_t) ( v >> 32 ); + const uint64_t t = kh__emulu( u1, v0 ) + ( w0 >> 32 ); + const uint64_t w1 = (uint32_t) t + kh__emulu( u0, v1 ); + + *rh = kh__emulu( u1, v1 ) + ( w1 >> 32 ) + ( t >> 32 ); + } + +#endif // defined( _MSC_VER ) + +// Common hashing round with 16-byte input, using the "r1l" and "r1h" +// temporary variables. + +#define KOMIHASH_HASH16( m ) \ + kh_m128( Seed1 ^ kh_lu64ec( m ), \ + Seed5 ^ kh_lu64ec( m + 8 ), &r1l, &r1h ); \ + Seed5 += r1h; \ + Seed1 = Seed5 ^ r1l; + +// Common hashing round without input, using the "r2l" and "r2h" temporary +// variables. + +#define KOMIHASH_HASHROUND() \ + kh_m128( Seed1, Seed5, &r2l, &r2h ); \ + Seed5 += r2h; \ + Seed1 = Seed5 ^ r2l; + +// Common hashing finalization round, with the final hashing input expected in +// the "r2l" and "r2h" temporary variables. + +#define KOMIHASH_HASHFIN() \ + kh_m128( r2l, r2h, &r1l, &r1h ); \ + Seed5 += r1h; \ + Seed1 = Seed5 ^ r1l; \ + KOMIHASH_HASHROUND(); + +/** + * KOMIHASH hash function. Produces and returns a 64-bit hash value of the + * specified message, string, or binary data block. Designed for 64-bit + * hash-table and hash-map uses. Produces identical hashes on both big- and + * little-endian systems. + * + * @param Msg0 The message to produce a hash from. The alignment of this + * pointer is unimportant. + * @param MsgLen Message's length, in bytes. + * @param UseSeed Optional value, to use instead of the default seed. To use + * the default seed, set to 0. The UseSeed value can have any bit length and + * statistical quality, and is used only as an additional entropy source. May + * need endianness-correction if this value is shared between big- and + * little-endian systems. + */ + +static inline uint64_t komihash( const void* const Msg0, size_t MsgLen, + const uint64_t UseSeed ) +{ + const uint8_t* Msg = (const uint8_t*) Msg0; + + // The seeds are initialized to the first mantissa bits of PI. + + uint64_t Seed1 = 0x243F6A8885A308D3 ^ ( UseSeed & 0x5555555555555555 ); + uint64_t Seed5 = 0x452821E638D01377 ^ ( UseSeed & 0xAAAAAAAAAAAAAAAA ); + uint64_t r1l, r1h, r2l, r2h; + + // The three instructions in the "KOMIHASH_HASHROUND" macro represent the + // simplest constant-less PRNG, scalable to any even-sized state + // variables, with the `Seed1` being the PRNG output (2^64 PRNG period). + // It passes `PractRand` tests with rare non-systematic "unusual" + // evaluations. + // + // To make this PRNG reliable, self-starting, and eliminate a risk of + // stopping, the following variant can be used, which is a "register + // checker-board", a source of raw entropy. The PRNG is available as the + // komirand() function. Not required for hashing (but works for it) since + // the input entropy is usually available in abundance during hashing. + // + // Seed5 += r2h + 0xAAAAAAAAAAAAAAAA; + // + // (the `0xAAAA...` constant should match register's size; essentially, + // it is a replication of the `10` bit-pair; it is not an arbitrary + // constant). + + KOMIHASH_HASHROUND(); // Required for PerlinNoise. + + if( KOMIHASH_LIKELY( MsgLen < 16 )) + { + KOMIHASH_PREFETCH( Msg ); + + r2l = Seed1; + r2h = Seed5; + + if( MsgLen > 7 ) + { + // The following two XOR instructions are equivalent to mixing a + // message with a cryptographic one-time-pad (bitwise modulo 2 + // addition). Message's statistics and distribution are thus + // unimportant. + + r2h ^= kh_lpu64ec_l3( Msg + 8, MsgLen - 8, + 1 << ( Msg[ MsgLen - 1 ] >> 7 )); + + r2l ^= kh_lu64ec( Msg ); + } + else + if( KOMIHASH_LIKELY( MsgLen != 0 )) + { + r2l ^= kh_lpu64ec_nz( Msg, MsgLen, + 1 << ( Msg[ MsgLen - 1 ] >> 7 )); + } + + KOMIHASH_HASHFIN(); + + return( Seed1 ); + } + + if( KOMIHASH_LIKELY( MsgLen < 32 )) + { + KOMIHASH_PREFETCH( Msg ); + + KOMIHASH_HASH16( Msg ); + + const uint64_t fb = 1 << ( Msg[ MsgLen - 1 ] >> 7 ); + + if( MsgLen > 23 ) + { + r2h = Seed5 ^ kh_lpu64ec_l4( Msg + 24, MsgLen - 24, fb ); + r2l = Seed1 ^ kh_lu64ec( Msg + 16 ); + } + else + { + r2l = Seed1 ^ kh_lpu64ec_l4( Msg + 16, MsgLen - 16, fb ); + r2h = Seed5; + } + + KOMIHASH_HASHFIN(); + + return( Seed1 ); + } + + if( MsgLen > 63 ) + { + uint64_t Seed2 = 0x13198A2E03707344 ^ Seed1; + uint64_t Seed3 = 0xA4093822299F31D0 ^ Seed1; + uint64_t Seed4 = 0x082EFA98EC4E6C89 ^ Seed1; + uint64_t Seed6 = 0xBE5466CF34E90C6C ^ Seed5; + uint64_t Seed7 = 0xC0AC29B7C97C50DD ^ Seed5; + uint64_t Seed8 = 0x3F84D5B5B5470917 ^ Seed5; + uint64_t r3l, r3h, r4l, r4h; + + do + { + KOMIHASH_PREFETCH( Msg ); + + kh_m128( Seed1 ^ kh_lu64ec( Msg ), + Seed5 ^ kh_lu64ec( Msg + 8 ), &r1l, &r1h ); + + kh_m128( Seed2 ^ kh_lu64ec( Msg + 16 ), + Seed6 ^ kh_lu64ec( Msg + 24 ), &r2l, &r2h ); + + kh_m128( Seed3 ^ kh_lu64ec( Msg + 32 ), + Seed7 ^ kh_lu64ec( Msg + 40 ), &r3l, &r3h ); + + kh_m128( Seed4 ^ kh_lu64ec( Msg + 48 ), + Seed8 ^ kh_lu64ec( Msg + 56 ), &r4l, &r4h ); + + Msg += 64; + MsgLen -= 64; + + // Such "shifting" arrangement (below) does not increase + // individual SeedN's PRNG period beyond 2^64, but reduces a + // chance of any occassional synchronization between PRNG lanes + // happening. Practically, Seed1-4 together become a single + // "fused" 256-bit PRNG value, having a summary PRNG period of + // 2^66. + + Seed5 += r1h; + Seed6 += r2h; + Seed7 += r3h; + Seed8 += r4h; + Seed2 = Seed5 ^ r2l; + Seed3 = Seed6 ^ r3l; + Seed4 = Seed7 ^ r4l; + Seed1 = Seed8 ^ r1l; + + } while( KOMIHASH_LIKELY( MsgLen > 63 )); + + Seed5 ^= Seed6 ^ Seed7 ^ Seed8; + Seed1 ^= Seed2 ^ Seed3 ^ Seed4; + } + + KOMIHASH_PREFETCH( Msg ); + + if( KOMIHASH_LIKELY( MsgLen > 31 )) + { + KOMIHASH_HASH16( Msg ); + KOMIHASH_HASH16( Msg + 16 ); + + Msg += 32; + MsgLen -= 32; + } + + if( MsgLen > 15 ) + { + KOMIHASH_HASH16( Msg ); + + Msg += 16; + MsgLen -= 16; + } + + const uint64_t fb = 1 << ( Msg[ MsgLen - 1 ] >> 7 ); + + if( MsgLen > 7 ) + { + r2h = Seed5 ^ kh_lpu64ec_l4( Msg + 8, MsgLen - 8, fb ); + r2l = Seed1 ^ kh_lu64ec( Msg ); + } + else + { + r2l = Seed1 ^ kh_lpu64ec_l4( Msg, MsgLen, fb ); + r2h = Seed5; + } + + KOMIHASH_HASHFIN(); + + return( Seed1 ); +} + +/** + * Simple, reliable, self-starting yet efficient PRNG, with 2^64 period. + * 0.62 cycles/byte performance. Self-starts in 4 iterations, which is a + * suggested "warming up" initialization before using its output. + * + * @param[in,out] Seed1 Seed value 1. Can be initialized to any value + * (even 0). This is the usual "PRNG seed" value. + * @param[in,out] Seed2 Seed value 2, a supporting variable. Best initialized + * to the same value as Seed1. + * @return The next uniformly-random 64-bit value. + */ + +static inline uint64_t komirand( uint64_t* const Seed1, uint64_t* const Seed2 ) +{ + uint64_t r1l, r1h; + + kh_m128( *Seed1, *Seed2, &r1l, &r1h ); + *Seed2 += r1h + 0xAAAAAAAAAAAAAAAA; + *Seed1 = *Seed2 ^ r1l; + + return( *Seed1 ); +} + +#endif // KOMIHASH_INCLUDED diff --git a/src/include/common/pagefeat.h b/src/include/common/pagefeat.h index d6a665cd8f..47331e71ed 100644 --- a/src/include/common/pagefeat.h +++ b/src/include/common/pagefeat.h @@ -17,6 +17,7 @@ /* revealed for GUCs */ extern int reserved_page_size; extern bool page_feature_page_checksums32; +extern bool page_feature_page_checksums64; extern bool page_feature_wasted_space; /* forward declaration to avoid circular includes */ @@ -30,6 +31,7 @@ extern PageFeatureSet cluster_page_features; /* bit offset for features flags */ typedef enum { PF_PAGE_CHECKSUMS32 = 0, + PF_PAGE_CHECKSUMS64, PF_WASTED_SPACE, PF_MAX_FEATURE /* must be last */ } PageFeature; diff --git a/src/include/storage/checksum.h b/src/include/storage/checksum.h index d0bcc01ca6..f131c4d492 100644 --- a/src/include/storage/checksum.h +++ b/src/include/storage/checksum.h @@ -21,5 +21,6 @@ */ extern uint16 pg_checksum_page(char *page, BlockNumber blkno); extern uint32 pg_checksum32_page(char *page, BlockNumber blkno, char*offset); +extern uint64 pg_checksum64_page(char *page, BlockNumber blkno, uint64*offset); #endif /* CHECKSUM_H */ diff --git a/src/include/storage/checksum_impl.h b/src/include/storage/checksum_impl.h index 87d9f96484..2ce2c8bb29 100644 --- a/src/include/storage/checksum_impl.h +++ b/src/include/storage/checksum_impl.h @@ -101,6 +101,7 @@ */ #include "storage/bufpage.h" +#include "common/komihash.h" /* number of checksums to calculate in parallel */ #define N_SUMS 32 @@ -138,6 +139,7 @@ do { \ (checksum) = __tmp * FNV_PRIME ^ (__tmp >> 17); \ } while (0) + /* * Block checksum algorithm. The page must be adequately aligned * (at least on 4-byte boundary). @@ -173,6 +175,8 @@ pg_checksum_block(const PGChecksummablePage *page) return result; } + + /* * Compute the checksum for a Postgres page. * @@ -270,3 +274,61 @@ pg_checksum32_page(char *page, BlockNumber blkno, char *cksum) * coset for group 1 here, but it's a nice property to preserve */ return (checksum == 0 ? 1 : checksum); } + +/* + * 64-bit block checksum algorithm. The page must be adequately aligned + * (at least on 4-byte boundary). + */ + +static uint64 +pg_checksum64_block(const PGChecksummablePage *page) +{ + /* ensure that the size is compatible with the algorithm */ + Assert(sizeof(PGChecksummablePage) == BLCKSZ); + + return (uint64)komihash(page, BLCKSZ, 0); +} + + +/* + * Compute and return a 64-bit checksum for a Postgres page. + * + * Beware that the 64-bit portion of the page that cksum points to is + * transiently zeroed, though it is restored. + * + * The checksum includes the block number (to detect the case where a page is + * somehow moved to a different location), the page header (excluding the + * checksum itself), and the page data. + * + * The high bits of this are stored in the overflow storage area of the page + * pointed to by *cksum, leaving the pd_checksum field with the same checksum + * you'd expect if running the pg_checksum_page function. + */ +uint64 +pg_checksum64_page(char *page, BlockNumber blkno, uint64 *cksumloc) +{ + PGChecksummablePage *cpage = (PGChecksummablePage *) page; + uint64 saved; + uint64 checksum; + + /* We only calculate the checksum for properly-initialized pages */ + Assert(!PageIsNew((Page) page)); + /* Ensure that the cksum pointer is in the page range on this page */ + Assert((char*)cksumloc >= page && (char*)cksumloc <= (page + BLCKSZ - sizeof(uint64))); + + saved = *cksumloc; + *cksumloc = 0; + + checksum = pg_checksum64_block(cpage); + + /* restore */ + *cksumloc = saved; + + /* Mix in the block number to detect transposed pages */ + checksum ^= blkno; + + /* ensure in the extremely unlikely case that we have non-zero return + * value here; this does double-up on our coset for group 1 here, but it's + * a nice property to preserve */ + return (checksum == 0 ? 1 : checksum); +} -- 2.37.0 (Apple Git-136)