MySQL NDB Cluster 8.0.22 is a new release of NDB 8.0, based on
MySQL Server 8.0 and including features in version 8.0 of the
NDB storage engine, as well as fixing
recently discovered bugs in previous NDB Cluster releases.
Obtaining NDB Cluster 8.0. NDB Cluster 8.0 source code and binaries can be obtained from https://dev.mysql.com/downloads/cluster/.
For an overview of changes made in NDB Cluster 8.0, see What is New in MySQL NDB Cluster 8.0.
This release also incorporates all bug fixes and changes made in previous NDB Cluster releases, as well as all bug fixes and feature changes which were added in mainline MySQL 8.0 through MySQL 8.0.22 (see Changes in MySQL 8.0.22 (2020-10-19, General Availability)).
To provide protection against unauthorized recovery of data from backups, this release adds support for
NDBnative encrypted backup using AES-256-CBC. Encrypted backup files are protected by a user-supplied password.
NDBdoes not save this password; this needs to be done by the user or application. To create an encrypted backup, use
ENCRYPT PASSWORD=with the ndb_mgm client
START BACKUPcommand (in addition to any other options which may be required). You can also initiate an encrypted backup in applications by calling the MGM API
To restore from an encrypted backup, use ndb_restore with both of the options
--backup-password=. ndb_print_backup_file can also read encrypted files using the
-Poption added in this release.
The encryption password used with this feature can be any string of up to 256 characters from the range of printable ASCII characters other than
^. When a password is supplied for encryption or decryption, it must be quoted using either single or double quotation marks. It is possible to specify an empty password using
""but this is not recommended.
You can encrypt existing backup files using the ndbxfrm utility which is added to the NDB Cluster distribution in this release; this program can also decrypt encrypted backup files. ndbxfrm also compresses and decompresses NDB Cluster backup files. The compression method is the same as used by NDB Cluster for creating compressed backups when
It is also possible to require encrypted backups using
RequireEncryptedBackup. When this parameter is enabled (by setting it equal to 1), the management client rejects any attempt to perform a backup that is not encrypted.
For more information, see Using The NDB Cluster Management Client to Create a Backup, as well as ndbxfrm — Compress, Decompress, Encrypt, and Decrypt Files Created by NDB Cluster. (WL #13474, WL #13499, WL #13548)
NDB Client Programs: Effective with this release, the MySQL NDB Cluster Auto-Installer (ndb_setup.py) has been deprecated and is subject to removal in a future version of NDB Cluster. (Bug #31888835)
ndbmemcacheis deprecated in this release of NDB Cluster, and is scheduled for removal in the next release. (Bug #31876970)
Important Change: The
Ndb_metadata_blacklist_sizestatus variable was renamed as
Ndb_metadata_excluded_count. (Bug #31465469)
Packaging: Made the following improvements to the
server-minimalRPM for NDB Cluster and the NDB Cluster Docker image:
Added ndb_import and other helpful utilities.
Included NDB utilities are now linked dynamically.
The NDB Cluster Auto-Installer is no longer included.
ndbmemcacheis no longer included.
NDB Replication: Batching of updates to rows containing columns of MySQL type
LONGTEXT(“Blob”) by NDB Cluster. This affects
DELETEstatements of either of the following types:
Statements which modify multiple blob columns in the same row
Statements which modify multiple rows containing blob columns in the same statement
This is accomplished by greatly reducing the number of round trips required between an SQL or other API node and the data nodes in the replica cluster, in some cases by a factor of 10 or more.
Other SQL statements may also see performance benefits from these improvements. Such statements include
LOAD DATA INFILEand
CREATE TABLE ... SELECT ...when acting on tables containing one or more Blob columns. In addition, an
ALTER TABLE ... ENGINE = NDBstatement which changes the storage engine of a table that previously used one other than
NDBand that contains one or more Blob columns may also execute more efficiently than before this enhancement was implemented.
The performance of some SQL statements which update Blob columns is not noticeably improved by this enhancement, due to the fact that they require scans of table Blob columns, which breaks up batching. Such statements include those of the types listed here:
SELECTwhich filters rows by matching on a primary key or unique key column which uses a Blob type
WHEREcondition which does not depend on a unique value
ALTER TABLEstatement on a table which already used the
NDBstorage engine prior to executing the statement
To take maximum advantage of this improvement, you must enable
slave_allow_batching. It is also recommended that you increase the values used with the
--ndb-blob-write-batch-bytesMySQL server options to minimize the number of round trips required by the replica cluster to apply epoch transactions. (Bug #27765184, WL #13043)
Added the CMake option
NDB_UTILS_LINK_DYNAMIC, to allow dynamic linking of NDB utilities with
ndbclient. (Bug #31668306)
IPv6 addressing is now supported for connections to management and data nodes, including connections between management and data nodes with SQL nodes. For IPv6 addressing to work, the operating platform and network on which the cluster is deployed must support IPv6. Hostname resolution to IPv6 addresses must be provided by the operating platform (this is the same as when using IPv4 addressing).
Management node configured with IPv6, data nodes configured with IPv4: This works if the data nodes are started with
--ndb-connectstringset to the IPv4 address of the management nodes.
Management node configured with IPv4, data nodes configured with IPv6: This works if the data nodes are started with
--ndb-connectstringset to the IPv6 address of the management node.
When upgrading from an NDB version that does not support IPv6 addressing to a version that does so, it is necessary that the network already support both IPv4 and IPv6. The software upgrade must be performed first; after this, you can update the IPv4 addresses used in the
config.iniconfiguration file with the desired IPv6 addresses. Finally, in order for the configuration changes to take effect, perform a system restart of the cluster. (WL #12963)
Important Change; NDB Cluster APIs: The NDB Cluster adapter for Node.js was built against an obsolete version of the runtime. Now it is built using Node.js 12.18.3, and only that version or a later version of Node.js is supported by
NDB. (Bug #31783049)
Important Change: In order to synchronize excluded metadata objects, it was necessary to correct the underlying issue, if any, and then trigger the synchronization of the objects again. This could be achieved though discovery of individual tables, which does not scale well with an increase in the number of tables and SQL nodes. It could also be done by reconnecting the SQL node to the cluster, but doing so also incurs extra overhead.
To fix this issue, the list of database objects excluded due to synchronization failure is cleared when
ndb_metadata_syncis enabled by the user. This makes all such objects eligible for synchronization in the subsequent detection run, which simplifies retrying the synchronization of all excluded objects.
This fix also removes the validation of objects to be retried which formerly took take place at the beginning of each detection run. Since these objects are of interest only while
ndb_metadata_syncis enabled, the list of objects to be retried is cleared when this variable is disabled, signalling that all changes have been synchronized. (Bug #31569436)
Packaging: The Dojo library included with NDB Cluster has been upgraded to version 1.15.4. (Bug #31559518)
NDB Disk Data: ndbmtd sometimes terminated unexpectedly when it could not complete a lookup for a log file group during a restore operation. (Bug #31284086)
NDB Disk Data: While upgrading a cluster having 3 or 4 replicas after creating sufficient disk data objects to fill up the tablespace, and while performing inserts on the disk data tables, trying to stop some data nodes caused others to exit improperly. (Bug #30922322)
NDB Replication: On Unix-based operating systems, binary logs can be flushed by sending a
SIGHUPsignal to the server, but
NDBCLUSTERexpected one of the SQL statements
SHOW BINLOG EVENTSonly. (Bug #31242689)
NDB Cluster APIs: In certain cases, the
Table::getColumn()method returned the wrong
Columnobject. This could happen when the full name of one table column was a prefix of the name of another, or when the names of two columns had the same hash value. (Bug #31774685)
NDB Cluster APIs: It was possible to make invalid sequences of NDB API method calls using blobs. This was because some method calls implicitly cause transaction execution inline, to deal with blob parts and other issues, which could cause user-defined operations not to be handled correctly due to the use of a method executing operations relating to blobs while there still user-defined blob operations pending. Now in such cases, NDB raises a new error 4558 Pending blob operations must be executed before this call. (Bug #27772916)
--remap-columndid not handle columns containing
NULLvalues correctly. Now any offset specified by the mapping function used with this option is not applied to
NULL, so that
NULLis preserved as expected. (Bug #31966676)
The ndb_print_backup_file utility did not respect byte order for row data. This tool now performs byte swapping on row page information to ensure the same results on both big-endian and little-endian platforms. (Bug #31831438)
References: See also: Bug #32470157.
In some cases following an upgrade from a version of NDB Cluster previous to 8.0.18 to a later one, writing the
sysfile(see NDB Cluster Data Node File System Directory) and reading back from it did not work correctly. This could occur when explicit node group assignments to data nodes had been made (using the
NodeGroupparameter); it was possible for node group assignments to change spontaneously, and even possible for node groups not referenced in the configuration file to be added. This was due to issues with version 2 of the
sysfileformat introduced in NDB 8.0.18. (Bug #31828452, Bug #31820201)
References: See also: Bug #31726653.
After encountering the data node in the configuration file which used
NodeGroup=65536, the management server stopped assigning data nodes lacking an explicit
NodeGroupsetting to node groups. (Bug #31825181)
Data nodes in certain cases experienced fatal memory corruption in the
PGMANkernel block due to an invalid assumption that pages were 32KB aligned, when in fact they are normally aligned to the system page size (4096 or 8192 bytes, depending on platform). (Bug #31768450, Bug #31773234)
Fixed a misspelled define introduced in NDB 8.0.20 which made an internal function used to control adaptive spinning non-operational. (Bug #31765660)
When executing undo log records during undo log recovery it was possible when hitting a page cache miss to use the previous undo log record multiple times. (Bug #31750627)
When an SQL node or cluster shutdown occurred during schema distribution while the coordinator was still waiting for the participants, the schema distribution was aborted halfway but any rows in
ndb_schema_resultrelated to this schema operation were not cleared. This left open the possibility that these rows might conflict with a future reply from a participant if a DDL operation having the same schema operation ID originated from a client using the same node ID.
To keep this from happening, we now clear all such rows in
NDBbinary log setup. This assures that there are no DDL distributions in progress and any rows remaining in the
ndb_schema_resulttable are already obsolete. (Bug #31601674)
Help output from the MySQL Cluster Auto-Installer displayed incorrect version information. (Bug #31589404)
In certain rare circumstances,
NDBmissed checking for completion of a local checkpoint, leaving it uncompleted, which meant that subsequent local checkpoints could not be executed. (Bug #31577633)
A data definition statement can sometimes involve reading or writing of multiple rows (or both) from tables;
NdbTransactionto perform these operations. When such a statement was rolled back,
NDBCLUSTERattempted to roll back the schema change before rolling back the
NdbTransactionand closing it; this led to the rollback hanging indefinitely while the cluster waited for the
NdbTransactionobject to close before it was able to roll back the schema change.
Now in such cases,
NDBCLUSTERrolls back the schema change only after rolling back and closing any open Ndb
Transactionassociated with the change. (Bug #31546868)
Adding a new user was not always synchronized correctly to all SQL nodes when the
NDB_STORED_USERprivilege was granted to the new user. (Bug #31486931)
SUMAon a node that is starting up should not send a
DICT_UNLOCK_ORDsignal to the
DICTblock on the master node until both all
SUMA_HANDOVER_REQsignals sent have had
SUMA_HANDOVER_CONFsignals sent in response, and every switchover bucket set up on receiving a
SUMA_HANDOVER_CONFhas completed switchover. In certain rare cases using
NoOfReplicas> 2, and in which the delay between global checkpoints was unusually short, it was possible for some switchover buckets to be ready for handover before others, and for handover to proceed even though this was the case. (Bug #31459930)
Attribute ID mapping needs to be performed when reading data from an
NDBtable using indexes or a primary key whose column order is different than that of the table. For unique indexes, a cached attribute ID map is created when the table is opened, and is then used for each subsequent read, but for primary key reads, the map was built for every read. This is changed so that an attribute ID map for primary key is built and cached when opening the table, and used whenever required for any subsequent reads. (Bug #31452597)
References: See also: Bug #24444899.
During different phases of the restore process, ndb_restore used different numbers of retries for temporary errors as well as different sleep times between retries. This is fixed by implementing consistent retry counts and sleep times across all restore phases. (Bug #31372923)
Removed warnings generated when compiling
NDBCLUSTERwith Clang 10. (Bug #31344788)
SPJblock contains a load throttling mechanism used when generating
LQHKEYREQsignals. When these were generated from parent rows from a scan, and this scan had a bushy topology with multiple children performing key lookups, it was possible to overload the job queues with too many
LQHKEYREQsignals, causing node shutdowns due to full job buffers. This problem was originally fixed by Bug #14709490. Further investigation of this issue showed that job buffer full errors could occur even if the
SPJquery was not bushy. Due to the increase in the internal batch size for
SPJworkers in NDB 7.6.4 as part of work done to implement use of multiple fragments when sending
SCAN_FRAGREQsignals to the
SPJblock, even a simple query could fill up the job buffers when a relatively small number of such queries were run in parallel.
To fix this problem, we no longer send any further
LQHKEYREQsignals once the number of outstanding signals in a given request exceeds 256. Instead, the parent row from which the
LQHKEYREQis produced is buffered, and the correlation ID of this row is stored in the collection of operations to be resumed later. (Bug #31343524)
References: This issue is a regression of: Bug #14709490.
MaxDiskWriteSpeedOwnRestartwas not honored as an upper bound for local checkpoint writes during a node restart. (Bug #31337487)
References: See also: Bug #29943227.
Under certain rare circumstances,
DROP TABLEof an
NDBtable triggered an assert. (Bug #31336431)
During a node restart, the
SUMAblock of the node that is starting must get a copy of the subscriptions (events with subscribers) and subscribers (
NdbEventOperationinstances which are executing) from a node already running. Before the copy is complete, nodes which are still starting ignore any user-level
SUB_STOPrequests; after the copy is done, they can participate in such requests. While the copy operation is in progress, user-level
SUB_STOPrequests are blocked using a
An issue was found whereby a starting node could participate in
SUB_STOPrequests after the lock was requested, but before it is granted, which resulted in unsuccessful
SUB_STOPrequests. This fix ensures that the nodes cannot participate in these requests until after the
DICTlock has actually been granted. (Bug #31302657)
Backups errored out with FsErrInvalidParameters when the filesystem was running with
O_DIRECTand a data file write was not aligned with the 512-byte block size used by
O_DIRECTwrites. If the total fragment size in the data file is not aligned with the
NDBpads the last write to the required size, but when there were no fragments to write,
BACKUPwrote only the header and footer to the data file. Since the header and footer are less than 512 bytes, leading to the issue with the
This is fixed by padding out the generic footer to 512 bytes if necessary, using an
EMPTY_ENTRY, when closing the data file. (Bug #31180508)
When employing an execution strategy which requires it to buffer received key rows for later use,
DBSPJnow manages the buffer memory allocation tree node by tree node, resulting in a significant drop in CPU usage by the
DBSPJblock. (Bug #31174015)
DBSPJnow uses linear memory instead of segmented memory for storing and handling
TRANSID_AIsignals, which saves approximately 10% of the CPU previously consumed. Due to this change, it is now possible for
TRANSID_AIsignals in the short signal format; this is more efficient than the long signal format which requires segmented memory. (Bug #31173582, Bug #31173766)
Altering the table comment of a fully replicated table using
ALGORITHM=INPLACEled to an assertion. (Bug #31139313)
A local data manager (LDM) has a mechanism for ensuring that a fragment scan does not continue indefinitely when it finds too few rows to fill the available batch size in a reasonable amount of time (such as when a ScanFilter evaluates to false for most of the scanned rows). When this time limit, set in
DBLQHas 10 ms, has expired, any rows found up to that point are returned, independent of whether the specified batch size has been filled or not. This acts as a keep-alive mechanism between data and API nodes, as well as to avoid keeping any locks held during the scan for too long.
A side effect of this is that returning result row batches to the
DBSPJblock which are filled well below the expected limit could cause performance issues. This was due not only to poor utilization of the space reserved for batches, requiring more
NEXTREQround trips, but because it also caused
DBSPJinternal parallelism statistics to become unreliable.
DBSPJblock never requests locks when performing scans, overly long locks are not a problem for SPJ requests. Thus it is considered safe to let scans requested by
DBSPJto continue for longer than the 10 ms allowed previously, and the limit set in
DBLQHhas been increased to 100 ms. (Bug #31124065)
For a pushed join, the output from
EXPLAIN FORMAT=TREEdid not indicate whether the table access was an index range scan returning multiple rows, or a single-row lookup on a primary or unique key.
This fix provides also a minor optimization, such that the handler interface is not accessed more than once in an attempt to return more than a single row if the access type is known to be
Unique. (Bug #31123930)
A previous change (made in NDB 8.0.20) made it possible for a pushed join on tables allowing
READ_BACKUPto place two SPJ workers on the data node local to the
DBTCblock while placing no SPJ workers on some other node; this sometime imbalance is intentional, as the SPJ workload (and possible introduced imbalance) is normally quite low compared to the gains of enabling more local reads of the backup fragments. As an unintended side effect of the same change, these two colocated SPJ workers might scan the same subset of fragments in parallel; this broke an assumption in the
DBSPJblock that only a single SPJ worker is instantiated on each data node on which the logic for insuring that each SPJ worker starts its scans from a different fragment depends.
To fix this problem, the starting fragment for each SPJ worker is now calculated based on the root fragment ID from which the worker starts, which is unique among all SPJ workers even when some of them reside on the same node. (Bug #31113005)
References: See also: Bug #30639165.
When upgrading a cluster from NDB 8.0.17 or earlier to 8.0.18 or later, data nodes not yet upgraded could shut down unexpectedly following upgrade of the management server (or management servers) to the new software version. This occurred when a management client
STOPcommand was sent to one or more of the data nodes still running the old version and the new master node (also running the old version of the
NDBsoftware) subsequently underwent an unplanned shutdown.
It was found that this occurred due to setting the signal length and number of signal sections incorrectly when sending a
GSN_STOP_REQ—one of a number of signals whose length has been increased in NDB 8.0 as part of work done to support greater numbers of data nodes—to the new master. This happened due to the use of stale data retained from sending a
GSN_STOP_REQto the previous master node. To prevent this from happening, ndb_mgmd now sets the signal length and number of sections explicitly each time, prior to sending a
GSN_STOP_REQsignal. (Bug #31019990)
In some cases, when failures occurred while replaying logs and restoring tuples, ndb_restore terminated instead of returning an error. In addition, the number of retries to be attempted for some operations was determined by hard-coded values. (Bug #30928114)
During schema distribution, if the client was killed after a DDL operation was already logged in the
ndb_schematable, but before the participants could reply, the client simply marked all participants as failed in the
NDB_SCHEMA_OBJECTand returned. Since the distribution protocol was already in progress, the coordinator continued to wait for the participants, received their
ndb_schema_resultinsert and processed them; meanwhile, the client was open to send another DDL operation; if one was executed and distribution of it was begun before the coordinator could finish processing the previous schema change, this triggered an assertion there should be only one distribution of a schema operation active at any given time.
In addition, when the client returned having detected a thread being killed, it also released the global schema lock (GSL); this could also lead to undefined issues since the participant could make the changes under the assumption that the GSL was still being held by the coordinator.
In such cases, the client should not return after the DDL operation has been logged in the
ndb_schematable; from this point, the coordinator has control and the client should wait for it to make a decision. Now the coordinator aborts the distribution only in the event of a server or cluster shutdown, and otherwise waits for all participants either to reply, or to time out and mark the schema operation as completed. (Bug #30684839)
When, during a restart, a data node received a
GCP_SAVEREQsignal prior to beginning start phase 9, and thus needed to perform a global checkpoint index write to a local data manager's local checkpoint control file, it did not record information from the
DIHblock originating with the node that sent the signal as part of the data written. This meant that, later in start phase 9, when attempting to send a
GCP_SAVECONFsignal in response to the
GCP_SAVEREQ, this information was not available, which meant the response could not be sent, resulting in an unplanned shutdown of the data node. (Bug #30187949)
References: See also: Bug #31337487.
BLOBvalue is stored by
NDBin multiple parts; when reading such a value, one read operation is executed per part. If a part is not found, the read fails with a row not found error, which indicates a corrupted
BLOB, since a
BLOBshould never have any missing parts. A problem can arise because this error is reported as the overall result of the read operation, which means that mysqld sees no error and reports zero rows returned.
This issue is fixed by adding a check specifically for the case in wich a blob part is not found. Now, when this occurs, overwriting the row not found error with corrupted blob, which causes the originating
SELECTstatement to fail as expected. Users of the NDB API should be aware that, despite this change, the
NdbBlob::getValue()method continues to report the error as row not found in such cases. (Bug #28590428)
Data nodes did not start when the
RealtimeSchedulerconfiguration parameter was set to 1. This was due to the fact that index builds during startup are performed by temporarily diverting some I/O threads for use as index building threads, and these threads inherited the realtime properties of the I/O threads. This caused a conflict (treated as a fatal error) when index build thread specifications were checked to ensure that they were not realtime threads. This is fixed by making sure that index build threads are not treated as realtime threads regardless of any settings applying to their host I/O threads, which is as actually intended in their design. (Bug #27533538)
Using an in-place
ALTER TABLEto drop an index could lead to the unplanned shutdown of an SQL node. (Bug #24444899)
As the final step when executing
ALTER TABLE ... ALGORITHM=INPLACE,
NDBCLUSTERperformed a read of the table metadata from the
NDBdictionary, requiring an extra round trip between the SQL nodes and data nodes, which unnecessarily both slowed down execution of the statement and provided an avenue for errors which
NDBCLUSTERwas not prepared to handle correctly. This issue is fixed by removing the read of
NDBtable metadata during the final phase of executing an in-place
ALTER TABLEstatement. (Bug #99898, Bug #31497026)
AllowUnresolvedHostNamesconfiguration parameter. When set to
true, this parameter overrides the fatal error normally raised when ndb_mgmd cannot connect to a given host name, allowing startup to continue and generating only a warning instead. To be effective, the parameter must be set in the cluster global configuration file's
[tcp default]section. (WL #13860)