Important Change: It is now possible to divide a backup into slices and to restore these in parallel using two new options implemented for the ndb_restore utility, making it possible to employ multiple instances of ndb_restore to restore subsets of roughly the same size of the backup in parallel, which should help to reduce the length of time required to restore an NDB Cluster from backup.
--num-slicesoptions determines the number of slices into which the backup should be divided;
--slice-idprovides the ID of the slice (0 to 1 less than the number of slices) to be restored by ndb_restore.
Up to 1024 slices are supported.
Important Change: To increase the rate at which update operations can be processed,
NDBnow supports and by default makes use of multiple transporters per node group. By default, the number of transporters used by each node group in the cluster is equal to the number of the number of local data management (LDM) threads. While this number should be optimal for most use cases, it can be adjusted by setting the value of the
NodeGroupTransportersdata node configuration parameter which is introduced in this release. The maximum is the greater of the number of LDM threads or the number of TC threads, up to an overall maximum of 32 transporters.
See Multiple Transporters, for additional information.
NDBnow supports versioning for
ndbinfotables, and maintains the current definitions for its tables internally. At startup,
NDBcompares its supported
ndbinfoversion with the version stored in the data dictionary. If the versions differ,
NDBdrops any old
ndbinfotables and recreates them using the current definitions.
Many outer joins and semijoins which previously could not be pushed down to the data nodes can now pushed (see Engine Condition Pushdown Optimization).
Outer joins which can now be pushed include those which meet the following conditions:
There are no unpushed conditions on this table
There are no unpushed conditions on other tables in the same join nest, or in upper join nests on which it depends
All other tables in the same join nest, or in upper join nests on which it depends are also pushed
A semijoin using an index scan can now be pushed if it meets the the conditions just noted for a pushed outer join, and it uses the
References: See also: Bug #28728603, Bug #28672214, Bug #29296615, Bug #29232744, Bug #29161281, Bug #28728007.
A new and simplified interface is implemented for enabling and configuring adaptive CPU spin. The
SpinMethoddata node parameter, added in this release, provides the following four settings:
StaticSpinning: Disables adaptive spinning; uses the static spinning employed in previous NDB Cluster releases
CostBasedSpinning: Enables adaptive spinning using a cost-based model
LatencyOptimisedSpinning: Enables adaptive spinning optimized for latency
DatabaseMachineSpinning: Enables adaptive spinning optimized for machines hosting databases, where each thread has its own CPU
Each of these settings causes the data node to use a set of predetermined values, as needed, for one or more of the spin parameters listed here:
SchedulerSpinTimer: The data node configuration parameter of this name.
EnableAdaptiveSpinning: Enables or disables adaptive spinning; cannot be set directly in the cluster configuration file, but can be controlled directly using
SetAllowedSpinOverhead: CPU time to allow to gain latency; cannot be set directly in the
config.inifile, but possible to change directly, using
The presets available from
SpinMethodshould cover most use cases, but you can fine-tune the adaptive spin behavior using the
SchedulerSpinTimerdata node configuration parameter and the
DUMPcommands just listed, as well as additional
DUMPcommands in the ndb_mgm cluster management client; see the description of
SchedulerSpinTimerfor a complete listing.
NDB 8.0.20 also adds a new TCP configuration parameter
TcpSpinTimewhich sets the time to spin for a given TCP connection. This can be used to enable adaptive spinning for any such connections between data nodes, management nodes, and SQL or API nodes.
The ndb_top tool is also enhanced to provide spin time information per thread; this is displayed in green in the terminal window.
Important Change: When
lower_case_table_nameswas set to 0, issuing a query in which the lettercase of any foreign key names differed from the case with which they were created led to an unplanned shutdown of the cluster. This was due to the fact that mysqld treats foreign key names as case insensitive, even on case-sensitive file systems, whereas the manner in which the
NDBdictionary stored foreign key names depended on the value of
lower_case_table_names, such that, when this was set to 0, during lookup,
NDBexpected the lettercase of any foreign key names to match that with which they were created. Foreign key names which differed in lettercase could then not be found in the
NDBdictionary, even though it could be found in the MySQL data dictionary, leading to the previously described issue in
This issue did not happen when
lower_case_table_nameswas set to 1 or 2.
The problem is fixed by making foreign key names case insensitive and removing the dependency on
lower_case_table_names. This means that the following two items are now always true:
Foreign key names are now stored using the same lettercase with which they are created, without regard to the value of
Lookups for foreign key names by
NDBare now always case insensitive.
Packaging: Removed an unnecessary dependency on Perl from the
mysql-cluster-community-server-minimalRPM package. (Bug #30677589)
NDBdid not compile successfully on Ubuntu 16.04 with GCC 5.4 due to the use of
std::isnan(). (Bug #30396292)
References: This issue is a regression of: Bug #30338980.
OS X: Removed the variable
SCHEMA_UUID_VALUE_LENGTHwhich was used only once in the
NDBsources, and which caused compilation warnings when building on Mac OSX. The variable has been replaced with
UUID_LENGTH. (Bug #30622139)
NDB Disk Data: Allocation of extents in tablespace data files is now performed in round-robin fashion among all data files used by the tablespace. This should provide more even distribution of data in cases where multiple storage devices are used for Disk Data storage. (Bug #30739018)
NDB Disk Data: Under certain conditions, checkpointing of Disk Data tables could not be completed, leading to an unplanned data node shutdown. (Bug #30728270)
NDB Disk Data: An uninitialized variable led to issues when performing Disk Data DDL operations following a restart of the cluster. (Bug #30592528)
The fix for a previous issue in the MySQL Optimizer adversely affected engine condition pushdown for the
NDBstorage engine. (Bug #303756135)
References: This issue is a regression of: Bug #97552, Bug #30520749.
When restoring signed auto-increment columns, ndb_restore incorrectly handled negative values when determining the maximum value included in the data. (Bug #30928710)
On an SQL node which had been started with
--ndbcluster, before any other nodes in the cluster were started, table creation succeeded while creating the
ndbinfoschema, but creation of views did not, raising
HA_ERR_NO_CONNECTIONinstead. (Bug #30846678)
Formerly (prior to NDB 7.6.4) an
SPJworker instance was activated for each fragment of the root table of the pushed join, but in NDB 7.6 and later, a single worker is activated for each data node and is responsible for all fragments on that data node.
Before this change was made, it was sufficient for each such worker to scan a fragment with parallelism equal to 1 for all
SPJworkers to keep all local data manager threads busy. When the number of workers was reduced as result of the change, the minimum parallelism should have been increased to equal the number of fragments per worker to maintain the degree of parallelism.
This fix ensures that this is now done. (Bug #30639503)
ndb_metadata_syncsystem variable is set to true to trigger synchronization of metadata between the MySQL data dictionary and the
NDBdictionary; when synchronization is complete, the variable is automatically reset to false to indicate that this has been done. One scenario involving the detection of a schema not present in the MySQL data dictionary but in use by the
NDBDictionary sometimes led to
ndb_metadata_syncbeing reset before all tables belonging to this schema were successfully synchronized. (Bug #30627292)
When using shared user and grants, all
ALTER USERstatements were distributed as snapshots, whether they contained plaintext passwords or not.
SHOW CREATE USERdid not include resource limits (such as
MAX_QUERIES_PER_HOUR) that were set to zero, which meant that these were not distributed among SQL nodes. (Bug #30600321)
Two buffers used for logging in
QMGRwere of insufficient size. (Bug #30598737)
References: See also: Bug #30593511.
Removed extraneous debugging output relating to
SPJfrom the node out logs. (Bug #30572315)
When performing an initial restart of an NDB Cluster, each MySQL Server attached to it as an SQL node recognizes the restart, reinstalls the
ndb_schematable from the data dictionary, and then clears all NDB schema definitions created prior to the restart. Because the data dictionary was cleared only after
ndb_schemais reinstalled, installation sometimes failed due to
ndb_schemahaving the same table ID as one of the tables from before the restart was performed. This issue is fixed by ensuring that the data dictionary is cleared before the
ndb_schematable is reinstalled. (Bug #30488610)
NDBsometimes made the assumption that the list of nodes containing index statistics was ordered, but this list is not always ordered in the same way on all nodes. This meant that in some cases
NDBignored a request to update index statistics, which could result in stale data in the index statistics tables. (Bug #30444982)
When the optimizer decides to presort a table into a temporary table, before later tables are joined, the table to be sorted should not be part of a pushed join. Although logic was present in the abstract query plan interface to detect such query plans, that this did not detect correctly all situations using
filesort into temporary table. This is changed to check whether a filesort descriptor has been set up; if so, the table content is sorted into a temporary file as its first step of accessing the table, which greatly simplifies interpretation of the structure of the join. We now also detect when the table to be sorted is a part of a pushed join, which should prevent future regressions in this interface. (Bug #30338585)
When a node ID allocation request failed with NotMaster temporary errors, the node ID allocation was always retried immediately, without regard to the cause of the error. This caused a very high rate of retries, whose effects could be observed as an excessive number of Alloc node id for node
nnnfailed log messages (on the order of 15,000 messages per second). (Bug #30293495)
NDBtables having no explicit primary key,
NdbReceiverBuffercould be allocated with too small a size. This was due to the fact that the attribute bitmap sent to
NDBfrom the data nodes always includes the primary key. The extra space required for hidden primary keys is now taken into consideration in such cases. (Bug #30183466)
When translating an
NDBtable created using
.frmfiles in a previous version of NDB Cluster and storing it as a table object in the MySQL data dictionary, it was possible for the table object to be committed even when a mismatch had been detected between the table indexes in the MySQL data dictionary and those for the same table's representation the
NDBdictionary. This issue did not occur for tables created in NDB 8.0, where it is not necessary to upgrade the table metadata in this fashion.
This problem is fixed by making sure that all such comparisons are actually performed before the table object is committed, regardless of whether the originating table was created with or without the use of
.frmfiles to store its metadata. (Bug #29783638)
An error raised when obtaining cluster metadata caused a memory leak. (Bug #97737, Bug #30575163)