Documentation Home
MySQL NDB Cluster 6.1 - 7.1 Release Notes
Download these Release Notes
PDF (US Ltr) - 3.0Mb
PDF (A4) - 3.0Mb

MySQL NDB Cluster 6.1 - 7.1 Release Notes  /  Changes in MySQL Cluster NDB 6.1  /  Changes in MySQL Cluster NDB 6.1.10 (5.1.15-ndb-6.1.10) (2007-05-30)

Changes in MySQL Cluster NDB 6.1.10 (5.1.15-ndb-6.1.10) (2007-05-30)

This is a bugfix release, fixing recently discovered bugs in the previous MySQL Cluster NDB 6.1 release.

MySQL Cluster NDB 6.1 no longer in development.  MySQL Cluster NDB 6.1 (formerly known as MySQL Cluster Carrier Grade Edition 6.1.x) is no longer being developed or maintained; if you are using a MySQL Cluster NDB 6.1 release, you should upgrade to the latest version of MySQL Cluster, which is available from .

This Beta release incorporates all bugfixes and changes made in previous MySQL Cluster releases, as well as all bugfixes and feature changes which were added in mainline MySQL 5.1 through MySQL 5.1.15 (see Changes in MySQL 5.1.15 (2007-01-25)).


Please refer to our bug database at for more details about the individual bugs fixed in this version.

Functionality Added or Changed

  • A new times printout was added in the ndbd watchdog thread.

  • Some unneeded printouts in the ndbd out file were removed.

  • The names of some log and other files were changed to avoid issues with the tar command's 99-character file name limit.

Bugs Fixed

  • A regression in the heartbeat monitoring code could lead to node failure under high load. This issue affected MySQL 5.1.19 and MySQL Cluster NDB 6.1.10 only. (Bug #28783)

  • A corrupt schema file could cause a File already open error. (Bug #28770)

  • Setting InitialNoOfOpenFiles equal to MaxNoOfOpenFiles caused an error. This was due to the fact that the actual value of MaxNoOfOpenFiles as used by the cluster was offset by 1 from the value set in config.ini. (Bug #28749)

  • A race condition could result when nonmaster nodes (in addition to the master node) tried to update active status due to a local checkpoint (that is, between NODE_FAILREP and COPY_GCIREQ events). Now only the master updates the active status. (Bug #28717)

  • A fast global checkpoint under high load with high usage of the redo buffer caused data nodes to fail. (Bug #28653)

  • Disk Data: When loading data into a cluster following a version upgrade, the data nodes could forcibly shut down due to page and buffer management failures (that is, ndbrequire failures in PGMAN). (Bug #28525)