This is a bugfix release, fixing recently discovered bugs in the previous MySQL Cluster NDB 6.1 release.
MySQL Cluster NDB 6.1 no longer in development. MySQL Cluster NDB 6.1 (formerly known as “MySQL Cluster Carrier Grade Edition 6.1.x”) is no longer being developed or maintained; if you are using a MySQL Cluster NDB 6.1 release, you should consider upgrading to MySQL Cluster NDB 6.2 or 6.3.
This Beta release incorporates all bugfixes and changes made in previous MySQL Cluster NDB 6.1 releases, as well as all bugfixes and feature changes which were added in mainline MySQL 5.1 through MySQL 5.1.15 (see Section C.1.37, “Changes in MySQL 5.1.15 (25 January 2007)”).
Please refer to our bug database at http://bugs.mysql.com/ for more details about the individual bugs fixed in this version.
Functionality added or changed:
Cluster Replication: Incompatible Change:
The schema for the ndb_apply_status
table in
the mysql
system database has changed. When
upgrading to this release from a previous MySQL Cluster NDB 6.x
or mainline MySQL 5.1 release, you must drop the
mysql.ndb_apply_status
table, then restart
the server in order for the table to be re-created with the new
schema.
See Section 17.6.4, “MySQL Cluster Replication Schema and Tables”, for additional information.
Bugs fixed:
The cluster waited 30 seconds instead of 30 milliseconds before reading table statistics. (Bug#28093)
Under certain rare circumstances, ndbd could get caught in an infinite loop when one transaction took a read lock and then a second transaction attempted to obtain a write lock on the same tuple in the lock queue. (Bug#28073)
Under some circumstances, a node restart could fail to update the Global Checkpoint Index (GCI). (Bug#28023)
An INSERT
followed by a delete
DELETE
on the same
NDB
table caused a memory leak.
(Bug#27756)
This regression was introduced by Bug#20612.
Under certain rare circumstances performing a
DROP TABLE
or
TRUNCATE TABLE
on an
NDB
table could cause a node
failure or forced cluster shutdown.
(Bug#27581)
Memory usage of a mysqld process grew even while idle. (Bug#27560)
Performing a delete followed by an insert during a local checkpoint could cause a Rowid already allocated error. (Bug#27205)
Cluster Replication: Disk Data: An issue with replication of Disk Data tables could in some cases lead to node failure. (Bug#28161)
Disk Data: Changes to a Disk Data table made as part of a transaction could not be seen by the client performing the changes until the transaction had been committed. (Bug#27757)
Disk Data: When restarting a data node following the creation of a large number of Disk Data objects (approximately 200 such objects), the cluster could not assign a node ID to the restarting node. (Bug#25741)
Disk Data:
Changing a column specification or issuing a
TRUNCATE TABLE
statement on a
Disk Data table caused the table to become an in-memory table.
This fix supersedes an incomplete fix that was made for this issue in MySQL 5.1.15. (Bug#24667, Bug#25296)
Cluster Replication: Some queries that updated multiple tables were not backed up correctly. (Bug#27748)
Cluster Replication: It was possible for API nodes to begin interacting with the cluster subscription manager before they were fully connected to the cluster. (Bug#27728)
Cluster Replication: Under very high loads, checkpoints could be read or written with checkpoint indexes out of order. (Bug#27651)
Cluster API:
An issue with the way in which the
NdbDictionary::Dictionary::listEvents()
method freed resources could sometimes lead to memory
corruption.
(Bug#27663)
The --with-readline
option for
configure did not work for commercial source
packages, but no error message was printed to that effect. Now a
message is printed.
(Bug#25530)
User Comments
Add your own comment.