Set up a MySQL IaaS NDB cluster with 4 nodes on Azure Linux

 

Introduction

 

In an earlier blog post, I wrote about setting up a MySQL single server instance on Azure Linux (Ubuntu) Virtual Machine. We also did some performance bench-marking around READ-s with lots of joins.

However, most enterprises would be interested in exploring a full-fledged MySQL cluster as opposed to just a single server. MySQL cluster uses the NDB engine to manage the different nodes, and data is saved in the data nodes in such a way that it is extremely fault-resilient, scalable and high performing.

In this blog post, I will explore how to set up a 4 node MySQL IAAS cluster on Azure.

Note: ClearDB now provides PAAS MySQL Clusters on Azure as well. In fact, their offering seems very robust and technically a lot more attractive than setting your own IAAS instance because of the recently announced option of Premium Dedicated Clusters. I said "technically" because I have not had the chance to compare the cost model of going with them versus hosting your own IAAS cluster on Azure. In this blog post, we will focus on the IAAS option.

 

Architecture

 

These steps are listed on mysql website in a generic fashion - https://dev.mysql.com/doc/mysql-cluster-excerpt/5.1/en/mysql-cluster-installation.html

However, these steps need some modifications on Azure Virtual Machines running Ubuntu LTS 14.04. We will try to capture the differences here.

Here is the conceptual architecture diagram of what we are trying to achieve. We will, as is obvious from this diagram, we will spin up 4 Virtual Machines - 2 for the data nodes, 1 for SQL Node, and 1 for the NDB Management node. You can set up a 5th node as a client if you want to run performance tests, and we will also briefly cover that. However, to hold against a large performance test run on the cluster, the cluster nodes need to be beefy, as MySQL has a tendency to consume large amounts of memory unless tuned well - especially queries with JOINS needs AQL. Discussing such optimizations in not within the scope of this blog post.

 

 

 

Set up the cluster on Azure

 

  1. Set up the 1st data node ("mysqlclusterdatanodeA") from Azure portal with its own resource group, vNet, storage account and availability Set. Recommended minimum size SKU is D1 Standard
     

  2. Set up the 2nd data node ("mysqlclusterdatanodeB") - join it to the same resource group, vNET, storage account and availability Set as the data node A. Recommended minimum size SKU is D1 Standard
     

  3. Set up the SQL Node ("mysqlclustersqlnode") - join it to the same resource group, vNET and storage account as the data nodes. No need for Availability Set. Recommended minimum size SKU is A1
     

  4. Set up the Management Node ("mysqlclustermgmtnode") - join it to the same resource group, vNET and storage account as the data nodes. No need for Availability Set. Recommended minimum size SKU is A0
     

  5. From the MySQL cluster download area (https://dev.mysql.com/downloads/cluster/) - download the file mysql-cluster-gpl-7.4.8-linux-glibc2.5-x86_64.tar.gz. Copy it to each of the 4 VM-s created above
     

  6. [At this stage, we have moved on to the section 4.1.1 here: https://dev.mysql.com/doc/mysql-cluster-excerpt/5.1/en/mysql-cluster-install-linux-binary.html. However, instead of saving the downloaded and copied tar.gz file in /var/tmp, we can proceed with it being present in the home directory of the primary used being used to log into the Azure VM-s]
     

  7. Follow all the steps in section 4.1.1. You will need to deviate in the following steps:
     

    1. Step 3 - "cd mysql" should read "/usr/local/mysql"

    2. Step 3 - The step "scripts/mysql_install_db --user=mysql" will fail as it cannot find libaio.so, it turns out that the 64 bit version of the library is not installed on the Azure VM by default, "sudo apt-get install libaio1" fixes it

    3. Step 5 - Instead of /etc/rc.d/init.d use /etc/init.d

    4. Step 5 - instead of the chkconfig command use: sudo update-rc.d mysql.server defaults

  8. Do not restart the SQL node now, just move on to the data node A

  9. Follow steps for data node A, data node B from the same section. DO NOT RESTART ANY OF THESE, move on to the next node

  10. Follow steps for management node from the same section. Do not restart. Move on

  11. [At this stage, we have moved on to the section 4.3 here: https://dev.mysql.com/doc/mysql-cluster-excerpt/5.1/en/mysql-cluster-install-configuration.html\]

  12. Log into the SQL Node and do the following:

    1. sudo vi /etc/my.cnf (or use your favorite editor)

    2. Paste this:

      
       [mysqld] 
       # Options for mysqld process: 
      ndbcluster # run NDB storage engine
       [mysql_cluster] 
       # Options for MySQL Cluster processes: 
      ndb-connectstring=10.2.0.7 # location of management server
      
    3.  [Remember to use the correct Private IP address for the management server in the last line.
      Get that from Azure Portal. Make sure that you DO NOT use the VM's Public IP Address here.
      Use the private one]
      
    4.  sudo reboot (so that the mysql process can start as we put it in /etc/init.d)
      This is also a test whether we did that right!
      
  13. While that is rebooting, repeat the same steps (12 a, b) on the two data nodes (but no need to reboot
    them, as data nodes are not expected to run any daemon that needs restarting or starting)

  14. Log into the management Node and do the following:

    1. sudo mkdir /var/lib/mysql-cluster

    2. cd /var/lib/mysql-cluster

    3. sudo vi config.ini

    4. Paste this:

      
       [ndbd default] 
       # Options affecting ndbd processes on all data nodes: 
      NoOfReplicas=2 # Number of replicas
      DataMemory=1024M # How much memory to allocate for data storage
      IndexMemory=160M # How much memory to allocate for index storage
        # For DataMemory and IndexMemory, we have used the
        # default values. Since the "world" database takes up
        # only about 500KB, this should be more than enough for
        # this example Cluster setup. 
       [tcp default] 
       # TCP/IP options: 
      portnumber=2202 # This the default; however, you can use any
        # port that is free for all the hosts in the cluster
        # Note: It is recommended that you do not specify the port
        # number at all and simply allow the default value to be used
        # instead
       [ndb_mgmd] 
       # Management process options: 
      hostname=10.2.0.7 # Hostname or IP address of MGM node
      datadir=/var/lib/mysql-cluster # Directory for MGM node log files
       [ndbd] 
       # Options for data node "A": 
        # (one [ndbd] section per data node) 
      hostname=10.2.0.4 # Hostname or IP address
      datadir=/usr/local/mysql/data # Directory for this data node's data files
       [ndbd] 
       # Options for data node "B": 
      hostname=10.2.0.5 # Hostname or IP address
      datadir=/usr/local/mysql/data # Directory for this data node's data files
       [mysqld] 
       # SQL node options: 
      hostname=10.2.0.6 # Hostname or IP address
        # (additional mysqld connections can be
        # specified for this node for various
        # purposes such as running ndb_restore) 
      
      
    5.  NOTE: 
      (i) NoOfReplicas = how many data nodes we have
      
      (ii) DataMemory, IndexMemory - Key point to note here apart from the explanations available
      in the documentation is that these memory limits will be imposed on the DATA NODES.
      Not on the management node. In our case, the management node is small (A0) and has
      only 0.75 GB of memory. Obviously it cannot handle 1024M as DataMemory - it does
      not even have that much. The reason we have used 1024M here is that: if you try to install
      the standard publicly available employees database, which is around 160 MB in size,
      you will still be good
      
      (iii) For IndexMemory, the thumb rule is to allocate between 1/6 and 1/8 of the DataMemory,
      but if there are too many UNIQUE in the table definitions, it will need more to store all the
      unique hash indexes
      
      (iv) portnumber - internal communication port for all the hosts in the cluster,
      we will have to open this port up from the portal
      
      (v) hostname - in the 4 sections "ndb_mgmd", "ndbd", "ndbd" and "mysqld" - make sure
      that you use the CORRECT PRIVATE IP ADDRESS of the management node,
      data node A, data node B and SQL node respectively
      
      (v) Remember that the way we have set up the data nodes, there is no directory
      called "/usr/local/mysql/data" on them, but the "datadir" entry for these 2 servers are
      pointing to this value. Which means we have to log into our data hosts, create this
      directory and set its permissions correctly 
      
  15.  So we have a few follow-up tasks from the above config.ini. Let us execute these one by one:

    1.  Go to the portal and browse to the NSG for each of the 4 nodes one by one
      (each host will have one Network Security Group created for itself. Select that NSG,
      add 1 inbound security rule - Source ANY, Destination Port 2202 (overwrite 80 with 2202),
      Protocol ANY and 1 outbound security rule - Source Any, Destination Any (overwrite 80
      with *), Protocol Any. Repeat for each of the 4 NSG-s
      
    2.  Log onto the data nodes are run these commands on each:
      
       sudo mkdir /usr/local/mysql
      sudo mkdir /usr/local/mysql/data
      sudo chmod a+w /usr/local/mysql
      sudo chmod a+w /usr/local/mysql/data/
      
  16. Now is a good time to log into your SQL Node and make sure that it rebooted. Once in, run
    "sudo netstat -tap | grep mysql" and make sure that mysqld is running
     

  17. [At this stage, we have moved on to the section 4.4 here: https://dev.mysql.com/doc/mysql-cluster-excerpt/5.1/en/mysql-cluster-install-first-start.html\]

  18. Initial Startup of cluster:

    1. On management node: Step 1 under section 4.4 will fail unless you run this first: "sudo mkdir /usr/local/mysql" and then run ndb_mgmd with sudo: "ndb_mgmd -f /var/lib/mysql-cluster/config.ini". Verify with "ps aux | grep ndb"

    2. On Data nodes: sudo ndbd. Verify with "ps aux | grep ndbd"

    3. On SQL Node: mysqld should be running already as we set it to run at startup and we rebooted the node. If you have not done that, do it now

    4. On management node: (VERIFY THAT CLUSTER IS RUNNING): "ndb_mgm" (no need to use sudo) - use verb SHOW and then EXIT

  19. [At this stage, we have moved on to the section 4.5 here: https://dev.mysql.com/doc/mysql-cluster-excerpt/5.1/en/mysql-cluster-install-example-data.html\]

  20. By now, you can safely exit from any terminal(s) on the data and management nodes. All interaction with the cluster can happen through the SQL node, as if you are talking to a MySQL server located on the SQL Node. Remember, however, that the underlying engine must be specified as "NDBCLUSTER" instead of InnoDB or MyISAM for every table residing on every database in the cluster for it to be replicated across the cluster. Therefore, when creating any table on the cluster, we must specify ENGINE=NDBCLUSTER option. This goes for any mode of creation - even if we are importing databases from dumps. If we have not done so, we must ALTER TABLE each table and set ENGINE=NDBCLUSTER later on

  21. On SQL Node:
     

    1. cd /usr/local/mysql

    2. sudo bin/mysql_secure_installation (As root password is not set yet, just press ENTER when prompted for current root password. Then, set the root password properly)

      NOTE: It is worth taking into account that Cluster nodes do not make use of the MySQL privilege system when accessing one another. Setting or changing MySQL user accounts (including the root account) affects only applications that access the SQL node, not interaction between nodes

    3.  log into mysql client prompt using root user (assuming you are still inside /usr/local/mysql): bin/mysql -uroot -p<root password set above>

  22. Add /usr/local/mysql/bin to PATH so that you can just log into the SQL Node VM and start typing "mysql..." without bothering to change directories.

    For an Azure Linux (Ubuntu) VM, this can be achieved most easily by executing this:

     cd ~; ln -s /usr/local/mysql/bin/ ./bin
    
    What this will do is create a symbolic link called "bin" in your home directory that links to
     /usr/local/mysql/bin/ . Azure Ubuntu VM-s already come with a ~/.profile file that adds
    your home directory's bin subdirectory to PATH while logging in. Log out and log back in
    and you should have access to mysql at your command prompt from anywhere.
    
  23.  The cluster is ready. Treat it like a single server where the network address
    of the SQL node is the equivalent of the "server" in the single-server world.
    You can now point another VM (a client box) to this SQL node and connect,
    and the steps would be the same as if you were connecting to a single server,
    which are covered in my previous blog post<br> 
    

 

Conclusion

 

This blog post only handles the setting up of the cluster. I am actually going ahead and running some basic performance tests on this cluster - I will update best practice guidelines for Azure shortly.

 Thanks!