pecl:mysqlnd_ms

This is an old revision of the document!


Mysqlnd replication and load balancing community wiki page

This is a community wiki page on the mysqlnd replication and load balancing plugin (short: mysqlnd_ms).

Overview

Description

PECL/mysqlnd_ms adds MySQL replication and load balancing support to all PHP MySQL extensions (mysql, mysqli, PDO_MySQL), if compiled to use the mysqlnd library (PHP 5.3+). The plugin does read-write splitting to send write statements to a MySQL master and read statements to a MySQL slave. A built-in load balancer distributes the statements over the the configured MySQL servers. Multiple load balancing policies are available.

Application -> PHP MySQL extension (mysql, mysqli, PDO_MySQL) -> mysqlnd library -> mysqlnd_ms (read-write split, load balancing) -> MySQL

PECL/mysqlnd_ms extends the mysqlnd library. The PHP MySQL extensions (mysql, mysqli, PDO_MySQL) can be compiled to use the mysqlnd library. The mysqlnd library is part of the PHP source tree as of version 5.3 (ext/mysqlnd). Mysqlnd is a compile time default for all PHP MySQL extensions as of PHP 5.4. Because the PHP MySQL extensions, the mysqlnd library and the mysqlnd_ms library plugin are written in C, they operate transparent from an applications point of view.

Please, find further details in the reference manual, http://php.net/mysqlnd_ms .

Advantages:

  • transparent from an applications point of view, depending on application use
  • no application changes required for basic use cases
  • easy to add to existing set ups
  • client-side plugin
    • horizontal scale-out
    • no single point of failure
Contributors

... blame: Andrey Hristov, Ulf Wendel, Johannes Schlüter

Download

The plugin is available from PECL, http://pecl.php.net/package/mysqlnd_ms . The plugin is available under the terms of the PHP License.

Repository
History
  • 04/2011 - 1.0.1 alpha, first release of practical relevance
  • 09/2010 - Initial check in, proof-of-concept

Further resources (manual, blogs, presentations, ...)

The manual is available from http://php.net/mysqlnd_ms .

Blog postings
Presentations

Development steps (release planning)

The following should be read as a proposal.

1.0 - Cover replication basics to test user feeedback

Coding/Features
  • [done] implement read/write splitter
    • logic: everything which does not begin with “SELECT” (or hint) is a write
  • [done] implement load balancer
    • policies: random, round-robin, random-once
  • [done] implement SQL hint parser
    • support hint to force use of any slave
    • support hint to force use of the master
    • support hint to force use of last used server for transaction support
QA
  • [done] >= 85% code coverage through tests
  • [done] add micro benchmark framework to test directory
  • [done] check if plugin and SQL hints cause <5% slowdown in worst case micro benchmarks (SELECT 1 FROM DUAL [WHERE 1=0])
Other
  • [done] publish through PECL
Documentation, Blogs, Articles etc.
  • [done] write and publish initial documentation covering all features
  • [done]] get documentation reviewed
Open tasks from previous releases

1.1 - Cover replication basics with production quality

Refactored, stable 1.0 successor.

Coding/Features
  • [done] develop new JSON-based plugin config syntax
QA
  • [done] significant higher number of tests (achieved: almost 5x more compared to 1.0 alpha)
  • [done] 75-90% code coverage through tests
  • [done] automated cross-platform building and testing
Documentation, Blogs, Articles etc.
  • [done] High-level overview presentation
  • [open] Documentation update

1.2 - Global Transaction ID injection and quality-of-service concept

Coding/Features
  • [done] Introduce quality-of-service (QoS) filter. Service levels provided by QoS filter:
    • [done] eventual consistency, optional option slave lag
    • [done] session consistency, optional option GTID
    • [done] strong consistency
  • [done] Allow setting connection quality at run time
  • [done] Client-side global transaction ID injection (GTID)
  • [done] Support fetching of most recent GTID
QA
  • [done] 85-90% code coverage through tests
Documentation, Blogs, Articles etc.
  • [done] Documentation update
  • [done] Load balancing for PHP MySQL presentation
  • [done] Assorted blog postings on GTID and QoS

1.3 - Query caching through quality-of-service concept

Coding/Features
  • [done] Introduce QoS filter for caching
  • [done] Update GTID related QoS logic for MySQL 5.6
QA
  • [done] 85-90% code coverage through tests
Documentation, Blogs, Articles etc.
  • [done] Documentation update
  • [done] Presentations on GTID and MySQL 5.6
  • [done] Assorted blog postings

1.4 - Tweaking based on user feedback

Coding/Features
  • [open] More fail over options
    • [open] Silent and automatic connection fail over if server returns configured error code
      • NOTE: this may require refactoring of four filters.
    • [open] Automatic on-connect fail over, if activated, shall be done in a loop until a connection can be opened. Currently we stop after the first attempt. If automatic fail over is on, we try “next” and in case of a failure of “next” we stop. With the new logic there shall be no stop until all possible “next” have been tried.
  • [open] Support directing statements manually to a group of nodes for more efficient server cache usage
  • [open] Refine QoS session consistency server selection policy
    • [open] Support “wait for GTID”. Currently we loop over all servers until we find a matching one. MySQL 5.6 allows SQL users either to fetch the latest GTID or SQL users can ask for a GTID and their request will block until the GTID has been replicated on the server. We should support the latter logic as well.
    • [open] Remember the most current server and test this one first when searching for a GTID (a synchronous server). Use of cached information is possible for the duration of a read-only request sequence. The cache must be flushed and refreshed for every write.
  • [open] Improve load balancing
    • [open] Allow users to set a ratio/weight for query distribution. There may be machines in a cluster/config which can handle twice as many requests as others. Such machines shall be given a weight of, for example, two to indicate that the load balancer shall direct twice as many requests to them as to a machine with the weight of one.
  • [open] Work with php.net Windows team to provide PECL/mysqlnd_ms windows builds
QA
  • [open] 85-90% code coverage through tests
  • [open] Cross-platform testing on all major MySQL build hosts
Documentation, Blogs, Articles etc.
  • [open] Update documentation
  • [open] Assorted blog postings

Open and delayed tasks

Coding/Features
  • [incomplete - 1.2] make load balancer schemata aware
    • for each statement detect database/table
    • follow schemata based partitioning rules
    • cancelled/delayed: SQL parser is not stable and can crash plugin
  • [cancelled] improve trx_stickiness using MYSQL*/MYSQLND* in_transaction flag
  • [open] bail on unknown config settings. For example, bail if someone uses round_robin instead of roundrobin
QA
  • [open] >= 90% code coverage through tests
  • [open] write report on parsing overhead, goal: <5% in worst case micro benchmarks
Documentation, Blogs, Articles etc.
  • [done] impact of connection limits such as allow_persistent, max_persistent, max_links

Raw Bin ideas (RFCs)

If you got any feature proposal or, if you have any replication related client-side task which you want to see solved, please, describe and add it here. Use the existing entries as a pattern.

Idea

Problem to solve / Idea
Feature description

Eventual consistent: automatic chaining of PECL/mysqlnd_qc query cache

Problem to solve / Idea

MySQL replication is asynchronous. Slaves are eventual consistent because there is a time lag between an update on the master and the ability of all clients to read that update from the slaves. Slaves may serve stale data.

MySQL replication used for read scale-out requires applications to be able to cope with stale data. If stale data is acceptable, a replication system may replace a MySQL slave read access with an access to a local, eventually stale, cache. The cache access will lower network latency resulting in faster reply to the query. Furthermore, it will reduce the load on the MySQL replication slaves, thus requiring less machines for a given read workload.

For certain queries, PECL/mysqlnd_ms shall automatically populate a local cache and fetch results from the cache. PECL/mysqlnd_ms shall use PECL/mysqlnd_qc, if possible, in a transparent way.

(MySQL semi-synchronous replication does not change the matter as it requires only one slave to have received but not yet applied the update.)

PECL/mysqlnd_qc supports Memcache, APC and other storage media.

Feature description

Applications shall be able to define a “quality of service” (the term is borrowed from Tungsten) for a statement, a connection or globally. The “quality of service” shall be guaranteed by PECL/mysqlnd_ms. The service level defines if eventual consistency is allowed:

  • [RW_]CONSISTENT
  • [RO_]EVENTUAL_CONSISTENT

With CONSISTENT every statement shall be run on the master to avoid stale data. EVENTUAL_CONSISTENT allows the use of slaves and thus, stale, non-consistent data may be read.

PECL/mysqlnd_ms supports the logic of service levels already as of 1.0.0, but does not call them that way. PECL/mysqlnd_ms features SQL hints (USE_MASTER) to enforce the use of the master (CONSISTENT), the use of the last used server (USE_LAST_USED) for consistent reads from slaves (CONSISTENT) and has an ini-setting (master_on_write) to automatically use only the master after the first write (CONSISTENT). It also features the concept of EVENTUAL_CONSISTENT which is the default operation mode.

However, PECL/mysqlnd_ms does not call it “quality of service” and has not extra API calls for setting the service level. It may be convenient to introduce an extra API call, which removes the need of having to prefix statements with SQL hint. Prefixing is inconvenient.

/* Force use of master - consistent read although no SQL hint is used */
mysqlnd_ms_set_service_level($link, CONSISTENT);
$link->query("SELECT ...");
$link->query("SELECT ...");

/* Use master or slaves for reads - allow read stale, non-consistent data is acceptable */  
mysqlnd_ms_set_service_level($link, EVENTUAL_CONSISTENT);
$link->query("SELECT ...");

The service level EVENTUAL_CONSISTENT can be combined with automatic, local caching. It shall be configurable if and for which queries caching shall be used. Configuration can, for example, be done via the PECL/mysqlnd_ms plugin configuration file. It may be also useful to introduce a EVENTUAL_CONSISTENT_CACHE service level to control caching via API call.

If PECL/mysqlnd_ms runs in caching mode, it shall automatically populate and read from the PECL/mysqlnd_qc query cache. PECL/mysqlnd_qc is TTL based. The TTL could either be set manually via PECL/mysqlnd_ms function call, globally in the PECL/mysqlnd_ms configuration file or automatically derived from SHOW SLAVE STATUS Seconds_Behind_Master.

Use and integration of PECL/mysqlnd_qc shall be as transparent as possible from an applications point of view. For example, PECL/mysqlnd_ms shall inject all SQL hints required to control caching by PECL/mysqlnd_qc.

Global Transaction ID

Problem to solve / Idea

In the scope of MySQL master - slave replication global transaction IDs can be understood as a measure to identify a database state. Replicas/slaves can use global transaction ids to identify change sets (from the master) that need to be applied. Clients can use global transaction ids to monitor the replication status of the replicas in the database cluster.

The MySQL built-in replication does not feature global transaction IDs. MySQL replication uses transaction log files and log file positions to identify change sets that replicas (slaves) need to apply. Log file and log file position are local to a machine. A tuple of (log file, log file position) does not uniquely identify a change set in the whole cluster.

If, for example, your setup is A -> B -> C (A is the master, B is a slave reading from A, C is a slave reading from B) and B fails you are faced with the challenge to identify the log file and log file position on A from which C shall continue to read and apply change sets. C cannot continue reading from the same (log file, log position) as it was reading from B. The tuple (log file=2, log position=17) on B can identify a different change set than the tuple (log file=2, log position=17) on A. C might get out of sync, loose change sets or apply them multiple times, if it would blindly continue reading from the same position on A.

Today, it can become a bit of a time consuming guessing game to identify the next change set that B needs to process in this fail-over example.

Fail-over and circular-replication is much easier to do if the database cluster maintains a global transaction ID.

The plugin shall inject SQL to manually maintain a global transaction ID on the master.

Feature description

Global transaction IDs can be used for multiple purposes:

  • Simplify server fail-over (server-side, not client-side)
  • Monitor slave replication status (client-side, plugin relevant)
    • Identify “up-to-date” slave for “synchronous” reads
    • Identify slaves which are “recent enough” (quality of service)
    • Let the load balancer skip slaves which have fallen to far behind to reduce their load

The plugin shall support using a global transaction ID for:

  1. server fail-over (e.g. promoting new master after master failure)
  2. selection of “recent enough” (quality of service) slaves for reading

Gobal transaction ID generation shall be done by SQL injection. The plugin shall insert SQL at the end of each (write) transaction (see http://jan.kneschke.de/projects/mysql/mysql-proxy-and-a-global-transaction-id/) on the master. The inserted SQL shall generate an ID which helps with the desired use cases.

The SQL used for globals transaction ID generation can be different for the different purposes. It may be good enough to inject SET @_g_trx_id=UUID_SHORT() on the master before a SQL transaction commits to help with fail-over. Whereas the selection of slaves of a certain state (“up-to-date, “recent enough”) may require maintaining a sequence number or a tuple of server-id and sequence number. Cluster topology and replication method (statement based, row based, mixed).

Thus, the user shall be able to configure:

  • SQL used for ID generation
  • SQL used to retrieve ID

If no SQL is given to retrieve the ID, the plugin does not feature '2) selection of “recent enough” (quality of service) slaves for reading'. If given, the user can restrict the choice of slaves for reading by setting a minimum global transaction id which has to be already replicated on the slaves.

This is useful if, for example, an application does a write and needs to read the modified data “immediately” after the write. Not all slaves my have replicated the write “immediately”. Thus, the read can only be performed on the master or some, selected slaves which have replicated the changes already. Using a global transaction id it is possible to identify those slaves.

/* write to master */
$link->query("INSERT ...");
$write_trx_id = mysqlnd_ms_get_last_global_transaction_id();

/* Force use of master - consistent read although no SQL hint is used */
mysqlnd_ms_set_service_level($link, CONSISTENT);
$link->query("SELECT ...");
$link->query("SELECT ...");
 
/* Read from a slave which has replicated global transaction ID or, if no slave found, use master */
mysqlnd_ms_set_service_level($link, MIN_GLOBAL_TRX_ID, $write_trx_id);
$link->query("SELECT ...");
$link->query("SELECT ...");

/* Use master or slaves for reads - allow read stale, non-consistent data is acceptable */  
mysqlnd_ms_set_service_level($link, EVENTUAL_CONSISTENT);
$link->query("SELECT ...");

The global transaction ID feature of the plugin is turned off by default. The feature is seen as an experts topic.

mysqlnd_ms_change_user()

Problem to solve / Idea

Give users a way to overrule all automatic decisions, including the use of host credentials from the config when broadcasting COM_CHANGE_USER after mysqlnd_change_user(master_user, master_pass, master_db).

Feature description
bool mysqlnd_ms_change_user(mixed $connection, mixed $where, $user, $password, $db)

$where: MYSQLND_MS_ALL_CONNECTIONS, MYSQLND_MS_LAST_USED_CONNECTION

Consider:

$where: MYSQLND_MS_ALL_CONNECTIONS, MYSQLND_MS_LAST_USED_CONNECTION, name of server

mysqlnd_ms_get_errors()

Problem to solve / Idea

Not all MS errors can be bubbled up through the standard APIs. For example, failing to parse parts of the MS config may result in connection error. Upon connect, the user will get the error code for the failed connect. However, the failure to parse the config, the cause of the problem, is only shown as a warning.

Feature description
array mysqlnd_ms_get_errors([mixed $connection])

TDB.

More fail over configuration options

Problem to solve / Idea

Whenever the plugin tries to connect to a node it may fail to do so. Connection attempts can be made when opening a connection or later, when executing a statement (lazy connections). By default the plugin bails out if the connection attempt fails. To make using the plugin as transparent as possible it is possible to optionally enable automatic fail over in case of failed connection attempts.

Feature description

Up to version 1.3 the automatic fail over stops after trying one alternative. For example, if a connection to A fails we try B. If connecting to B fails we stop and bail out. In that case the user must handle the error although there may be node C and D, ... which could be used.

In the future the search for an alternative shall not stop after B but continue until a connection has been established or there are no alternatives. If a connect to A fails, the plugin shall try B, C, D and so forth. Automatic fail over, automatic connection attempts shall not stop after trying B and failing to connect to B.

Whether this shall become a new default or become configurable is to be decided.

Furthermore, it shall be possible to link the automatic fail over to certain error codes: if A returns error n, try next. If A returns m, stop the search and bail out. This way we can differentiate between a server that rejects new requests, e.g. because its overloaded, and a server that has failed.

Support directing statements manually to a group of nodes for more efficient server cache usage

Problem to solve / Idea

In large clusters users can improve performance by optimizing query distribution and server selection using criteria such as cache usage, distance or latency. Application developers shall be allowed to annotate a statement in a way that its executed on a certain group of nodes.

Server cache usage can be optimized by distributing queries in a way that they hit hot caches. For example, clients may want to run all accesses to table A, B and C on the node group 1 and table accesses to D preferably on group 2. Because we do not support automatic table filtering/partitioning clients shall be allowed to manually route requests to the appropriate groups.

Some nodes may be co-located to clients whereas others may be remote. This is often done in large systems when storing multiple copies, e.g. on the machine, on the rack, in the same data center, in a different data center. In case of MySQL Replication its unlikely to find such highly optimized setups, however, there may be nodes closer to the client than others. Nodes closer to a client may be given a certain alias or group name and application developers shall be allowed to hint routing to a group of such nodes.

Feature description

For every node in the configuration users shall be able to set one or more group names. A SQL hint, for example, MS_GROUP=name can be used to hint the load balancer to direct a request to a certain node group.

Refine QoS session consistency server selection policy

Problem to solve / Idea

Users can request session consistency. Session consistency guarantees that a user will only be redirected to nodes that have already replicated his changes. Currently we check the status of all configured nodes before we pick a node for statement execution. Checking the status causes extra load on the nodes. The load shall be optionally reduced with tweaking settings.

Feature description
pecl/mysqlnd_ms.1334663291.txt.gz · Last modified: 2017/09/22 13:28 (external edit)