Recently we have worked with a customer on replication latency issue with transactional replication. Customer has over 30 published databases on a single server. All of them are very active. Periodically, they will see up to 30 minutes latency from publisher to distributor. When they see waits on REPL_SCHEMA_ACCESS. Below is a sample screenshot of sys.dm_exec_requests during problem period.


What does this wait type mean?

Our current online document states "Occurs during synchronization of replication schema version information. This state exists when DDL statements are executed on the replicated object, and when the log reader builds or consumes versioned schema based on DDL occurrence." But this wait type is also used by synchronize memory access to prevent multiple log reader to corrupt internal structures on the publishers. Each time log reader agent runs sp_replcmds, it needs to access memory buffer. If it results in growing the buffer, the action needs to be synchronized among log reader agents with REPL_SCHEMA_ACCESS.

Contention can be seen on this wait type if you have many published databases on a single publisher with transactional replication and the published databases are very active.

Troubleshooting and reducing contention on REPL_SCHEMA_ACCESS waits?

This issue is dependent on number of log reader agents accessing the same publisher and transaction rate. If you have a single log reader agent access the publisher, you shouldn't see this type of contention.

In general, you can watch transactions/sec performance counter for all published databases to measure how active your system is. The higher your transaction rate is, the more likely you hit the issue assuming you have multiple log reader agents accessing the same publisher.

We charted the waits on REPL_SCHEMA_ACCESS and transaction/sec for multiple published databases. We saw a very clear correlation.

SQL Nexus report


Transaction/sec for one example published database



Here are a few things you can do to reduce contention:

  1. Do not use large transactions. Large transaction that results in many commands can make the situation worse because of higher requirement of memory buffer. If you do have large transactions, experiment MaxCmdsIn value.
  2. Try to spread out transactions among different published database to different time. for example, if you have batch load jobs for different databases, don't schedule them at the same time
  3. Reduce number of log readers. In this customer, they have over 35 published databases on the same publisher and they are all active. The wait is at server level. So if you split your published databases into two different instances (even on the same hardware), contention can be reduced
  4. Experiment decreasing –PollingInterval for your log reader agent. The default is 5 seconds. If you reduce the PollingInterval. This will allow log reader agent to catch up more frequently once the wait gets cleared.



Jack Li

Senior Escalation Engineer | Microsoft SQL Server Support

Comments (4)

  1. JohnW says:

    Jack, the hyperlnk to the SQL Nexus screenshot points to an internal server (mce_host) and doesn't show up outside MS.

  2. Aditya says:

    Hi Jack,

    thanks for explaining this wait type in clear , we were suffering with this wait types in past bnut really couldnt figure out what is happening.

    I have one more doubt here., what if i have huge trasanctions/second on the database which is published and i have CDC enabled on the same database, what is the probability of we hitting this wait type, basically is there a correlation between this wait type and CDC that can cause Log reader latency.

  3. Ashish says:

    Is it possible to allocate the higher memory upfront on the logreader start up itself to reduce this problem?

  4. Try running Log Reader or CDC Scan Job at schedule intervals. One it may reduce number of Log Reader Agents running at any one time, and second, should Agent fail, the Log Reader Agents will restart at next scheduled interval.

Skip to main content