SQL Server Errors or Failures from Error: 20646 to Error: 20665

Error: 20646, Severity: 16, Peer-to-peer publications do not support %s. Change the value for parameter ‘%s’.
Error: 20647, Severity: 16, Cannot modify property ‘%s’. The publication is used in a peer-to-peer topology, which does not allow this property to be modified after the publication is created.
Error: 20648, Severity: 16, An article already exists for table “%s” with a different value for the @delete_tracking property. The value must be the same for all publications in which the table is published. Use the stored procedures sp_helpmergearticle and sp_changemergearticle to

Error: 20649, Severity: 16, Publications enabled for heterogeneous subscriptions do not support %s. Please change the ‘%s’ parameter value.
Error: 20650, Severity: 16, Cannot enable data definition language (DDL) replication (a value of “true” for the @replicate_ddl parameter) for publication “%s”. This is because the compatibility level of the publication is lower than 90RTM. For new publications, in the stored procedu
Error: 20651, Severity: 16, Publication “%s” “%s”. Therefore the compatibility level of the publication cannot be set to lower than %d. To set the compatibility level lower, disable the feature and then call the stored procedure sp_changemergepublication to lower the

compatibility l
Error: 20652, Severity: 16, Required metadata for publication ‘%s’ could not be found in the sysmergeschemachange system table. Run the Snapshot Agent again.
Error: 20653, Severity: 16, Cannot have a dynamic snapshot job with both dynamic_filter_login and dynamic_filter_hostname being NULL.
Error: 20654, Severity: 16, Dynamic snapshots are only valid for merge publications.
Error: 20655, Severity: 16, The partitioned snapshot process cannot complete. Cannot retrieve the maximum timestamp information from the MSsnapshot_history table in the distribution database. Ensure that a standard snapshot is up-to-date and available.
Error: 20656, Severity: 16, The @subset_filterclause parameter cannot reference a computed column.
Error: 20657, Severity: 16, The value for the @pub_identity_range parameter must be a multiple of the increment for the identity column. The increment for table “%s” and identity column “%s” is %s.
Error: 20658, Severity: 16, The value for the @identity_range parameter must be a multiple of the increment for the identity column. The increment for table “%s” and identity column “%s” is %s.
Error: 20659, Severity: 11, The value of IDENT_CURRENT (“%s”) is greater than the value in the max_used column of the MSmerge_identity_range system table.
Error: 20660, Severity: 16, The republisher’s republishing range obtained from its publisher is not large enough to allocate the specified @pub_identity_range.
Error: 20661, Severity: 16, The republisher’s republishing range obtained from its publisher is not large enough to allocate the specified @identity_range.
Error: 20662, Severity: 16, The republisher does not have a range of identity values from the root Publisher ‘%s’ that it can assign to its Subscribers. Ensure that the republisher has a server subscription to the publication at the root Publisher, and then run the Merge Agent to sy
Error: 20663, Severity: 16, The identity range allocation entry for the Publisher could not be found in the system table MSmerge_identity_range. Ensure that the value for the @identityrangemanagementoption property is “auto”.
Error: 20664, Severity: 16, The Publisher cannot be assigned a new range of identity values, because the values for the identity column’s data type have all been used. Change the data type in the identity column.
Error: 20665, Severity: 16, The republisher does not have a range of identity values from the root Publisher that it can assign to its Subscribers. Run the Merge Agent to synchronize with the root Publisher.

With codeplex shutting down, we have moved SQL Nexus to github with a new release (6.0). Now both Pssdiag/SQLDiag manager and SQL Nexus are on github. Where to get it As you navigate to SQL Nexus, you can download code and released binary files. If you choose to download binary files, you can go to...

SQL Server : large RAM and DB Checkpointing Hi everyone, This post’s purpose is to establish a summary of the specific behaviors with relation to DB Checkpoint that may happen within SQL Server when running with a large quantity of allocated memory and when applicable, how to best address them. SQL Server 2016 improves...

Recently we got an inquiry from a customer who received the following message in errorlog and wanted to know why. [INFO] HkDatabaseTryAcquireUserMemory(): Database ID: [7]. Out of user memory quota: requested = 131200; available = 74641; quota = 34359738368; operation = 1. This is my first time to see this error. As usual, I relied...

Recently I assisted on a customer issue where customer wasn’t able to alter a memory optimized table with the following error Msg 41317, Level 16, State 3, Procedure ddl, Line 4 [Batch Start Line 35]A user transaction that accesses memory optimized tables or natively compiled modules cannot access more than one user database or databases...

In a previous blog, I talked about memory optimized table consumes memory until end of the batch. In this blog, I want to make you aware of cardinality estimate of memory optimized table as we have had customers who called in for clarifications. By default memory optimized table variable behaves the same way as...

I worked on an interesting issue today where a user couldn’t restore a backup. Here is what this customer did: backed up a database from an on-premises server (2008 R2) copied the file to an Azure VM tried to restore the backup on the Azure VM (2008 R2 with exact same build#) But he got...

Recently, I looked an In-Memory OLTP issue with Principal Software Engineer Bob Dorr who is still my office neighbor. After restoring a database that had just one memory optimized table, we dropped the table. Even without any memory optimized tables,number of checkpoint files keep going up every time we issue a checkpoint. For a while,...

In this blog Added per-operator level performance stats for Query Processing, Senior PM in QP talks about extending operator level performance stats. They include stats related to reads, CPU and elapse time. These are very helpful to track down query performance issues. We worked on recent case where we put ActualElapsedms in a good...

In blog “Importance of choosing correct bucket count of hash indexes on a memory optimized table”, I talked about encountering performance issues with incorrect sized bucket count. I was actually investigating an out of memory issues with the following error. Msg 701, Level 17, State 103, Line 11There is insufficient system memory in resource pool...

I was working with a customer to troubleshoot memory optimized table issues. In this scenario, our customer uses a memory optimized table variable. He put 1 million rows of data into the table variable and then process it. Based on what he said, I tried to come up with a repro to see if...