Why yet another replication system?

Slony-I was born from an idea to create a replication system that was not tied
to a specific version of PostgreSQL, and allowed to be started and stopped on
an existing database with out the need for a dump/reload cycle.

What Slony-I is:

Slony-I is a "master to multiple slaves" replication system with cascading and
slave promotion.  The big picture for the development of Slony-I is a
master-slave system that includes all features and capabilities needed to
replicate large databases to a reasonably limited number of slave systems.
Slony-I is a system for data centers and backup sites, where the normal mode
of operation is that all nodes are available. 

What Slony-I is not:

Slony-I is not a network management system.  Slony-I does not have any
functionality within it to detect a node failure, or automatically promote a
node to a master or other data origin.  Slony-I is not multi-master; it's not
a connection broker, and it doesn't make you coffee and toast in the morning.

Why doesn't Slony-I do automatic fail-over/promotion?

This is the job of network monitoring software, not Slony-I.  Every site's
configuration and fail-over path is different.  For example, keep-alive
monitoring with redundant NIC's and intelligent HA switches that guarantee
race-condition-free takeover of a network address and disconnecting the
"failed" node vary in every network setup, vendor choice, hardware/software
combination.  This is clearly the realm of network management software and not
Slony-I.  Let Slony-I do what it does best: provide database replication.

Current Limitations:

Slony-I does not automatically propagate schema changes, nor does it have any
ability to replicate large objects. Slony currently works with PostgreSQL 
versions 7.3 and newer.
