[db-wg] NRTM replication inefficiencies
- Previous message (by thread): [db-wg] NRTM replication inefficiencies
- Next message (by thread): [db-wg] NRTM replication inefficiencies
Messages sorted by: [ date ] [ thread ] [ subject ] [ author ]
Edward Shryane
eshryane at ripe.net
Fri Dec 8 15:05:10 CET 2017
Hi Agoston, > On 8 Dec 2017, at 13:43, Horváth Ágoston János via db-wg <db-wg at ripe.net> wrote: > > Or you could use TCP's built-in keepalive feature: > > http://tldp.org/HOWTO/TCP-Keepalive-HOWTO/overview.html > Yes, this is already possible, you can set the SO_KEEPALIVE option on the socket. However, at least on CentOS7 (Linux), the default is to wait 2 hours before sending a keepalive probe, then 9 probes have to fail (each 75s apart) before declaring the connection is broken. Changing this default behaviour is applied system-wide. Adding a protocol-specific keepalive mechanism may still be useful. Regards Ed
- Previous message (by thread): [db-wg] NRTM replication inefficiencies
- Next message (by thread): [db-wg] NRTM replication inefficiencies
Messages sorted by: [ date ] [ thread ] [ subject ] [ author ]
[ db-wg Archives ]