/* Set to true if you are behind CloudFlare (not recommended) or behind http-reverse
proxy to enable IP detection from X-Forwarded-For header.
Advanced users only. It’s tricky to make it right and secure.
*/
“behindReverseProxy”: false,

// Try to get new job from geth in this interval
“blockRefreshInterval”: “120ms”,
“stateUpdateInterval”: “3s”,
// Require this share difficulty from miners
“difficulty”: 2000000000,

/* Reply error to miner instead of job if redis is unavailable.
Should save electricity to miners if pool is sick and they didn’t set up failovers.
*/
“healthCheck”: true,
// Mark pool sick after this number of redis failures.
“maxFails”: 100,
// TTL for workers stats, usually should be equal to large hashrate window from API section
“hashrateExpiration”: “3h”,

/* If you are running API node on a different server where this module
is reading data from redis writeable slave, you must run an api instance with this option enabled in order to purge hashrate stats from main redis node.
Only redis writeable slave will work properly if you are distributing using redis slaves.
Very advanced. Usually all modules should share same redis instance.
*/
“purgeOnly”: false
},

// Check health of each geth node in this interval
“upstreamCheckInterval”: “5s”,

/* List of geth nodes to poll for new jobs. Pool will try to get work from
first alive one and check in background for failed to back up.
Current block template of the pool is always cached in RAM indeed.
*/
“upstream”: [
{
“name”: “main”,
“url”: “http://127.0.0.1:8545”,
“timeout”: “10s”
},
{
“name”: “backup”,
“url”: “http://127.0.0.2:8545”,
“timeout”: “10s”
}
],