Age | Commit message (Collapse) | Author |
|
Soledad now creates user-dbs, which has been done by tapicero
in the past. we need to remove any leftovers from tapicero.
|
|
|
|
- create soledad-admin user
- deploy netrc file for userdb creation
- Move soledad-server.conf from /etc/leap to /etc/soledad
- make soledad-server.conf group-accessible for the soledad group, so
the soledad-admin user can read it
- Resolves: #7502
|
|
|
|
|
|
The bigcouch specific class ordering from site_couchdb::create_dbs
needed to move to site_couchdb::bigcouch, otherwise a plain couchdb
setup would try to include bigcouch classes and fail.
Change-Id: I06742d4a12c5b40c9c9faa90441734e6926d422d
|
|
Change-Id: I9e46286c402adc06f3f815f8a1eea11fe82c7c39
|
|
use this to run a single, plain couchdb node, using couchdb 1.6
from the leap repo:
"couch": {
"master": true,
"pwhash_alg": "pbkdf2"
}
Change-Id: Ie4f34c2c5cb9feca7a10450bcf0bc260c8aa9d33
|
|
|
|
|
|
Change-Id: Ifd665eb7efc06c1cf9e7b8fcfd5f42dbb34ac1b1
|
|
|
|
descriptor limits to account for bigcouch sync spikes (#4935)
Change-Id: I242fba31f961b6139ec641e1708b170f5c0d009b
|
|
In a multi-node couch deployment, it was observed that the Service['stunnel']
would be activated, and then later a stunnel::client was created which would
trigger an Exec['refresh_stunnel']. Because of this, and the ordering hints
that were in place, the service would get started, and then the couchdb
databases, users, designs, etc. were being put into place and then a stunnel
client was created, triggering the refresh_stunnel exec, which would cause
an interruption in the connectivity and result in failures.
This change replaces the Service['stunnel'] hint with the the
Exec['refresh_stunnel'] to make sure that the stunnels are fully setup before
attempting couch operations.
Change-Id: I33ddd24884b3c23a1df5555ca53ca65cd703da50
|
|
Change-Id: I318944a6872a53ff9c533704514da339426d9401
|
|
|
|
Site_couchdb::Bigcouch::Settle_cluster/Exec[wait_for_couch_nodes] waits
60s for all nodes to be member of the cluster. Because we deploy to
multiple nodes in parallel, not all nodes are ready at the same time,
so we increased the timeout from 60s to 120s.
|
|
|
|
This only works with the latest patch to puppet_couchdb
|
|
|
|
|
|
out into separate files.
|
|
needed shorewall will be automatically set up. requires new leap_cli
|
|
|
|
|
|
|
|
class to be more visually logical (#5269, #4590, #3712)
Change-Id: I58c28c3bc62e67b25f33da3378e8146110471613
|
|
. make the couchdb service start after the stunnels have been
setup. This may improve the cluster membership coming online
faster
. replace the two Couchdb::Create_db ordering hints (for the
'users' and 'tokens' databases) with a generic
Class['site_config::create_dbs'] hint. This makes it so we get
the ordering hint for all databases, which we were not before,
without having to individually list them
. replace the two Couchdb::Add_user ordering hints (for the
$couchdb_webapp_user and the $couchdb_soledad_user) with a
generic ordering hint for Class['site_couchdb::add_users']
ordering hint. This makes it so we get the ordering hint for all
the users, which we were not before, without having to
individually list them
Change-Id: Ia63e62d68d24e77a49d4ef928a2a8130ab7bccb9
|
|
cluster membership to settle, before attempting any operations
(#5269, #4590, #3712)
Change-Id: Ic9826dda1c242e705ce85ae218766496bdd8ecbd
|
|
|
|
|
|
design docs json files (Feature #5359)
|
|
|
|
|
|
|
|
reviewed by azul
|
|
|
|
local checks
|
|
|
|
|
|
|
|
|
|
This change is a result of https://github.com/leapcode/leap_web/pull/133.
Both should be deployed at the same time to prevent conflicts.
|
|
|
|
|
|
i.e. sessions + tokens (#4794)
|
|
Change-Id: I704a1d46069f975711323ff8e6a634d06c267157
|
|
Change-Id: I5b3d073aec0522cf464ff41905be1ee326f13197
|
|
Change-Id: Ic0e9f5f6a1f28d865d7757a9de0d9399a6a9a5e3
Conflicts:
puppet/modules/site_couchdb/manifests/init.pp
|
|
Change-Id: I77061054f4768f0677ca9c498e6cd6d5df4ff806
|