Quarry is a web service that allows to perform SQL queries against Wikipedia and sister projects databases.
Quarry uses Docker to set up a local environment. You can set it up by:
- Download and install Docker and docker-compose (already ships with docker on Windows and Mac)
- Clone the Quarry repository
- Run
docker-compose up
A web server will be setup, available at http://localhost:5000. Change to python files will trigger an automatic reload of the server, and your modifications will imediatelly be taken into account. A worker node is also created to execute your queries in the background (uses the same image). Finally, redis and two database instances are also started.
To stop, run docker-compose stop
or hit CTRL-C on the terminal your docker-compose
is running in. After that, to start with code changes, you'll want to docker-compose down
to clean up. Also, this creates a docker volume where sqlite versions of query
results are found. That will not be cleaned up unless you run docker-compose down -v
It is possible to run a quarry system inside minikube! At this time, you need to set it up with a cluster version before 1.22, most likely.
First build the containers:
eval $(minikube docker-env)
docker build . -t quarry:01
cd docker-replica/
docker build . -t mywiki:01
You will need to install minikube (tested on minikube 1.23) and helm and kubectl on your system. When you are confident those are working, start minikube with:
minikube start --kubernetes-version=v1.23.15
minikube addons enable ingress
kubectl create namespace quarry
helm -n quarry install quarry helm-quarry -f helm-quarry/dev-env.yaml
The rest of the setup instructions will display on screen as long as the install is successful.
Both local setups will create two databases.
One database is your quarry database the other is a wikireplica-like database
named mywiki
. This (or mywiki_p
) is the correct thing to enter in the
database field on all local test queries.
The other database is the Quarry internal db. In your local environment, you can query Quarry internal db itself. Use then "quarry" as database name.
If you had already run a dev environment (that is, ran docker-compose up
) you might want to update
the containers with the new dependencies by running docker-compose build
before running
docker-compose up
again.
To pre-compile nunjucks templates:
nunjucks-precompile quarry/web/static/templates/ > quarry/web/static/templates/compiled.js
See also commands listed in the mainters documentation: https://wikitech.wikimedia.org/wiki/Portal:Data_Services/Admin/Quarry
To have a PR make comments to an associated phabricator ticket have the last line of the commit look like:
Bug:
For example: Bug: T317566
git-crypt is used to encrypt the config.yaml file. We're using the "symmetric key mode" of git-crypt, with a single secret key.
To decrypt ask a maintainer for the secret key and:
git clone https://github.com/toolforge/quarry.git
cd quarry
git-crypt unlock <path to secret key>
A copy of the decryption key is stored in Pwstore.
From quarry-bastion.quarry.eqiad1.wikimedia.cloud
:
git clone https://github.com/toolforge/quarry.git
cd quarry
git checkout <branch> # If not deploying main
git-crypt unlock <path to encryption key>
bash deploy.sh
After a PR has been reviewed, and if the CI runs successfully, the current procedure is to deploy it to production before merging the PR, so that you can verify that it is working correctly.
This procedure makes sure that we never merge non-working code to main
, but
you have to be careful because if you deploy a branch that is stale, you might
undo some recent changes deployed by somebody else.
- Make sure your PR is not stale:
- Github should note that the PR has updates or conflicts and offer to fix them.
- Alternatively:
git checkout main
git pull
git checkout <branch>
git fetch
git rebase origin/main
git push --force
- Wait for the CI to build the image
- Deploy with
bash deploy.sh
- If everything works, merge the PR. No need to redeploy after merging.
- If something breaks, revert your change:
git checkout main
git pull
- Re-deploy with
bash deploy.sh
For a completely fresh deploy, an nfs server will need to be setup. Add its hostname to helm-quarry/prod-env.yaml.
And an object store will need to be generated for the tofu state file. Named "tofu-state"
And setup mysql:
mysql -uquarry -h <trove hostname created in by tofu> -p < schema.sql
After a fresh deploy, go to Horizon and point the web proxy at the new cluster.
- Copy the primary tofu file (today that is tofu/127b.tf) to a new name, e.g. 127c.tf
- In the new file, search and replace the old name with the new name (s/127b/127c/g)
- Remove the 'resource "local_file" "kube_config"' definition from the old .tf file
Now you have an existing blue deployment (127b) and a potential green deployment (127c).
Now when you run deploy.sh, tofu will ignore the old (blue) magnum cluster, and build a new (green) cluster next to it. The old kube.config file will be overwritten with the bindings for the new cluster, and ansible will deploy paws to the new cluster.
If you get to the end of that without warnings, then you have two largely interchangeable clusters. Adjust the web proxy entry in Horizon to point to node-0 in the new magnum cluster and make sure it works. If it doesn't flip the proxy back, destroy the new (green) magnum cluster and repeat as needed.
If all is well, then remove the old/blue tofu file (127.b.tf) and re-deploy, and which point the blue cluster will be torn down by tofu and the remaining setup left intact.
The failover between clusters is largely harmless to users since most state is stored outside the cluster.
If ansible doesn't detect a change for quarry helm the following can be run:
helm -n quarry upgrade --install quarry helm-quarry -f helm-quarry/prod-env.yaml
For shell access, a debug pod can be created on a running node with something lie $ kubectl debug node/quarry-127a-g4ndvpkr5sro-node-0 -it --image debian:stable