Add prometheus, alertmanager, pushgateway configs

Change-Id: I7f146e0207ada58f0723b14d0a625a6bce67ea9f
diff --git a/README.rst b/README.rst
index 0565a0a..754a94e 100644
--- a/README.rst
+++ b/README.rst
@@ -5,6 +5,110 @@
 Power your metrics and alerting with a leading open-source monitoring
 solution.
 
+Sample pillars
+==============
+
+Configure prometheus server
+---------------------------
+
+.. code-block:: yaml
+
+  prometheus:
+    server:
+      enabled: true
+      dir:
+        config: /srv/volumes/prometheus-config
+        config_in_container: /opt/prometheus/config
+      bind:
+        port: 9090
+        address: 0.0.0.0
+      external_port: 15010
+      target:
+        kubernetes:
+          api_ip: 127.0.0.1
+          ssl_dir: /opt/prometheus/config
+          cert_name: kubelet-client.crt
+          key_name: kubelet-client.key
+        etcd: ${etcd:server:members}
+      alert:
+        PrometheusTargetDown:
+          if: 'up != 1'
+          labels:
+            severity: down
+          annotations:
+            summary: 'Prometheus target down'
+      storage:
+        local:
+          engine: "persisted"
+          retention: "360h"
+          memory_chunks: 1048576
+          max_chunks_to_persist: 524288
+          num_fingerprint_mutexes: 4096
+      alertmanager:
+        notification_queue_capacity: 10000
+      config:
+        global:
+          scrape_interval: "15s"
+          scrape_timeout: "15s"
+          evaluation_interval: "1m"
+          external_labels:
+            region: 'region1'
+
+Configure alertmanager
+----------------------
+
+.. code-block:: yaml
+
+  prometheus:
+    alertmanager:
+      enabled: true
+      dir:
+        config: /srv/volumes/prometheus-config
+      bind:
+        address: 0.0.0.0
+        port: 9093
+      external_port: 15011
+      config:
+        global:
+          resolve_timeout: 5m
+        route:
+          group_by: ['alertname', 'region', 'service']
+          group_wait: 60s
+          group_interval: 5m
+          repeat_interval: 3h
+          receiver: HTTP-notification
+        inhibit_rules:
+          - source_match:
+              severity: 'down'
+            target_match:
+              severity: 'critical'
+            equal: ['region', 'service']
+          - source_match:
+              severity: 'down'
+            target_match:
+              severity: 'warning'
+            equal: ['region', 'service']
+          - source_match:
+              severity: 'critical'
+            target_match:
+              severity: 'warning'
+            equal: ['alertname', 'region', 'service']
+        receivers:
+          - name: 'HTTP-notification'
+            webhook_configs:
+              - url: http://127.0.0.1
+                send_resolved: true
+
+Configure pushgateway
+---------------------
+
+.. code-block:: yaml
+
+  prometheus:
+    pushgateway:
+      enabled: true
+      external_port: 15012
+
 Documentation and Bugs
 ======================
 
diff --git a/metadata/service/alertmanager/container.yml b/metadata/service/alertmanager/container.yml
new file mode 100644
index 0000000..1fcf610
--- /dev/null
+++ b/metadata/service/alertmanager/container.yml
@@ -0,0 +1,39 @@
+applications:
+  - alertmanager
+parameters:
+  prometheus:
+    alertmanager:
+      enabled: true
+      bind:
+        address: 0.0.0.0
+        port: 9093
+      config:
+        global:
+          resolve_timeout: 5m
+        route:
+          group_by: ['alertname', 'region', 'service']
+          group_wait: 60s
+          group_interval: 5m
+          repeat_interval: 3h
+          receiver: HTTP-notification
+        inhibit_rules:
+          - source_match:
+              severity: 'down'
+            target_match:
+              severity: 'critical'
+            equal: ['region', 'service']
+          - source_match:
+              severity: 'down'
+            target_match:
+              severity: 'warning'
+            equal: ['region', 'service']
+          - source_match:
+              severity: 'critical'
+            target_match:
+              severity: 'warning'
+            equal: ['alertname', 'region', 'service']
+        receivers:
+          - name: 'HTTP-notification'
+            webhook_configs:
+              - url: http://127.0.0.1
+                send_resolved: true
diff --git a/metadata/service/server.yml b/metadata/service/pushgateway/container.yml
similarity index 64%
rename from metadata/service/server.yml
rename to metadata/service/pushgateway/container.yml
index 79283d9..28c24e6 100644
--- a/metadata/service/server.yml
+++ b/metadata/service/pushgateway/container.yml
@@ -1,6 +1,6 @@
 applications:
-  - prometheus
+  - pushgateway
 parameters:
   prometheus:
-    server:
+    pushgateway:
       enabled: true
diff --git a/metadata/service/server/container.yml b/metadata/service/server/container.yml
new file mode 100644
index 0000000..2269896
--- /dev/null
+++ b/metadata/service/server/container.yml
@@ -0,0 +1,27 @@
+applications:
+  - prometheus
+classes:
+ - service.prometheus.support
+parameters:
+  prometheus:
+    server:
+      enabled: true
+      bind:
+        port: 9090
+        address: 0.0.0.0
+      storage:
+        local:
+          engine: "persisted"
+          retention: "360h"
+          memory_chunks: 1048576
+          max_chunks_to_persist: 524288
+          num_fingerprint_mutexes: 4096
+      alertmanager:
+        notification_queue_capacity: 10000
+      config:
+        global:
+          scrape_interval: "15s"
+          scrape_timeout: "15s"
+          evaluation_interval: "1m"
+          external_labels:
+            region: 'region1'
diff --git a/metadata/service/support.yml b/metadata/service/support.yml
new file mode 100644
index 0000000..76476fd
--- /dev/null
+++ b/metadata/service/support.yml
@@ -0,0 +1,5 @@
+parameters:
+  prometheus:
+    _support:
+      grafana:
+        enabled: true
diff --git a/prometheus/alertmanager.sls b/prometheus/alertmanager.sls
new file mode 100644
index 0000000..576c349
--- /dev/null
+++ b/prometheus/alertmanager.sls
@@ -0,0 +1,12 @@
+{% from "prometheus/map.jinja" import alertmanager with context %}
+{%- if alertmanager.enabled %}
+
+{%- if pillar.docker.host is defined %}
+
+{{alertmanager.dir.config}}/alertmanager.yml:
+  file.managed:
+  - source: salt://prometheus/files/alertmanager.yml
+  - template: jinja
+
+{%- endif %}
+{%- endif %}
diff --git a/prometheus/files/alertmanager.yml b/prometheus/files/alertmanager.yml
new file mode 100644
index 0000000..d699de3
--- /dev/null
+++ b/prometheus/files/alertmanager.yml
@@ -0,0 +1,20 @@
+{% from "prometheus/map.jinja" import alertmanager with context %}
+global:
+{%- if alertmanager.config.global is defined %}
+{{ alertmanager.config.global | yaml(False) | indent(2, true) }}
+{%- endif %}
+
+route:
+{%- if alertmanager.config.route is defined %}
+{{ alertmanager.config.route | yaml(False) | indent(2, true) }}
+{%- endif %}
+
+inhibit_rules:
+{%- if alertmanager.config.inhibit_rules is defined %}
+{{ alertmanager.config.inhibit_rules | yaml(False) | indent(2, true) }}
+{%- endif %}
+
+receivers:
+{%- if alertmanager.config.receivers is defined %}
+{{ alertmanager.config.receivers | yaml(False) | indent(2, true) }}
+{%- endif %}
diff --git a/prometheus/files/alerts.yml b/prometheus/files/alerts.yml
new file mode 100644
index 0000000..ebafbed
--- /dev/null
+++ b/prometheus/files/alerts.yml
@@ -0,0 +1,23 @@
+{%- from "prometheus/map.jinja" import server with context %}
+
+{%- if server.alert is defined %}
+{%- for alertname, alert in server.alert.iteritems() %}
+ALERT {{ alertname }}
+  IF {{ alert.if }}
+  {%- if alert.for is defined %}FOR {{ alert.for }}{%- endif %}
+  {%- if alert.labels is defined %}
+  LABELS {
+    {%- for name, value in alert.labels.iteritems() %}
+    {{ name }} = "{{ value }}"{%- if not loop.last %},{%- endif %}
+    {%- endfor %}
+  }
+  {%- endif %}
+  {%- if alert.annotations is defined %}
+  ANNOTATIONS {
+    {%- for name, value in alert.annotations.iteritems() %}
+    {{ name }} = "{{ value }}"{%- if not loop.last %},{%- endif %}
+    {%- endfor %}
+  }
+  {%- endif %}
+{%- endfor %}
+{%- endif %}
diff --git a/prometheus/files/prometheus.yml b/prometheus/files/prometheus.yml
new file mode 100644
index 0000000..09cc042
--- /dev/null
+++ b/prometheus/files/prometheus.yml
@@ -0,0 +1,174 @@
+{%- from "prometheus/map.jinja" import server with context %}
+
+global:
+{%- if server.get('config', {}).global is defined %}
+{{ server.config.global | yaml(False) | indent(2, true) }}
+{%- endif %}
+
+rule_files:
+- {{ server.dir.config_in_container }}/alerts.yml
+
+scrape_configs:
+{%- set telegraf_nodes = [] %}
+{%- for node_name, node_grains in salt['mine.get']('*', 'grains.items').iteritems() %}
+  {%- if 'telegraf' in node_grains.get('services') %}
+    {%- set node_ip = node_grains.get('prometheus_client').get('address') %}
+    {%- set node_port = node_grains.get('prometheus_client').get('port') %}
+    {%- set telegraf_address = "'%s:%d'" | format(node_ip, node_port) %}
+    {%- do telegraf_nodes.append(telegraf_address) %}
+  {%- endif %}
+{%- endfor %}
+{%- if telegraf_nodes|length > 0 %}
+  - job_name: 'telegraf'
+    static_configs:
+    - targets: [{{ telegraf_nodes | join(',') }}]
+{%- endif %}
+
+{% if server.get('target', {}).etcd is defined %}
+  {%- set etcd_nodes = [] %}
+  {%- for node in server.target.etcd %}
+    {%- set etcd_address = "'%s:%d'" | format(node.host, node.port) %}
+    {%- do etcd_nodes.append(etcd_address) %}
+  {%- endfor %}
+  {%- if etcd_nodes|length > 0 %}
+  - job_name: 'etcd'
+    static_configs:
+    - targets: [{{ etcd_nodes | join(',') }}]
+  {%- endif %}
+{%- endif %}
+
+{% if server.get('target', {}).kubernetes is defined %}
+  - job_name: 'kubernetes-api'
+    scheme: https
+    tls_config:
+      insecure_skip_verify: true
+      {% if server.target.kubernetes.cert_name is defined %}cert_file: {{ server.target.kubernetes.ssl_dir }}/{{ server.target.kubernetes.cert_name }}{%- endif %}
+      {% if server.target.kubernetes.key_name is defined %}key_file: {{ server.target.kubernetes.ssl_dir }}/{{ server.target.kubernetes.key_name }}{%- endif %}
+    kubernetes_sd_configs:
+    - api_server: {{ server.target.kubernetes.api_ip }}
+      role: endpoints
+      tls_config:
+        insecure_skip_verify: true
+        {% if server.target.kubernetes.cert_name is defined %}cert_file: {{ server.target.kubernetes.ssl_dir }}/{{ server.target.kubernetes.cert_name }}{%- endif %}
+        {% if server.target.kubernetes.key_name is defined %}key_file: {{ server.target.kubernetes.ssl_dir }}/{{ server.target.kubernetes.key_name }}{%- endif %}
+    relabel_configs:
+    - source_labels: [__meta_kubernetes_namespace, __meta_kubernetes_service_name, __meta_kubernetes_endpoint_port_name]
+      action: keep
+      regex: default;kubernetes;https
+
+
+  - job_name: 'kubernetes-node'
+    scheme: https
+    tls_config:
+      insecure_skip_verify: true
+      {% if server.target.kubernetes.cert_name is defined %}cert_file: {{ server.target.kubernetes.ssl_dir }}/{{ server.target.kubernetes.cert_name }}{%- endif %}
+      {% if server.target.kubernetes.key_name is defined %}key_file: {{ server.target.kubernetes.ssl_dir }}/{{ server.target.kubernetes.key_name }}{%- endif %}
+    kubernetes_sd_configs:
+    - api_server: {{ server.target.kubernetes.api_ip }}
+      role: node
+      tls_config:
+        insecure_skip_verify: true
+        {% if server.target.kubernetes.cert_name is defined %}cert_file: {{ server.target.kubernetes.ssl_dir }}/{{ server.target.kubernetes.cert_name }}{%- endif %}
+        {% if server.target.kubernetes.key_name is defined %}key_file: {{ server.target.kubernetes.ssl_dir }}/{{ server.target.kubernetes.key_name }}{%- endif %}
+    relabel_configs:
+    - action: labelmap
+      regex: __meta_kubernetes_node_label_(.+)
+
+
+  - job_name: 'kubernetes-service-endpoint'
+    scheme: https
+    tls_config:
+      insecure_skip_verify: true
+      {% if server.target.kubernetes.cert_name is defined %}cert_file: {{ server.target.kubernetes.ssl_dir }}/{{ server.target.kubernetes.cert_name }}{%- endif %}
+      {% if server.target.kubernetes.key_name is defined %}key_file: {{ server.target.kubernetes.ssl_dir }}/{{ server.target.kubernetes.key_name }}{%- endif %}
+    kubernetes_sd_configs:
+    - api_server: {{ server.target.kubernetes.api_ip }}
+      role: endpoints
+      tls_config:
+        insecure_skip_verify: true
+        {% if server.target.kubernetes.cert_name is defined %}cert_file: {{ server.target.kubernetes.ssl_dir }}/{{ server.target.kubernetes.cert_name }}{%- endif %}
+        {% if server.target.kubernetes.key_name is defined %}key_file: {{ server.target.kubernetes.ssl_dir }}/{{ server.target.kubernetes.key_name }}{%- endif %}
+    relabel_configs:
+    - source_labels: [__meta_kubernetes_service_annotation_prometheus_io_scrape]
+      action: keep
+      regex: true
+    - source_labels: [__meta_kubernetes_service_annotation_prometheus_io_scheme]
+      action: replace
+      target_label: __scheme__
+      regex: (https?)
+    - source_labels: [__meta_kubernetes_service_annotation_prometheus_io_path]
+      action: replace
+      target_label: __metrics_path__
+      regex: (.+)
+    - source_labels: [__address__, __meta_kubernetes_service_annotation_prometheus_io_port]
+      action: replace
+      target_label: __address__
+      regex: (.+)(?::\d+);(\d+)
+      replacement: $1:$2
+    - action: labelmap
+      regex: __meta_kubernetes_service_label_(.+)
+    - source_labels: [__meta_kubernetes_service_namespace]
+      action: replace
+      target_label: kubernetes_namespace
+    - source_labels: [__meta_kubernetes_service_name]
+      action: replace
+      target_label: kubernetes_name
+    - source_labels: [__meta_kubernetes_pod_node_name]
+      action: replace
+      target_label: kubernetes_io_hostname
+    - source_labels: [__meta_kubernetes_pod_name]
+      action: replace
+      target_label: kubernetes_pod_name
+
+
+  - job_name: 'kubernetes-pod'
+    scheme: https
+    tls_config:
+      insecure_skip_verify: true
+      {% if server.target.kubernetes.cert_name is defined %}cert_file: {{ server.target.kubernetes.ssl_dir }}/{{ server.target.kubernetes.cert_name }}{%- endif %}
+      {% if server.target.kubernetes.key_name is defined %}key_file: {{ server.target.kubernetes.ssl_dir }}/{{ server.target.kubernetes.key_name }}{%- endif %}
+    kubernetes_sd_configs:
+    - api_server: {{ server.target.kubernetes.api_ip }}
+      role: pod
+      tls_config:
+        insecure_skip_verify: true
+        {% if server.target.kubernetes.cert_name is defined %}cert_file: {{ server.target.kubernetes.ssl_dir }}/{{ server.target.kubernetes.cert_name }}{%- endif %}
+        {% if server.target.kubernetes.key_name is defined %}key_file: {{ server.target.kubernetes.ssl_dir }}/{{ server.target.kubernetes.key_name }}{%- endif %}
+    relabel_configs:
+    - source_labels: [__meta_kubernetes_pod_annotation_prometheus_io_scrape]
+      action: keep
+      regex: true
+    - source_labels: [__meta_kubernetes_pod_annotation_prometheus_io_path]
+      action: replace
+      target_label: __metrics_path__
+      regex: (.+)
+    - source_labels: [__address__, __meta_kubernetes_pod_annotation_prometheus_io_port]
+      action: replace
+      regex: (.+):(?:\d+);(\d+)
+      replacement: ${1}:${2}
+      target_label: __address__
+    - action: labelmap
+      regex: __meta_kubernetes_pod_label_(.+)
+    - source_labels: [__meta_kubernetes_pod_namespace]
+      action: replace
+      target_label: kubernetes_namespace
+    - source_labels: [__meta_kubernetes_pod_name]
+      action: replace
+      target_label: kubernetes_pod_name
+{%- endif %}
+
+
+  - job_name: 'pushgateway'
+    dns_sd_configs:
+    - names:
+      - 'tasks.pushgateway'
+      type: 'A'
+      port: 9091
+
+
+  - job_name: 'prometheus'
+    dns_sd_configs:
+    - names:
+      - 'tasks.prometheus'
+      type: 'A'
+      port: {{ server.bind.port }}
diff --git a/prometheus/init.sls b/prometheus/init.sls
new file mode 100644
index 0000000..e69de29
--- /dev/null
+++ b/prometheus/init.sls
diff --git a/prometheus/map.jinja b/prometheus/map.jinja
new file mode 100644
index 0000000..f6a1c70
--- /dev/null
+++ b/prometheus/map.jinja
@@ -0,0 +1,16 @@
+{% set server = salt['grains.filter_by']({
+    'default': {
+        'dir': {
+            'config': '/srv/volumes/prometheus-config',
+            'config_in_container': '/opt/prometheus/config'
+        },
+    },
+}, merge=salt['pillar.get']('prometheus:server')) %}
+
+{% set alertmanager = salt['grains.filter_by']({
+    'default': {
+        'dir': {
+            'config': '/srv/volumes/prometheus-config',
+        },
+    },
+}, merge=salt['pillar.get']('prometheus:alertmanager')) %}
diff --git a/prometheus/server.sls b/prometheus/server.sls
new file mode 100644
index 0000000..fc0a181
--- /dev/null
+++ b/prometheus/server.sls
@@ -0,0 +1,17 @@
+{% from "prometheus/map.jinja" import server with context %}
+{%- if server.enabled %}
+
+{%- if pillar.docker.host is defined %}
+
+{{server.dir.config}}/prometheus.yml:
+  file.managed:
+  - source: salt://prometheus/files/prometheus.yml
+  - template: jinja
+
+{{server.dir.config}}/alerts.yml:
+  file.managed:
+  - source: salt://prometheus/files/alerts.yml
+  - template: jinja
+
+{%- endif %}
+{%- endif %}
diff --git a/tests/pillar/prometheus_alertmanager.sls b/tests/pillar/prometheus_alertmanager.sls
new file mode 100644
index 0000000..48a9d5d
--- /dev/null
+++ b/tests/pillar/prometheus_alertmanager.sls
@@ -0,0 +1,36 @@
+prometheus:
+  alertmanager:
+    enabled: true
+    bind:
+      address: 0.0.0.0
+      port: 9093
+    config:
+      global:
+        resolve_timeout: 5m
+      route:
+        group_by: ['alertname', 'region', 'service']
+        group_wait: 60s
+        group_interval: 5m
+        repeat_interval: 3h
+        receiver: HTTP-notification
+      inhibit_rules:
+        - source_match:
+            severity: 'down'
+          target_match:
+            severity: 'critical'
+          equal: ['region', 'service']
+        - source_match:
+            severity: 'down'
+          target_match:
+            severity: 'warning'
+          equal: ['region', 'service']
+        - source_match:
+            severity: 'critical'
+          target_match:
+            severity: 'warning'
+          equal: ['alertname', 'region', 'service']
+      receivers:
+        - name: 'HTTP-notification'
+          webhook_configs:
+            - url: http://127.0.0.1
+              send_resolved: true
diff --git a/tests/pillar/prometheus_pushgateway.sls b/tests/pillar/prometheus_pushgateway.sls
new file mode 100644
index 0000000..7921ba1
--- /dev/null
+++ b/tests/pillar/prometheus_pushgateway.sls
@@ -0,0 +1,3 @@
+prometheus:
+  pushgateway:
+    enabled: true
diff --git a/tests/pillar/prometheus_server.sls b/tests/pillar/prometheus_server.sls
new file mode 100644
index 0000000..c967b8f
--- /dev/null
+++ b/tests/pillar/prometheus_server.sls
@@ -0,0 +1,30 @@
+prometheus:
+  server:
+    enabled: true
+    bind:
+      port: 9090
+      address: 0.0.0.0
+    storage:
+      local:
+        engine: "persisted"
+        retention: "360h"
+        memory_chunks: 1048576
+        max_chunks_to_persist: 524288
+        num_fingerprint_mutexes: 4096
+    alertmanager:
+      notification_queue_capacity: 10000
+    config:
+      global:
+        scrape_interval: "15s"
+        scrape_timeout: "15s"
+        evaluation_interval: "1m"
+        external_labels:
+          region: 'region1'
+    alert:
+      PrometheusTargetDownKubernetesNodes:
+        if: 'up{job="kubernetes-nodes"} != 1'
+        labels:
+          severity: down
+          service: prometheus
+        annotations:
+          summary: 'Prometheus target down'
diff --git a/tests/run_tests.sh b/tests/run_tests.sh
new file mode 100755
index 0000000..3f42101
--- /dev/null
+++ b/tests/run_tests.sh
@@ -0,0 +1,162 @@
+#!/usr/bin/env bash
+
+set -e
+[ -n "$DEBUG" ] && set -x
+
+CURDIR="$( cd "$( dirname "${BASH_SOURCE[0]}" )" && pwd )"
+METADATA=${CURDIR}/../metadata.yml
+FORMULA_NAME=$(cat $METADATA | python -c "import sys,yaml; print yaml.load(sys.stdin)['name']")
+
+## Overrideable parameters
+PILLARDIR=${PILLARDIR:-${CURDIR}/pillar}
+BUILDDIR=${BUILDDIR:-${CURDIR}/build}
+VENV_DIR=${VENV_DIR:-${BUILDDIR}/virtualenv}
+DEPSDIR=${BUILDDIR}/deps
+
+SALT_FILE_DIR=${SALT_FILE_DIR:-${BUILDDIR}/file_root}
+SALT_PILLAR_DIR=${SALT_PILLAR_DIR:-${BUILDDIR}/pillar_root}
+SALT_CONFIG_DIR=${SALT_CONFIG_DIR:-${BUILDDIR}/salt}
+SALT_CACHE_DIR=${SALT_CACHE_DIR:-${SALT_CONFIG_DIR}/cache}
+
+SALT_OPTS="${SALT_OPTS} --retcode-passthrough --local -c ${SALT_CONFIG_DIR}"
+
+if [ "x${SALT_VERSION}" != "x" ]; then
+    PIP_SALT_VERSION="==${SALT_VERSION}"
+fi
+
+## Functions
+log_info() {
+    echo "[INFO] $*"
+}
+
+log_err() {
+    echo "[ERROR] $*" >&2
+}
+
+setup_virtualenv() {
+    log_info "Setting up Python virtualenv"
+    virtualenv $VENV_DIR
+    source ${VENV_DIR}/bin/activate
+    pip install salt${PIP_SALT_VERSION}
+}
+
+setup_pillar() {
+    [ ! -d ${SALT_PILLAR_DIR} ] && mkdir -p ${SALT_PILLAR_DIR}
+    echo "base:" > ${SALT_PILLAR_DIR}/top.sls
+    for pillar in ${PILLARDIR}/*; do
+        state_name=$(basename ${pillar%.sls})
+        echo -e "  ${state_name}:\n    - ${state_name}" >> ${SALT_PILLAR_DIR}/top.sls
+    done
+}
+
+setup_salt() {
+    [ ! -d ${SALT_FILE_DIR} ] && mkdir -p ${SALT_FILE_DIR}
+    [ ! -d ${SALT_CONFIG_DIR} ] && mkdir -p ${SALT_CONFIG_DIR}
+    [ ! -d ${SALT_CACHE_DIR} ] && mkdir -p ${SALT_CACHE_DIR}
+
+    echo "base:" > ${SALT_FILE_DIR}/top.sls
+    for pillar in ${PILLARDIR}/*.sls; do
+        state_name=$(basename ${pillar%.sls})
+        echo -e "  ${state_name}:\n    - ${FORMULA_NAME}" >> ${SALT_FILE_DIR}/top.sls
+    done
+
+    cat << EOF > ${SALT_CONFIG_DIR}/minion
+file_client: local
+cachedir: ${SALT_CACHE_DIR}
+verify_env: False
+
+file_roots:
+  base:
+  - ${SALT_FILE_DIR}
+  - ${CURDIR}/..
+  - /usr/share/salt-formulas/env
+
+pillar_roots:
+  base:
+  - ${SALT_PILLAR_DIR}
+  - ${PILLARDIR}
+EOF
+}
+
+fetch_dependency() {
+    dep_name="$(echo $1|cut -d : -f 1)"
+    dep_source="$(echo $1|cut -d : -f 2-)"
+    dep_root="${DEPSDIR}/$(basename $dep_source .git)"
+    dep_metadata="${dep_root}/metadata.yml"
+
+    [ -d /usr/share/salt-formulas/env/${dep_name} ] && log_info "Dependency $dep_name already present in system-wide salt env" && return 0
+    [ -d $dep_root ] && log_info "Dependency $dep_name already fetched" && return 0
+
+    log_info "Fetching dependency $dep_name"
+    [ ! -d ${DEPSDIR} ] && mkdir -p ${DEPSDIR}
+    git clone $dep_source ${DEPSDIR}/$(basename $dep_source .git)
+    ln -s ${dep_root}/${dep_name} ${SALT_FILE_DIR}/${dep_name}
+
+    METADATA="${dep_metadata}" install_dependencies
+}
+
+install_dependencies() {
+    grep -E "^dependencies:" ${METADATA} >/dev/null || return 0
+    (python - | while read dep; do fetch_dependency "$dep"; done) << EOF
+import sys,yaml
+for dep in yaml.load(open('${METADATA}', 'ro'))['dependencies']:
+    print '%s:%s' % (dep["name"], dep["source"])
+EOF
+}
+
+clean() {
+    log_info "Cleaning up ${BUILDDIR}"
+    [ -d ${BUILDDIR} ] && rm -rf ${BUILDDIR} || exit 0
+}
+
+salt_run() {
+    [ -e ${VEN_DIR}/bin/activate ] && source ${VENV_DIR}/bin/activate
+    salt-call ${SALT_OPTS} $*
+}
+
+prepare() {
+    [ -d ${BUILDDIR} ] && mkdir -p ${BUILDDIR}
+
+    which salt-call || setup_virtualenv
+    setup_pillar
+    setup_salt
+    install_dependencies
+}
+
+run() {
+    for pillar in ${PILLARDIR}/*.sls; do
+        state_name=$(basename ${pillar%.sls})
+        salt_run --id=${state_name} state.show_sls ${FORMULA_NAME} || (log_err "Execution of ${FORMULA_NAME}.${state_name} failed"; exit 1)
+    done
+}
+
+_atexit() {
+    RETVAL=$?
+    trap true INT TERM EXIT
+
+    if [ $RETVAL -ne 0 ]; then
+        log_err "Execution failed"
+    else
+        log_info "Execution successful"
+    fi
+    return $RETVAL
+}
+
+## Main
+trap _atexit INT TERM EXIT
+
+case $1 in
+    clean)
+        clean
+        ;;
+    prepare)
+        prepare
+        ;;
+    run)
+        run
+        ;;
+    *)
+        prepare
+        run
+        ;;
+esac