# my global config global: #scrape_interval: 1m # Set the interval to scrape data. Default is every 1 minute. #evaluation_interval: 1m # Set interval to evaluate rules. The default is every 1 minute. #scrape_timeout: 10s # Set scrape timeout period. global default (10s). # Alertmanager configuration alerting: alertmanagers: - static_configs: - targets: # - alertmanager:9093 # Load rules once and periodically evaluate them according to the global 'evaluation_interval'. rule_files: # - "first_rules.yml" # - "second_rules.yml" # A scrape configuration containing exactly one endpoint to scrape: # Here it's Prometheus itself. scrape_configs: # The job name is added as a label `job=` to any timeseries scraped from this config. - job_name: 'prometheus' # metrics_path defaults to '/metrics' # scheme defaults to 'http'. static_configs: - targets: ['localhost:9090'] - job_name: 'node_exporter' scrape_interval: 5s static_configs: - targets: ['10.0.0.137:10091'] - job_name: 'blackbox' scrape_timeout: 20s metrics_path: /probe params: module: [http_2xx] # Look for a HTTP 200 response. static_configs: - targets: - https://jaytaala.com - https://bitwarden.jaytaala.com - https://confluence.jaytaala.com - https://chat.jaytaala.com - https://crowd.jaytaala.com - https://gitlab.jaytaala.com - https://url.jaytaala.com - https://opstoolkit.unswcanberraspace.com - https://opstoolkit-dev.unswcanberraspace.com - https://confluence.unswcanberraspace.com - https://bitbucket.unswcanberraspace.com - https://jira.unswcanberraspace.com - https://crowd.unswcanberraspace.com - https://rcode.unswcanberraspace.com - https://survey.unswcanberraspace.com relabel_configs: - source_labels: [__address__] target_label: __param_target - source_labels: [__param_target] target_label: instance - target_label: __address__ replacement: bbexp:9115 # The blackbox exporter's real hostname:port.